Literature DB >> 3372872

Learning the hidden structure of speech.

J L Elman1, D Zipser.   

Abstract

In the work described here, the backpropagation neural network learning procedure is applied to the analysis and recognition of speech. This procedure takes a set of input/output pattern pairs and attempts to learn their functional relationship; it develops the necessary representational features during the course of learning. A series of computer simulation studies was carried out to assess the ability of these networks to accurately label sounds, to learn to recognize sounds without labels, and to learn feature representations of continuous speech. These studies demonstrated that the networks can learn to label presegmented test tokens with accuracies of up to 95%. Networks trained on segmented sounds using a strategy that requires no external labels were able to recognize and delineate sounds in continuous speech. These networks developed rich internal representations that included units which corresponded to such traditional distinctions as vowels and consonants, as well as units that were sensitive to novel and nonstandard features. Networks trained on a large corpus of unsegmented, continuous speech without labels also developed interesting feature representations, which may be useful in both segmentation and label learning. The results of these studies, while preliminary, demonstrate that backpropagation learning can be used with complex, natural data to identify a feature structure that can serve as the basis for both analysis and nontrivial pattern recognition.

Mesh:

Year:  1988        PMID: 3372872     DOI: 10.1121/1.395916

Source DB:  PubMed          Journal:  J Acoust Soc Am        ISSN: 0001-4966            Impact factor:   1.840


  13 in total

1.  Systematic determination of order parameters for chain dynamics using diffusion maps.

Authors:  Andrew L Ferguson; Athanassios Z Panagiotopoulos; Pablo G Debenedetti; Ioannis G Kevrekidis
Journal:  Proc Natl Acad Sci U S A       Date:  2010-07-19       Impact factor: 11.205

2.  Development of a two-stage procedure for the automatic recognition of dysfluencies in the speech of children who stutter: II. ANN recognition of repetitions and prolongations with supplied word segment markers.

Authors:  P Howell; S Sackin; K Glenn
Journal:  J Speech Lang Hear Res       Date:  1997-10       Impact factor: 2.297

3.  Using neural networks to diagnose cancer.

Authors:  P S Maclin; J Dempsey; J Brooks; J Rand
Journal:  J Med Syst       Date:  1991-02       Impact factor: 4.460

4.  Auto-association by multilayer perceptrons and singular value decomposition.

Authors:  H Bourlard; Y Kamp
Journal:  Biol Cybern       Date:  1988       Impact factor: 2.086

5.  Direct Associations or Internal Transformations? Exploring the Mechanisms Underlying Sequential Learning Behavior.

Authors:  Todd M Gureckis; Bradley C Love
Journal:  Cogn Sci       Date:  2010

Review 6.  Relative cue encoding in the context of sophisticated models of categorization: Separating information from categorization.

Authors:  Keith S Apfelbaum; Bob McMurray
Journal:  Psychon Bull Rev       Date:  2015-08

7.  Mice can learn phonetic categories.

Authors:  Jonny L Saunders; Michael Wehr
Journal:  J Acoust Soc Am       Date:  2019-03       Impact factor: 1.840

8.  Using an artificial neural network to diagnose hepatic masses.

Authors:  P S Maclin; J Dempsey
Journal:  J Med Syst       Date:  1992-10       Impact factor: 4.460

9.  Statistical learning of phonetic categories: insights from a computational approach.

Authors:  Bob McMurray; Richard N Aslin; Joseph C Toscano
Journal:  Dev Sci       Date:  2009-04

10.  SORN: a self-organizing recurrent neural network.

Authors:  Andreea Lazar; Gordon Pipa; Jochen Triesch
Journal:  Front Comput Neurosci       Date:  2009-10-30       Impact factor: 2.380

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.