Literature DB >> 21682422

Processing speech signal using auditory-like filterbank provides least uncertainty about articulatory gestures.

Prasanta Kumar Ghosh1, Louis M Goldstein, Shrikanth S Narayanan.   

Abstract

Understanding how the human speech production system is related to the human auditory system has been a perennial subject of inquiry. To investigate the production-perception link, in this paper, a computational analysis has been performed using the articulatory movement data obtained during speech production with concurrently recorded acoustic speech signals from multiple subjects in three different languages: English, Cantonese, and Georgian. The form of articulatory gestures during speech production varies across languages, and this variation is considered to be reflected in the articulatory position and kinematics. The auditory processing of the acoustic speech signal is modeled by a parametric representation of the cochlear filterbank which allows for realizing various candidate filterbank structures by changing the parameter value. Using mathematical communication theory, it is found that the uncertainty about the articulatory gestures in each language is maximally reduced when the acoustic speech signal is represented using the output of a filterbank similar to the empirically established cochlear filterbank in the human auditory system. Possible interpretations of this finding are discussed.
© 2011 Acoustical Society of America

Entities:  

Mesh:

Year:  2011        PMID: 21682422      PMCID: PMC3135153          DOI: 10.1121/1.3573987

Source DB:  PubMed          Journal:  J Acoust Soc Am        ISSN: 0001-4966            Impact factor:   1.840


  10 in total

1.  Efficient coding of natural sounds.

Authors:  Michael S Lewicki
Journal:  Nat Neurosci       Date:  2002-04       Impact factor: 24.884

Review 2.  The molecular architecture of the inner ear.

Authors:  Andrew Forge; Tony Wright
Journal:  Br Med Bull       Date:  2002       Impact factor: 4.291

3.  Listening to speech activates motor areas involved in speech production.

Authors:  Stephen M Wilson; Ayşe Pinar Saygin; Martin I Sereno; Marco Iacoboni
Journal:  Nat Neurosci       Date:  2004-06-06       Impact factor: 24.884

4.  Electromagnetic midsagittal articulometer systems for transducing speech articulatory movements.

Authors:  J S Perkell; M H Cohen; M A Svirsky; M L Matthies; I Garabieta; M T Jackson
Journal:  J Acoust Soc Am       Date:  1992-12       Impact factor: 1.840

5.  Efficient auditory coding.

Authors:  Evan C Smith; Michael S Lewicki
Journal:  Nature       Date:  2006-02-23       Impact factor: 49.962

6.  Role of articulation in speech perception: clues from production.

Authors:  B Lindblom
Journal:  J Acoust Soc Am       Date:  1996-03       Impact factor: 1.840

Review 7.  Perception of the speech code.

Authors:  A M Liberman; F S Cooper; D P Shankweiler; M Studdert-Kennedy
Journal:  Psychol Rev       Date:  1967-11       Impact factor: 8.934

8.  The motor theory of speech perception revised.

Authors:  A M Liberman; I G Mattingly
Journal:  Cognition       Date:  1985-10

9.  Comparative morphometry of mammalian central auditory systems: variation in nuclei and form of the ascending system.

Authors:  K K Glendenning; R B Masterton
Journal:  Brain Behav Evol       Date:  1998       Impact factor: 1.808

10.  Bark frequency transform using an arbitrary order allpass filter.

Authors:  Prasanta Kumar Ghosh; Shrikanth S Narayanan
Journal:  IEEE Signal Process Lett       Date:  2010-03       Impact factor: 3.109

  10 in total
  5 in total

1.  Nonlinear auditory models yield new insights into representations of vowels.

Authors:  Laurel H Carney; Joyce M McDonough
Journal:  Atten Percept Psychophys       Date:  2019-05       Impact factor: 2.199

2.  Directly data-derived articulatory gesture-like representations retain discriminatory information about phone categories.

Authors:  Vikram Ramanarayanan; Maarten Van Segbroeck; Shrikanth S Narayanan
Journal:  Comput Speech Lang       Date:  2015-03-21       Impact factor: 1.899

3.  Speech Coding in the Brain: Representation of Vowel Formants by Midbrain Neurons Tuned to Sound Fluctuations

Authors:  Laurel H Carney; Tianhao Li; Joyce M McDonough
Journal:  eNeuro       Date:  2015-07-20

4.  Advances in real-time magnetic resonance imaging of the vocal tract for speech science and technology research.

Authors:  Asterios Toutios; Shrikanth S Narayanan
Journal:  APSIPA Trans Signal Inf Process       Date:  2016-03-31

5.  Articulating What Infants Attune to in Native Speech.

Authors:  Catherine T Best; Louis M Goldstein; Hosung Nam; Michael D Tyler
Journal:  Ecol Psychol       Date:  2016-11-01
  5 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.