OBJECTIVE: People who suffer from hearing impairments can find it difficult to follow a conversation in a multi-speaker environment. Current hearing aids can suppress background noise; however, there is little that can be done to help a user attend to a single conversation amongst many without knowing which speaker the user is attending to. Cognitively controlled hearing aids that use auditory attention decoding (AAD) methods are the next step in offering help. Translating the successes in AAD research to real-world applications poses a number of challenges, including the lack of access to the clean sound sources in the environment with which to compare with the neural signals. We propose a novel framework that combines single-channel speech separation algorithms with AAD. APPROACH: We present an end-to-end system that (1) receives a single audio channel containing a mixture of speakers that is heard by a listener along with the listener's neural signals, (2) automatically separates the individual speakers in the mixture, (3) determines the attended speaker, and (4) amplifies the attended speaker's voice to assist the listener. MAIN RESULTS: Using invasive electrophysiology recordings, we identified the regions of the auditory cortex that contribute to AAD. Given appropriate electrode locations, our system is able to decode the attention of subjects and amplify the attended speaker using only the mixed audio. Our quality assessment of the modified audio demonstrates a significant improvement in both subjective and objective speech quality measures. SIGNIFICANCE: Our novel framework for AAD bridges the gap between the most recent advancements in speech processing technologies and speech prosthesis research and moves us closer to the development of cognitively controlled hearable devices for the hearing impaired.
OBJECTIVE:People who suffer from hearing impairments can find it difficult to follow a conversation in a multi-speaker environment. Current hearing aids can suppress background noise; however, there is little that can be done to help a user attend to a single conversation amongst many without knowing which speaker the user is attending to. Cognitively controlled hearing aids that use auditory attention decoding (AAD) methods are the next step in offering help. Translating the successes in AAD research to real-world applications poses a number of challenges, including the lack of access to the clean sound sources in the environment with which to compare with the neural signals. We propose a novel framework that combines single-channel speech separation algorithms with AAD. APPROACH: We present an end-to-end system that (1) receives a single audio channel containing a mixture of speakers that is heard by a listener along with the listener's neural signals, (2) automatically separates the individual speakers in the mixture, (3) determines the attended speaker, and (4) amplifies the attended speaker's voice to assist the listener. MAIN RESULTS: Using invasive electrophysiology recordings, we identified the regions of the auditory cortex that contribute to AAD. Given appropriate electrode locations, our system is able to decode the attention of subjects and amplify the attended speaker using only the mixed audio. Our quality assessment of the modified audio demonstrates a significant improvement in both subjective and objective speech quality measures. SIGNIFICANCE: Our novel framework for AAD bridges the gap between the most recent advancements in speech processing technologies and speech prosthesis research and moves us closer to the development of cognitively controlled hearable devices for the hearing impaired.
Authors: James A O'Sullivan; Alan J Power; Nima Mesgarani; Siddharth Rajaram; John J Foxe; Barbara G Shinn-Cunningham; Malcolm Slaney; Shihab A Shamma; Edmund C Lalor Journal: Cereb Cortex Date: 2014-01-15 Impact factor: 5.357
Authors: Elana M Zion Golumbic; Nai Ding; Stephan Bickel; Peter Lakatos; Catherine A Schevon; Guy M McKhann; Robert R Goodman; Ronald Emerson; Ashesh D Mehta; Jonathan Z Simon; David Poeppel; Charles E Schroeder Journal: Neuron Date: 2013-03-06 Impact factor: 17.173
Authors: Daniel Lenz; Marcus Jeschke; Jeanette Schadow; Nicole Naue; Frank W Ohl; Christoph S Herrmann Journal: Brain Res Date: 2007-10-28 Impact factor: 3.252
Authors: Jane E Huggins; Christoph Guger; Erik Aarnoutse; Brendan Allison; Charles W Anderson; Steven Bedrick; Walter Besio; Ricardo Chavarriaga; Jennifer L Collinger; An H Do; Christian Herff; Matthias Hohmann; Michelle Kinsella; Kyuhwa Lee; Fabien Lotte; Gernot Müller-Putz; Anton Nijholt; Elmar Pels; Betts Peters; Felix Putze; Rüdiger Rupp; Gerwin Schalk; Stephanie Scott; Michael Tangermann; Paul Tubig; Thorsten Zander Journal: Brain Comput Interfaces (Abingdon) Date: 2019-12-10
Authors: James O'Sullivan; Jose Herrero; Elliot Smith; Catherine Schevon; Guy M McKhann; Sameer A Sheth; Ashesh D Mehta; Nima Mesgarani Journal: Neuron Date: 2019-10-21 Impact factor: 17.173
Authors: Blake S Wilson; Debara L Tucci; David A Moses; Edward F Chang; Nancy M Young; Fan-Gang Zeng; Nicholas A Lesica; Andrés M Bur; Hannah Kavookjian; Caroline Mussatto; Joseph Penn; Sara Goodwin; Shannon Kraft; Guanghui Wang; Jonathan M Cohen; Geoffrey S Ginsburg; Geraldine Dawson; Howard W Francis Journal: J Assoc Res Otolaryngol Date: 2022-04-20
Authors: Florian Denk; Marleen Grzybowski; Stephan M A Ernst; Birger Kollmeier; Stefan Debener; Martin G Bleichner Journal: Trends Hear Date: 2018 Jan-Dec Impact factor: 3.293
Authors: Sina Miran; Sahar Akram; Alireza Sheikhattar; Jonathan Z Simon; Tao Zhang; Behtash Babadi Journal: Front Neurosci Date: 2018-05-01 Impact factor: 4.677