Literature DB >> 32679578

Linear versus deep learning methods for noisy speech separation for EEG-informed attention decoding.

Neetha Das1, Jeroen Zegers, Hugo Van Hamme, Tom Francart, Alexander Bertrand.   

Abstract

OBJECTIVE: A hearing aid's noise reduction algorithm cannot infer to which speaker the user intends to listen to. Auditory attention decoding (AAD) algorithms allow to infer this information from neural signals, which leads to the concept of neuro-steered hearing aids. We aim to evaluate and demonstrate the feasibility of AAD-supported speech enhancement in challenging noisy conditions based on electroencephalography recordings. APPROACH: The AAD performance with a linear versus a deep neural network (DNN) based speaker separation was evaluated for same-gender speaker mixtures using three different speaker positions and three different noise conditions. MAIN
RESULTS: AAD results based on the linear approach were found to be at least on par and sometimes even better than pure DNN-based approaches in terms of AAD accuracy in all tested conditions. However, when using the DNN to support a linear data-driven beamformer, a performance improvement over the purely linear approach was obtained in the most challenging scenarios. The use of multiple microphones was also found to improve speaker separation and AAD performance over single-microphone systems. SIGNIFICANCE: Recent proof-of-concept studies in this context each focus on a different method in a different experimental setting, which makes it hard to compare them. Furthermore, they are tested in highly idealized experimental conditions, which are still far from a realistic hearing aid setting. This work provides a systematic comparison of a linear and non-linear neuro-steered speech enhancement model, as well as a more realistic validation in challenging conditions.

Entities:  

Mesh:

Year:  2020        PMID: 32679578     DOI: 10.1088/1741-2552/aba6f8

Source DB:  PubMed          Journal:  J Neural Eng        ISSN: 1741-2552            Impact factor:   5.379


  4 in total

Review 1.  Harnessing the Power of Artificial Intelligence in Otolaryngology and the Communication Sciences.

Authors:  Blake S Wilson; Debara L Tucci; David A Moses; Edward F Chang; Nancy M Young; Fan-Gang Zeng; Nicholas A Lesica; Andrés M Bur; Hannah Kavookjian; Caroline Mussatto; Joseph Penn; Sara Goodwin; Shannon Kraft; Guanghui Wang; Jonathan M Cohen; Geoffrey S Ginsburg; Geraldine Dawson; Howard W Francis
Journal:  J Assoc Res Otolaryngol       Date:  2022-04-20

2.  A Speech-Level-Based Segmented Model to Decode the Dynamic Auditory Attention States in the Competing Speaker Scenes.

Authors:  Lei Wang; Yihan Wang; Zhixing Liu; Ed X Wu; Fei Chen
Journal:  Front Neurosci       Date:  2022-02-10       Impact factor: 4.677

3.  Synchronization of ear-EEG and audio streams in a portable research hearing device.

Authors:  Steffen Dasenbrock; Sarah Blum; Paul Maanen; Stefan Debener; Volker Hohmann; Hendrik Kayser
Journal:  Front Neurosci       Date:  2022-09-01       Impact factor: 5.152

4.  A particle swarm optimization improved BP neural network intelligent model for electrocardiogram classification.

Authors:  Guixiang Li; Zhongwei Tan; Weikang Xu; Fei Xu; Lei Wang; Jun Chen; Kai Wu
Journal:  BMC Med Inform Decis Mak       Date:  2021-07-30       Impact factor: 2.796

  4 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.