| Literature DB >> 33643116 |
Karol Chlasta1,2, Krzysztof Wołk1.
Abstract
Dementia, a prevalent disorder of the brain, has negative effects on individuals and society. This paper concerns using Spontaneous Speech (ADReSS) Challenge of Interspeech 2020 to classify Alzheimer's dementia. We used (1) VGGish, a deep, pretrained, Tensorflow model as an audio feature extractor, and Scikit-learn classifiers to detect signs of dementia in speech. Three classifiers (LinearSVM, Perceptron, 1NN) were 59.1% accurate, which was 3% above the best-performing baseline models trained on the acoustic features used in the challenge. We also proposed (2) DemCNN, a new PyTorch raw waveform-based convolutional neural network model that was 63.6% accurate, 7% more accurate then the best-performing baseline linear discriminant analysis model. We discovered that audio transfer learning with a pretrained VGGish feature extractor performs better than the baseline approach using automatically extracted acoustic features. Our DepCNN exhibits good generalization capabilities. Both methods presented in this paper offer progress toward new, innovative, and more effective computer-based screening of dementia through spontaneous speech.Entities:
Keywords: affective computing; convolutional neural network; dementia detection; machine learning; mental health monitoring; prosodic analysis; speech technology; transfer learning
Year: 2021 PMID: 33643116 PMCID: PMC7907518 DOI: 10.3389/fpsyg.2020.623237
Source DB: PubMed Journal: Front Psychol ISSN: 1664-1078