Interspeech 2020 2020
DOI: 10.21437/interspeech.2020-2635
|View full text |Cite
|
Sign up to set email alerts
|

A Comparison of Acoustic and Linguistics Methodologies for Alzheimer’s Dementia Recognition

Abstract: In the light of the current COVID-19 pandemic, the need for remote digital health assessment tools is greater than ever. This statement is especially pertinent for elderly and vulnerable populations. In this regard, the INTERSPEECH 2020 Alzheimer's Dementia Recognition through Spontaneous Speech (ADReSS) Challenge offers competitors the opportunity to develop speech and language-based systems for the task of Alzheimer's Dementia (AD) recognition. The challenge data consists of speech recordings and their trans… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
27
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
2
1

Relationship

2
6

Authors

Journals

citations
Cited by 34 publications
(27 citation statements)
references
References 27 publications
0
27
0
Order By: Relevance
“…The top-performing model was proposed by Yuan et al [27] whose approach achieved an accuracy of 89.60% compared to the challenge baseline of 75.00%. The second placed model was proposed by us [26], and it achieved an accuracy of 85.42%, which was marginally better than the third placed model, proposed by Cummins et al [25], who achieved an accuracy of 85.20%. For the regression task of the challenge, only three participants were successful at improving the baseline RMSE score of 4.34.…”
Section: Introductionmentioning
confidence: 74%
See 1 more Smart Citation
“…The top-performing model was proposed by Yuan et al [27] whose approach achieved an accuracy of 89.60% compared to the challenge baseline of 75.00%. The second placed model was proposed by us [26], and it achieved an accuracy of 85.42%, which was marginally better than the third placed model, proposed by Cummins et al [25], who achieved an accuracy of 85.20%. For the regression task of the challenge, only three participants were successful at improving the baseline RMSE score of 4.34.…”
Section: Introductionmentioning
confidence: 74%
“…Cummins et al [25], proposed a multimodal fusion system as part of their solution for the ADReSS challenge. For the audio modality, they used three types of acoustic feature representations, which included (a) the popular bag-of-audio words (BoAW) feature aggregation method for acoustic lowlevel descriptors [37], (b) an end-to-end (e2e) convolutional neural network which learns to classify using raw audio waveforms, and (c) a siamese network which learns to classify using Mel spectrogram representation of the subjects' speech signal.…”
Section: Introductionmentioning
confidence: 99%
“…On the balanced DementiaBank dataset using both linguistic and paralinguistic features, an 87.5% classification accuracy was achieved using a Random Forest classifier (Farrús and Codina-Filbà, 2020 ) and an 85.2% using a fusion deep learning approach (Cummins et al, 2020 ). On a different subset of 167 samples from DementiaBank, combining linguistic and paralinguistic features yielded an 81% accuracy (Fraser et al, 2016 ).…”
Section: Discussionmentioning
confidence: 99%
“…The English baseline classifier with all features (on the same data set as Cummins et al, 2020;Farrús and Codina-Filbà, 2020) achieved an AUC of 0.72 and accuracy of 69.7% using a LR classifier. In comparison, the English classifier with generalizable language features achieved an AUC of 0.87 and an accuracy of 76.4% using a LR model.…”
Section: Comparison To Baselinementioning
confidence: 99%
“…Previous work has been done using the ADReSS dataset. Some researchers only participated in the AD classification task (Edwards et al, 2020 ; Pompili et al, 2020 ; Yuan et al, 2020 ), others only participated in the Mini-Mental State Examination (MMSE) prediction task (Farzana and Parde, 2020 ), and others participated in both tasks (Balagopalan et al, 2020 ; Cummins et al, 2020 ; Koo et al, 2020 ; Luz et al, 2020 ; Martinc and Pollak, 2020 ; Pappagari et al, 2020 ; Rohanian et al, 2020 ; Sarawgi et al, 2020 ; Searle et al, 2020 ; Syed et al, 2020 ). The best performance on the AD classification task was achieved by Yuan et al ( 2020 ), who obtained an accuracy of 89.6% on the test set using linguistic features extracted from the transcripts, as well as encoded pauses.…”
Section: Introductionmentioning
confidence: 99%