Interspeech 2021 2021
DOI: 10.21437/interspeech.2021-1850
|View full text |Cite
|
Sign up to set email alerts
|

Automatic Detection and Assessment of Alzheimer Disease Using Speech and Language Technologies in Low-Resource Scenarios

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
17
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
3
2

Relationship

1
8

Authors

Journals

citations
Cited by 30 publications
(23 citation statements)
references
References 23 publications
0
17
0
Order By: Relevance
“…The used subset has 87 recordings from speakers with AD and 79 from control subjects. Considering the small amount of the utterances in the data, we ran 10-fold CV as used in [26] where each fold has class-balanced distribution.…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…The used subset has 87 recordings from speakers with AD and 79 from control subjects. Considering the small amount of the utterances in the data, we ran 10-fold CV as used in [26] where each fold has class-balanced distribution.…”
Section: Methodsmentioning
confidence: 99%
“…Using a multitask objective, e.g., emotion classification loss plus Alzheimer's disease (AD) classification loss, can be one way to increase the total number of samples, but an AD dataset as a medical data also usually has a small number of samples, thus making the multitask objective training still not solving the data scarcity problem better than transfer learning. In [25], [26], the authors used pre-trained speaker classification, and encoder-decoder ASR models for AD detection.…”
Section: Introductionmentioning
confidence: 99%
“…Pappagari et al ( 2021 ) proposed several acoustic and language models and fused the two modalities by using the output probabilities of the individual models as the inputs to a Logistic Regression classifier for obtaining a final prediction. Regarding the acoustic models, they used an end-to-end classifier by fine-tuning an x-vector model and trained a Logistic Regression and XGBoost classifier using features extracted via several open-source libraries.…”
Section: Related Workmentioning
confidence: 99%
“…2) EGEMAPSV02 : The extended Geneva Minimalistic Acoustic Parameter Set (eGeMAPS) features are a selected standardized set of statistical features that characterize affective physiological changes in voice production. We extracted these features for the entire recording, as this feature set was shown to be usable for atypical speech [22] and was successfully used for classifying AD from speech [23,24]. 3) WAV2VEC : in order to create audio representations using this approach, we make use of the huggingface 1 implementation of the wav2vec 2.0 [25] base model wav2vec2-base-960h.…”
Section: Feature Extractionmentioning
confidence: 99%