Interspeech 2020 2020
DOI: 10.21437/interspeech.2020-2049
|View full text |Cite
|
Sign up to set email alerts
|

Aphasic Speech Recognition Using a Mixture of Speech Intelligibility Experts

Abstract: Robust speech recognition is a key prerequisite for semantic feature extraction in automatic aphasic speech analysis. However, standard one-size-fits-all automatic speech recognition models perform poorly when applied to aphasic speech. One reason for this is the wide range of speech intelligibility due to different levels of severity (i.e., higher severity lends itself to less intelligible speech). To address this, we propose a novel acoustic model based on a mixture of experts (MoE), which handles the varyin… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

2
4
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 6 publications
(6 citation statements)
references
References 26 publications
2
4
0
Order By: Relevance
“…The differences between the performance in the different groups that establish the degree of the aphasia severity are quite significant, obtaining up to 2x error on the most severe groups when comparing with mild cases. These big differences between AQ level groups are in line with previous publications [23,38,39], which PER and WER results are summarized in Table 4.…”
Section: Evaluation Results and Discussionsupporting
confidence: 91%
See 4 more Smart Citations
“…The differences between the performance in the different groups that establish the degree of the aphasia severity are quite significant, obtaining up to 2x error on the most severe groups when comparing with mild cases. These big differences between AQ level groups are in line with previous publications [23,38,39], which PER and WER results are summarized in Table 4.…”
Section: Evaluation Results and Discussionsupporting
confidence: 91%
“…Hence, a fair and balanced comparison between systems and technological approaches cannot always be guaranteed. Nonetheless, in some cases, notable improvements can be appreciated between the 52.3 of PER in moderate aphasia test group presented in [25] and the more recent 41.7 of PER reported in [39]. These results seems to be in line with the 38.3 global Syllable Error Rate (SER) reported for the full test set in Cantonese [40], where more than 60% of the test set was composed of mild severity speech data.…”
Section: Related Work In Aphasic Speech Recognitionsupporting
confidence: 75%
See 3 more Smart Citations