2016
DOI: 10.1016/j.csl.2015.09.002
|View full text |Cite
|
Sign up to set email alerts
|

Preprocessing for elderly speech recognition of smart devices

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2018
2018
2024
2024

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 13 publications
(5 citation statements)
references
References 12 publications
0
5
0
Order By: Relevance
“…These errors may 'propagate' downstream (Errattahi, Hannani, & Ouahmane, 2018). Adapting ASR systems for older voices can help to reduce errors: Zhou et al (2016) found that using a small, domain specific dataset led to fewer errors than using large, out-of-domain data, and Kwon, Kim & Choeh (2016) improved accuracy by preprocessing data in-line with elderly speech patterns. Given that early detection of AD will rely on ASR capabilities in adult voices, as opposed to older, current systems may be appropriate.…”
Section: Automatic Speech Recognitionmentioning
confidence: 99%
“…These errors may 'propagate' downstream (Errattahi, Hannani, & Ouahmane, 2018). Adapting ASR systems for older voices can help to reduce errors: Zhou et al (2016) found that using a small, domain specific dataset led to fewer errors than using large, out-of-domain data, and Kwon, Kim & Choeh (2016) improved accuracy by preprocessing data in-line with elderly speech patterns. Given that early detection of AD will rely on ASR capabilities in adult voices, as opposed to older, current systems may be appropriate.…”
Section: Automatic Speech Recognitionmentioning
confidence: 99%
“…We note that, although high, the word error rate of the elderly participants in the current study is consistent with other studies using automated speech recognition on elderly speech, 42,43 even in controlled laboratory settings. 44 There are reports that pre-processing of elderly speech can decrease the word error rate by up to 12%, 45 although evidence suggests that natural language processing models are relatively impervious to high word error rate. 30 The robust performance of natural language processing models can be attributed to different normalizations of words between a human transcript and an automated speech recognition transcript, trivial word errors that would not change the meaning of a sentence (e.g., "one" vs. "1"), and that natural language processing models are generally trained with a diverse set of language features and are thus able to retain different facets of the language even in the context of a large word error rate.…”
Section: Discussionmentioning
confidence: 99%
“…The diversity of languages, vernaculars, dialects, and people understood and supported by CUIs is an important, yet incredibly difficult challenge. This shouldn't be seen just as a problem of improving speechto-text accuracy for specific populations (for example, adapting to the slower speech and inter-syllabic silence of elderly users [53]), but more widely on the understanding and adapting to how different groups of people speak; their idioms, tropes, and methods for imbuing emotional and social subtlety in language.…”
Section: Breakdowns and Recoverymentioning
confidence: 99%