2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2018
DOI: 10.1109/icassp.2018.8461289
|View full text |Cite
|
Sign up to set email alerts
|

Automatic Speech Assessment for Aphasic Patients Based on Syllable-Level Embedding and Supra-Segmental Duration Features

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
18
0

Year Published

2018
2018
2024
2024

Publication Types

Select...
4
2

Relationship

2
4

Authors

Journals

citations
Cited by 16 publications
(22 citation statements)
references
References 6 publications
0
18
0
Order By: Relevance
“…If the overall score is higher than 0.5, the test speaker is classified as High-AQ, otherwise the speaker is classified as Low-AQ. The baseline assessment system in this study follows a conventional twostep assessment approach proposed in our previous study [14]. A 5-dimensional feature vector of supra-segmental duration features is evaluated on the same task of binary classification using a random forest classifier.…”
Section: Speaker-level Classification Accuracymentioning
confidence: 99%
See 1 more Smart Citation
“…If the overall score is higher than 0.5, the test speaker is classified as High-AQ, otherwise the speaker is classified as Low-AQ. The baseline assessment system in this study follows a conventional twostep assessment approach proposed in our previous study [14]. A 5-dimensional feature vector of supra-segmental duration features is evaluated on the same task of binary classification using a random forest classifier.…”
Section: Speaker-level Classification Accuracymentioning
confidence: 99%
“…The timedelay layers stacked with bidirectional long short term memory layers (TDNN-BLSTM) are used as acoustic model of the ASR system and it is trained using multi-task learning strategy [15]. These ASR-generated features were shown to be effective to classify High-AQ speakers from Low-AQ ones in the aspect of acoustic impairment of PWA speech [14]. Table 6 lists the speaker-level binary classification results on 91 test speakers.…”
Section: Speaker-level Classification Accuracymentioning
confidence: 99%
“…In our previous study [5], a standard DNN based ASR for assessment was trained with limited domain-matched healthy speech. MTL provides a potential way to use the domainmismatched datasets to tackle the data scarcity problem.…”
Section: Asr System For Aphasia Assessment 31 Mt-tdnn-blstm Modelmentioning
confidence: 99%
“…In our previous study [5], a framework of fully automatic speech assessment for Cantonese-speaking PWA was developed. A domain-matched ASR system trained with unimpaired speech was used to decode PWA speech into syllable sequences with time alignment information.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation