Interspeech 2018 2018
DOI: 10.21437/interspeech.2018-1541
|View full text |Cite
|
Sign up to set email alerts
|

Development of the CUHK Dysarthric Speech Recognition System for the UA Speech Corpus

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

1
41
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
3
3
2

Relationship

1
7

Authors

Journals

citations
Cited by 47 publications
(54 citation statements)
references
References 0 publications
1
41
0
Order By: Relevance
“…Block 1 (B1) and block 3 (B3) were treated as the training set, leaving the block 2 (B2) as the test set. The feed forward DNN based acoustic model without using pitch features is determined as the baseline system, since more advanced forms of neural networks based acoustic models such as time delayed neural networks [39] and long short-term memory recurrent neural networks [40] did not produce lower WER over feed forward DNN on the UASpeech recognition task [23]. All the investigated neural network models on the UASpeech corpus were built in Pytorch [41].…”
Section: Incorporating Pitch Featurementioning
confidence: 99%
See 2 more Smart Citations
“…Block 1 (B1) and block 3 (B3) were treated as the training set, leaving the block 2 (B2) as the test set. The feed forward DNN based acoustic model without using pitch features is determined as the baseline system, since more advanced forms of neural networks based acoustic models such as time delayed neural networks [39] and long short-term memory recurrent neural networks [40] did not produce lower WER over feed forward DNN on the UASpeech recognition task [23]. All the investigated neural network models on the UASpeech corpus were built in Pytorch [41].…”
Section: Incorporating Pitch Featurementioning
confidence: 99%
“…Speech recognition for disordered speech is a challenging task in general [19]. Acoustic features play a major role in early and recent studies of disordered speech recognition [20,21,22,23]. To the best of our knowledge, the vast majority of speech recognition systems using pitch features are conducted on normal speech, very limited research incorporating pitch features has been found for disordered speech.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…In addition to traditional Gaussian mixture model (GMM)hidden Markov model (HMM) systems [6], deep neural networks have been applied to robust acoustic modeling of dysarthric speech recognition systems in recent years. For example, Kim et al [7] captured the distinct characteristics of dysarthric speech by taking advantage of convolutional neural network (CNN) for extracting effective local features and recurrent neural network (RNN) for modeling temporal dependencies of the acoustic features.…”
Section: Introductionmentioning
confidence: 99%
“…For instance, to reduce the influence caused by dysarthric speech variability in tandem ASR systems, bottleneck features have been proposed via DNNs using a large amount out-of-domain data [13], or convolutional neural networks have been employed [14]. Various advanced forms of DNN architecture were tested in [15] for both tandem and hybrid ASR systems for dysarthric speech. To fully exploit DNNbased acoustic modelling, data augmentation was successfully applied based on speed and tempo perturbation in the signal domain for dysarthric speech recognition [16].…”
Section: Introductionmentioning
confidence: 99%