2015
DOI: 10.1109/taslp.2015.2422573
|View full text |Cite
|
Sign up to set email alerts
|

Multi-task Learning of Deep Neural Networks for Low-resource Speech Recognition

Abstract: If it is the author's pre-published version, changes introduced as a result of publishing processes such as copy-editing and formatting may not be reflected in this document. For a definitive version of this work, please refer to the published version.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

2
62
0

Year Published

2016
2016
2023
2023

Publication Types

Select...
7
2

Relationship

0
9

Authors

Journals

citations
Cited by 62 publications
(64 citation statements)
references
References 51 publications
2
62
0
Order By: Relevance
“…However, the multilingual training using universal phone set does not show much improvement and in most cases it is even worse. The result is consistent with previous work [3,9,10]. Although the IPA-based multilingual modelling enjoys richer data resources, it has a larger set of units to model as well.…”
Section: Resultssupporting
confidence: 92%
See 1 more Smart Citation
“…However, the multilingual training using universal phone set does not show much improvement and in most cases it is even worse. The result is consistent with previous work [3,9,10]. Although the IPA-based multilingual modelling enjoys richer data resources, it has a larger set of units to model as well.…”
Section: Resultssupporting
confidence: 92%
“…Thus there is sufficient data to train the universal phones. However, it is usually found that the performance of the universal acoustic models is worse than the language-specific acoustic models unless the amount of training data for the target language is really small [9,10]. Although the universal model may share data among various languages, mixture of data creates more variation especially for those identical IPA symbols shared among different languages.…”
Section: Introductionmentioning
confidence: 99%
“…The predominant one being applied to ASR is heterogeneous transfer learning (Wang and Zheng, 2015) which involves training a base model on multiple languages (and tasks) simultaneously. While this achieves some competitive results (Chen and Mak, 2015;Knill et al, 2014), it still requires large amounts of data to yield robust improvements (Heigold et al, 2013).…”
Section: Related Workmentioning
confidence: 99%
“…In the MTL framework, several related tasks are jointly trained with shared hidden layers to improve the generalization power of each task [6]- [10]. In the proposed approach, the main task of VAD is jointly trained with a subsidiary task of feature enhancement.…”
Section: Introductionmentioning
confidence: 99%