2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2014
DOI: 10.1109/icassp.2014.6854676
|View full text |Cite
|
Sign up to set email alerts
|

Multilingual shifting deep bottleneck features for low-resource ASR

Abstract: We propose a framework that enables the acquisition of annotation-heavy resources such as syntactic dependency tree corpora for low-resource languages by importing linguistic annotations from high-quality English resources. We present a large-scale experiment showing that Chinese dependency trees can be induced by using an English parser, a word alignment package, and a large corpus of sentence-aligned bilingual text. As a part of the experiment, we evaluate the quality of a Chinese parser trained on the induc… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
5

Citation Types

0
12
0

Year Published

2014
2014
2021
2021

Publication Types

Select...
5
5

Relationship

1
9

Authors

Journals

citations
Cited by 17 publications
(12 citation statements)
references
References 19 publications
0
12
0
Order By: Relevance
“…While MBN features have been shown to be useful in several speech recognition tasks (e.g. [20,21]), learned audio features face the same issue as word embeddings, as humans learn to extract useful features from the audio signal as a result of learning to understand language and not as a separate process. However, the MBN features can still be useful where system performance is more important than cognitive plausibility, for instance in a low resource setting.…”
Section: Introductionmentioning
confidence: 99%
“…While MBN features have been shown to be useful in several speech recognition tasks (e.g. [20,21]), learned audio features face the same issue as word embeddings, as humans learn to extract useful features from the audio signal as a result of learning to understand language and not as a separate process. However, the MBN features can still be useful where system performance is more important than cognitive plausibility, for instance in a low resource setting.…”
Section: Introductionmentioning
confidence: 99%
“…[12] and thoroughly evaluated in [13] using five target lan guages and two language sets for multilingual training. Similar NN topology is used in [14], but the adaptation is not properly described and the evaluation is done on one language only.…”
Section: Introductionmentioning
confidence: 99%
“…Another way is to build bottleneck feature (BNF) extractor. The extracted target language BNFs are then used to train a Gaussian Mixture Model-Hidden Markov Model (GMM-HMM) or Deep Neural Network-Hidden Markov Model (DNN-HMM) recognizer [13], [14], [16], [19], [20], [23].…”
Section: Introductionmentioning
confidence: 99%