Interspeech 2016 2016
DOI: 10.21437/interspeech.2016-1200
|View full text |Cite
|
Sign up to set email alerts
|

Language Model Data Augmentation for Keyword Spotting in Low-Resourced Training Conditions

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
4
0

Year Published

2017
2017
2021
2021

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 11 publications
(4 citation statements)
references
References 21 publications
0
4
0
Order By: Relevance
“…The FCN and CNSA show similar trends across all experimental conditions, therefore we can perhaps infer that both of these classifiers are overfitting the training data. Since they are overfitting, we would normally expect to see improvements on these classifiers by applying web data augmentation ( Lamel et al, 2002 ; Gorin et al, 2016 ). However, we barely gained any benefit from data augmentation, apparently because the in-domain and out-domain datasets had different acoustic feature distributions.…”
Section: Discussionmentioning
confidence: 99%
“…The FCN and CNSA show similar trends across all experimental conditions, therefore we can perhaps infer that both of these classifiers are overfitting the training data. Since they are overfitting, we would normally expect to see improvements on these classifiers by applying web data augmentation ( Lamel et al, 2002 ; Gorin et al, 2016 ). However, we barely gained any benefit from data augmentation, apparently because the in-domain and out-domain datasets had different acoustic feature distributions.…”
Section: Discussionmentioning
confidence: 99%
“…Data augmentation provides a potential solution for data scarcity because it is less time-consuming than collecting and transcribing real speech data and people have shown in many contexts [4,5,6,7,8] that it improves results. Text generation -a data augmentation method -has been proposed in [9,10,11,12,13,14,15] with the aim of improving language models and therefore automatic speech recognition performance for CS speech.…”
Section: Introductionmentioning
confidence: 99%
“…Data augmentation via SMT has been explored in the past for keyword spotting [1] and ASR [8,9]. These studies primarily focus on incorporating raw translation output as a component in the LM.…”
Section: Introductionmentioning
confidence: 99%
“…There has been a growing interest in the area of data augmentation for ASR language modeling. Previous studies include training a recurrent neural network (RNN) based LM on transcriptions and using it to generate synthetic samples for augmentation [1]. SeqGAN, a generative adversarial model for sequences, has been employed for pretraining a code-switched LM [2].…”
Section: Introductionmentioning
confidence: 99%