2021
DOI: 10.48550/arxiv.2110.04908
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Targeted Subset Selection for Limited-data ASR Accent Adaptation

Abstract: We study the task of personalizing ASR models to a target non-native speaker/accent while being constrained by a transcription budget on the duration of utterances selected from a large unlabelled corpus. We propose a subset selection approach using the recently proposed submodular mutual information functions, in which we identify a diverse set of utterances that match the target speaker/accent. This is specified through a few target utterances and achieved by modelling the relationship between the target sub… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
2

Relationship

1
1

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 15 publications
(19 reference statements)
0
2
0
Order By: Relevance
“…Devising strategies for data pruning and constructing optimal subsets is a recent topic of interest in the area of optimization and active learning (Dong et al, 2019;Kaushal et al, 2019;Saadatfar et al, 2020;Durga et al, 2021;Kothawade et al, 2021;Killamsetty et al, 2021;Paul et al, 2021;Kothyari et al, 2021;Ahia et al, 2021). A few studies have examined the training landscape for drawing clues about the optimal subset creation (Toneva et al, 2018;Agarwal et al, 2020;Baldock et al, 2021;Paul et al, 2021;Schirrmeister et al, 2022).…”
Section: Related Workmentioning
confidence: 99%
“…Devising strategies for data pruning and constructing optimal subsets is a recent topic of interest in the area of optimization and active learning (Dong et al, 2019;Kaushal et al, 2019;Saadatfar et al, 2020;Durga et al, 2021;Kothawade et al, 2021;Killamsetty et al, 2021;Paul et al, 2021;Kothyari et al, 2021;Ahia et al, 2021). A few studies have examined the training landscape for drawing clues about the optimal subset creation (Toneva et al, 2018;Agarwal et al, 2020;Baldock et al, 2021;Paul et al, 2021;Schirrmeister et al, 2022).…”
Section: Related Workmentioning
confidence: 99%
“…[20] use the submodular information measures for active learning in the image classification setting to address realistic scenarios like imbalance, redundancy, and out-of-distribution data. Finally, [21] use the submodular information measures for personalized speech recognition. To our knowledge, this is the first work which proposes an active learning framework for object detection capable of handling rare slices of data.…”
Section: 6mentioning
confidence: 99%