Findings of the Association for Computational Linguistics: EMNLP 2020 2020
DOI: 10.18653/v1/2020.findings-emnlp.162
|View full text |Cite
|
Sign up to set email alerts
|

Active Learning Approaches to Enhancing Neural Machine Translation

Abstract: Active learning is an efficient approach for mitigating data dependency when training neural machine translation (NMT) models. In this paper we explore new training frameworks by incorporating active learning into various techniques such as transfer learning and iterative back-translation (IBT) under a limited human translation budget. We design a word frequency based acquisition function and combine it with a strong uncertainty based method. The combined method steadily outperforms all other acquisition funct… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 12 publications
(9 citation statements)
references
References 36 publications
(24 reference statements)
0
5
0
Order By: Relevance
“…We develop a word frequency-based technique that considers linguistic factors for the data driven function. It has been found that active NMT training is advantageous for both varieties of acquisition functions [Zhao et al, 2020]. In this way, we can build a non-redundant translated corpus on which NMT can be trained to achieve a better performance than models trained on randomly built corpora.…”
Section: Active Learningmentioning
confidence: 99%
“…We develop a word frequency-based technique that considers linguistic factors for the data driven function. It has been found that active NMT training is advantageous for both varieties of acquisition functions [Zhao et al, 2020]. In this way, we can build a non-redundant translated corpus on which NMT can be trained to achieve a better performance than models trained on randomly built corpora.…”
Section: Active Learningmentioning
confidence: 99%
“…Active Learning (AL) algorithms are built to select the most useful samples for improving the performance of a given model. Therefore, the criteria used by AL algorithms in NLP or MT tasks (Zhang et al, 2017;Zhao et al, 2020;Peris and Casacuberta, 2018;Dou et al, 2020;Zhan et al, 2021) could also serve as discriminative features when it comes to predicting the future NMT performance. We reuse some of the scoring functions introduced by previous work as specified in Section 4.2.…”
Section: Learning Curve and Domain Shift Predictionmentioning
confidence: 99%
“…Active Learning (AL) algorithms are built to select the most useful samples for improving the performance of a given model. Therefore, the criteria used by AL algorithms in NLP or MT tasks (Zhang et al, 2017;Zhao et al, 2020;Peris and Casacuberta, 2018;Dou et al, 2020;Zhan et al, 2021) could also serve as discriminative features when it comes to predicting the future NMT performance. We reuse some of the scoring functions introduced by previous work as specified in Section 4.2.…”
Section: Learning Curve and Domain Shift Predictionmentioning
confidence: 99%