Proceedings of the 2019 Conference of the North 2019
DOI: 10.18653/v1/n19-1208
|View full text |Cite
|
Sign up to set email alerts
|

Reinforcement Learning based Curriculum Optimization for Neural Machine Translation

Abstract: We consider the problem of making efficient use of heterogeneous training data in neural machine translation (NMT). Specifically, given a training dataset with a sentence-level feature such as noise, we seek an optimal curriculum, or order for presenting examples to the system during training. Our curriculum framework allows examples to appear an arbitrary number of times, and thus generalizes data weighting, filtering, and fine-tuning schemes. Rather than relying on prior knowledge to design a curriculum, we … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
27
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
5
4
1

Relationship

0
10

Authors

Journals

citations
Cited by 51 publications
(32 citation statements)
references
References 15 publications
0
27
0
Order By: Relevance
“…Platanios et al (2019) propose competence-based curriculum learning that select training samples based on sample difficulty and model competence. Kumar et al (2019) use reinforcement learning to learn the curriculum automatically. propose a norm-based curriculum learning method based on the norm of word embedding to improve the efficiency of training an NMT system.…”
Section: Curriculum Learningmentioning
confidence: 99%
“…Platanios et al (2019) propose competence-based curriculum learning that select training samples based on sample difficulty and model competence. Kumar et al (2019) use reinforcement learning to learn the curriculum automatically. propose a norm-based curriculum learning method based on the norm of word embedding to improve the efficiency of training an NMT system.…”
Section: Curriculum Learningmentioning
confidence: 99%
“…Curriculum Learning In recent years, curriculum learning (Bengio et al, 2009), which enables the models to gradually proceed from easy samples to more complex ones in training (Elman, 1993), has received growing research interests in natural language processing field, e.g., neural machine translation (Platanios et al, 2019;Kumar et al, 2019;Zhao et al, 2020;Liu et al, 2020b;Kocmi and Bojar, 2017;Xu et al, 2020) and computer vision field, e.g., image classification (Weinshall et al, 2018), human attribute analysis and visual question answering (Li et al, 2020). For example, in neural machine translation, Platanios et al (2019) proposed to utilize the training samples in order of easy-to-hard and to describe the "difficulty" of a training sample using the sentence length or the rarity of the words appearing in it (Zhao et al, 2020).…”
Section: Image Captioning and Paragraph Generationmentioning
confidence: 99%
“…Neural machine translation model training may combine data selection and model training, taking advantage of the increasing quality of the model to better detect noisy data or to increasingly focus on cleaner parts of the data (Wang et al, 2018;Kumar et al, 2019).…”
Section: Impact Of Noise On Neural Machine Translationmentioning
confidence: 99%