Proceedings of the 2019 Conference of the North 2019
DOI: 10.18653/v1/n19-1119
|View full text |Cite
|
Sign up to set email alerts
|

Competence-based Curriculum Learning for Neural Machine Translation

Abstract: Current state-of-the-art NMT systems use large neural networks that are not only slow to train, but also often require many heuristics and optimization tricks, such as specialized learning rate schedules and large batch sizes. This is undesirable as it requires extensive hyperparameter tuning. In this paper, we propose a curriculum learning framework for NMT that reduces training time, reduces the need for specialized heuristics or large batch sizes, and results in overall better performance. Our framework con… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
101
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 198 publications
(139 citation statements)
references
References 15 publications
1
101
0
Order By: Relevance
“…Thanks to the use of a transfer learning approach inspired by curriculum learning [20,10], we are now able to reach stateof-the-art performance. A curriculum learning approach has also been recently proposed with success for machine translation in [21]. At the end, we also address the issue of domain portability for SLU systems [13,22] that can obviously be tackled as a transfer learning problem.…”
Section: Related Workmentioning
confidence: 99%
“…Thanks to the use of a transfer learning approach inspired by curriculum learning [20,10], we are now able to reach stateof-the-art performance. A curriculum learning approach has also been recently proposed with success for machine translation in [21]. At the end, we also address the issue of domain portability for SLU systems [13,22] that can obviously be tackled as a transfer learning problem.…”
Section: Related Workmentioning
confidence: 99%
“…training the model using easy instances first and increasing the difficulty during the training procedure, has been empirically demonstrated in different machine learning problems, e.g. image classification [11,14], machine translation [21,30,60] and answer generation [23].…”
Section: Related Workmentioning
confidence: 99%
“…Sentence length Machine translation [30], language generation [42], reading comprehension [58] Word rarity Machine translation [30,60], language modeling [1] External model confidence Machine translation [60], image classification [14,49], ad-hoc retrieval [9] Supervision signal intensity Facial expression recognition [12], ad-hoc retrieval [9] Noise estimate Speaker identification [34], image classification [5] Human annotation Image classification [45] (through weak supervision)…”
Section: Difficulty Criteria Tasksmentioning
confidence: 99%
See 2 more Smart Citations