Proceedings of the 22nd Conference on Computational Natural Language Learning 2018
DOI: 10.18653/v1/k18-1033
|View full text |Cite
|
Sign up to set email alerts
|

Learning to Actively Learn Neural Machine Translation

Abstract: Traditional active learning (AL) methods for machine translation (MT) rely on heuristics. However, these heuristics are limited when the characteristics of the MT problem change due to e.g. the language pair or the amount of the initial bitext. In this paper, we present a framework to learn sentence selection strategies for neural MT. We train the AL query strategy using a high-resource language-pair based on AL simulations, and then transfer it to the lowresource language-pair of interest. The learned query s… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
22
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
7
2

Relationship

0
9

Authors

Journals

citations
Cited by 30 publications
(23 citation statements)
references
References 22 publications
0
22
0
Order By: Relevance
“…They assume that the classifier is a convolutional neural network and use expected gradient length (Settles et al, 2008) to choose sentences that contain words with the most label-discriminative embeddings. Besides text classification, AL has been applied to neural models for semantic parsing (Duong et al, 2018), named entity recognition (Shen et al, 2018), and machine translation (Liu et al, 2018).…”
Section: Related Workmentioning
confidence: 99%
“…They assume that the classifier is a convolutional neural network and use expected gradient length (Settles et al, 2008) to choose sentences that contain words with the most label-discriminative embeddings. Besides text classification, AL has been applied to neural models for semantic parsing (Duong et al, 2018), named entity recognition (Shen et al, 2018), and machine translation (Liu et al, 2018).…”
Section: Related Workmentioning
confidence: 99%
“…Peris and Casacuberta (2018) applied attention based acquisition functions for NMT. Liu et al (2018) introduced reinforcement learning to actively train an NMT model. Wang and Neubig (2019) proposed a method to select relevant sentences from other languages to bring performance gains in low resource NMT.…”
Section: Related Workmentioning
confidence: 99%
“…More recent work has explored bandit optimization for scheduling tasks in a multi-task problem (Graves et al, 2017), and reinforcement learning for selecting examples in a co-trained classifier . Finally, Liu et al (2018) apply imitation learning to actively select monolingual training sentences for labeling in NMT, and show that the learned strategy can be transferred to a related language pair.…”
Section: Related Workmentioning
confidence: 99%