Proceedings of the 22nd Conference on Computational Natural Language Learning 2018
DOI: 10.18653/v1/k18-1015
|View full text |Cite
|
Sign up to set email alerts
|

Active Learning for Interactive Neural Machine Translation of Data Streams

Abstract: We study the application of active learning techniques to the translation of unbounded data streams via interactive neural machine translation. The main idea is to select, from an unbounded stream of source sentences, those worth to be supervised by a human agent. The user will interactively translate those samples. Once validated, these data is useful for adapting the neural machine translation model.We propose two novel methods for selecting the samples to be validated. We exploit the information from the at… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
47
1

Year Published

2018
2018
2022
2022

Publication Types

Select...
5
4

Relationship

2
7

Authors

Journals

citations
Cited by 32 publications
(49 citation statements)
references
References 39 publications
1
47
1
Order By: Relevance
“…Considering the -greedy-like strategy of the regulator and the strong role of the cost factor shown in Figure 4, the regulator module does not appear to choose individual actions based e.g., on the difficulty of inputs, but rather composes mini-batches with a feedback ratio according to the feedback type's statistics. This confirms the observations of Peris and Casacuberta (2018), who find that the subset of instances selected for labeling is secondaryit is rather the mixing ratio of feedback types that matters. This finding is also consistent with the mini-batch update regime that forces the regulator to take a higher-level perspective and optimize the expected improvement at the granularity of (minibatch) updates rather than at the input level.…”
Section: Resultssupporting
confidence: 88%
“…Considering the -greedy-like strategy of the regulator and the strong role of the cost factor shown in Figure 4, the regulator module does not appear to choose individual actions based e.g., on the difficulty of inputs, but rather composes mini-batches with a feedback ratio according to the feedback type's statistics. This confirms the observations of Peris and Casacuberta (2018), who find that the subset of instances selected for labeling is secondaryit is rather the mixing ratio of feedback types that matters. This finding is also consistent with the mini-batch update regime that forces the regulator to take a higher-level perspective and optimize the expected improvement at the granularity of (minibatch) updates rather than at the input level.…”
Section: Resultssupporting
confidence: 88%
“…A key component of AL is the choice of the sampling strategy, which curates the samples in order to maximize the model's performance with a minimum amount of user interaction. Many AL sampling strategies have proven effective for human-supervised natural language processing tasks other than compression (Hahn et al, 2012;Peris and Casacuberta, 2018;Liu et al, 2018).…”
Section: Introductionmentioning
confidence: 99%
“…Following the spirit of the Keras library, we developed NMT-Keras, released under MIT license, that aims to provide a highly-modular and extensible framework to NMT. https://github.com/lvapeab/nmt-keras NMT-Keras supports advanced features, including support of interactive-predictive NMT (INMT) (Barrachina et al, 2009;Peris et al, 2017c) protocols, continuous adaptation (Peris et al, 2017a) and active learning (Peris and Casacuberta, 2018b) strategies. An additional goal, is to ease the usage of the library, but allowing the user to configure most of the options involving the NMT process.…”
Section: Introductionmentioning
confidence: 99%