Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing 2018
DOI: 10.18653/v1/d18-1337
|View full text |Cite
|
Sign up to set email alerts
|

Prediction Improves Simultaneous Neural Machine Translation

Abstract: Simultaneous speech translation aims to maintain translation quality while minimizing the delay between reading input and incrementally producing the output. We propose a new general-purpose prediction action which predicts future words in the input to improve quality and minimize delay in simultaneous translation. We train this agent using reinforcement learning with a novel reward function. Our agent with prediction has better translation quality and less delay compared to an agent-based simultaneous transla… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
48
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
4
4
1

Relationship

0
9

Authors

Journals

citations
Cited by 73 publications
(50 citation statements)
references
References 11 publications
(16 reference statements)
0
48
0
Order By: Relevance
“…Other measures of lagging include Average proportion (AP) [8] and Differentiable Average Lagging (DAL) [18]. AP is unfavorable to short sequences and is incapable of highlighting improvement as it occupies a narrow range [1,18,29]. DAL is a differentiable version of AL used to regularize trainable decoders, and behaves similarly to AL.…”
Section: Training Wait-k Modelsmentioning
confidence: 99%
“…Other measures of lagging include Average proportion (AP) [8] and Differentiable Average Lagging (DAL) [18]. AP is unfavorable to short sequences and is incapable of highlighting improvement as it occupies a narrow range [1,18,29]. DAL is a differentiable version of AL used to regularize trainable decoders, and behaves similarly to AL.…”
Section: Training Wait-k Modelsmentioning
confidence: 99%
“…Knowing that a very-low-latency wait-1 system incurs at best an AP of 0.5 also implies that much of the metric's dynamic range is wasted; in fact, Alinejad et al (2018) report that AP is not sufficiently sensitive to detect their improvements to simultaneous MT.…”
Section: Previous Latency Metricsmentioning
confidence: 99%
“…A separate policy model could avoid these issues. However, previous policy-learning methods either depends on reinforcement learning (RL) (Grissom II et al, 2014;Gu et al, 2017;Alinejad et al, 2018), which makes the training process unstable and inefficient due to exploration, or applies advanced attention mechanisms (Arivazhagan et al, 2019), which requires its training process to be autoregressive, and hence inefficient. Furthermore, each such learned policy cannot change its behaviour according to different latency requirements at testing time, and we will need to train multiple policy models for scenarios with different latency requirements.…”
Section: Introductionmentioning
confidence: 99%