Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics 2019
DOI: 10.18653/v1/p19-1289
|View full text |Cite
|
Sign up to set email alerts
|

STACL: Simultaneous Translation with Implicit Anticipation and Controllable Latency using Prefix-to-Prefix Framework

Abstract: Simultaneous translation, which translates sentences before they are finished, is useful in many scenarios but is notoriously difficult due to word-order differences. While the conventional seq-to-seq framework is only suitable for full-sentence translation, we propose a novel prefix-to-prefix framework for simultaneous translation that implicitly learns to anticipate in a single translation model. Within this framework, we present a very simple yet surprisingly effective "wait-k" policy trained to generate th… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

4
348
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
4
2
1

Relationship

1
6

Authors

Journals

citations
Cited by 143 publications
(352 citation statements)
references
References 16 publications
(32 reference statements)
4
348
0
Order By: Relevance
“…6 Our test-time wait-results are better than those in Ma et al (2019), because the added source side <eos> token helps the full-sentence model to learn when the source sentence higher BLEU scores when latency AL is small, which we think will be the most useful scenarios of simultaneous translation. Furthermore, this figure also shows that our model can achieve good performance on different latency conditions by controlling the threshold , so we do not need to train multiple models for different latency requirements.…”
Section: Performance Comparisonmentioning
confidence: 88%
See 1 more Smart Citation
“…6 Our test-time wait-results are better than those in Ma et al (2019), because the added source side <eos> token helps the full-sentence model to learn when the source sentence higher BLEU scores when latency AL is small, which we think will be the most useful scenarios of simultaneous translation. Furthermore, this figure also shows that our model can achieve good performance on different latency conditions by controlling the threshold , so we do not need to train multiple models for different latency requirements.…”
Section: Performance Comparisonmentioning
confidence: 88%
“…
Simultaneous translation is widely useful but remains challenging. Previous work falls into two main categories: (a) fixed-latency policies such as Ma et al (2019) and (b) adaptive policies such as Gu et al (2017). The former are simple and effective, but have to aggressively predict future content due to diverging source-target word order; the latter do not anticipate, but suffer from unstable and inefficient training.
…”
mentioning
confidence: 99%
“…• The schedule is simple and fixed and can thus be easily integrated into MT training, as typified by wait-k approaches (Dalvi et al, 2018;Ma et al, 2018).…”
Section: Introductionmentioning
confidence: 99%
“…2. We extend the recently-proposed Average Lagging latency metric (Ma et al, 2018), making it differentiable and calculable in expectation, which allows it to be used as a training objective.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation