Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing 2021
DOI: 10.18653/v1/2021.emnlp-main.537
|View full text |Cite
|
Sign up to set email alerts
|

It Is Not As Good As You Think! Evaluating Simultaneous Machine Translation on Interpretation Data

Abstract: Most existing simultaneous machine translation (SiMT) systems are trained and evaluated on offline translation corpora. We argue that SiMT systems should be trained and tested on real interpretation data. To illustrate this argument, we propose an interpretation test set and conduct a realistic evaluation of SiMT trained on offline translations. Our results, on our test set along with 3 existing smaller scale language pairs, highlight the difference of up-to 13.83 BLEU score when SiMT models are evaluated on t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(1 citation statement)
references
References 9 publications
0
1
0
Order By: Relevance
“…We use the test portion for evaluation, only for En→X. It has been argued that simultaneous translation is better evaluated (and trained, if possible) on interpreted data (Zhao et al, 2021). However such data is hard to come by, and ESIC is the only such resource for European languages.…”
Section: Datamentioning
confidence: 99%
“…We use the test portion for evaluation, only for En→X. It has been argued that simultaneous translation is better evaluated (and trained, if possible) on interpreted data (Zhao et al, 2021). However such data is hard to come by, and ESIC is the only such resource for European languages.…”
Section: Datamentioning
confidence: 99%