The platform will undergo maintenance on Sep 14 at about 7:45 AM EST and will be unavailable for approximately 2 hours.
Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Confer 2021
DOI: 10.18653/v1/2021.acl-long.200
|View full text |Cite
|
Sign up to set email alerts
|

Beyond Sentence-Level End-to-End Speech Translation: Context Helps

Abstract: Document-level contextual information has shown benefits to text-based machine translation, but whether and how context helps endto-end (E2E) speech translation (ST) is still under-studied. We fill this gap through extensive experiments using a simple concatenationbased context-aware ST model, paired with adaptive feature selection on speech encodings for computational efficiency. We investigate several decoding approaches, and introduce inmodel ensemble decoding which jointly performs document-and sentence-le… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
7
1

Relationship

0
8

Authors

Journals

citations
Cited by 8 publications
(3 citation statements)
references
References 40 publications
0
3
0
Order By: Relevance
“…We focus on improving translation quality of conversations by speaker-turn and cross-talk detection, yet using the context information could also help. In addition, within each MT-MS segment, the inter-utterance context could have already been leveraged (Zhang et al, 2021). We leave analysis of the interand intra-segment context as future work.…”
Section: Discussionmentioning
confidence: 99%
“…We focus on improving translation quality of conversations by speaker-turn and cross-talk detection, yet using the context information could also help. In addition, within each MT-MS segment, the inter-utterance context could have already been leveraged (Zhang et al, 2021). We leave analysis of the interand intra-segment context as future work.…”
Section: Discussionmentioning
confidence: 99%
“…Context-aware ST models have been shown to be robust towards error-prone automatic segmentations of the test set at inference time (Zhang et al, 2021a). Our method bears similarities with Gaido et al (2020b); Papi et al (2021) in that it re-segments the train set to create synthetic data.…”
Section: Relevant Researchmentioning
confidence: 99%
“…With regard to exploiting streaming history, or more generally sentence context, it is worth mentioning the significant amount of previous work in offline MT at sentence level (Tiedemann and Scherrer, 2017;Agrawal et al, 2018), document level (Scherrer et al, 2019;Ma et al, 2020a;Zheng et al, 2020b;Li et al, 2020;Maruf et al, 2021;Zhang et al, 2021), and in related areas such as language modelling (Dai et al, 2019) that has proved to lead to quality gains. Also, as reported in (Li et al, 2020), more robust ST systems can be trained by taking advantage of the context across sentence boundaries using a data augmentation strategy similar to the prefix training methods proposed in (Niehues et al, 2018;Ma et al, 2019).…”
Section: Introductionmentioning
confidence: 99%