Proceedings of the 18th International Conference on Spoken Language Translation (IWSLT 2021) 2021
DOI: 10.18653/v1/2021.iwslt-1.20
|View full text |Cite
|
Sign up to set email alerts
|

ON-TRAC’ systems for the IWSLT 2021 low-resource speech translation and multilingual speech translation shared tasks

Abstract: This paper describes the ON-TRAC Consortium translation systems developed for two challenge tracks featured in the Evaluation Campaign of IWSLT 2021, low-resource speech translation and multilingual speech translation. The ON-TRAC Consortium is composed of researchers from three French academic laboratories and an industrial partner: LIA (Avignon Université), LIG (Université Grenoble Alpes), LIUM (Le Mans Université), and researchers from Airbus. A pipeline approach was explored for the lowresource speech tran… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
5
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(5 citation statements)
references
References 17 publications
(14 reference statements)
0
5
0
Order By: Relevance
“…Optimal Transport (OT) (Peyré and Cuturi, 2019), traditionally used in NLP and MT (Chen et al, 2019;Alqahtani et al, 2021), has recently found its way into ST. Zhou et al (2023) used OT to find the alignment between speech and text features to apply Mixup. Le et al (2023) applied OT in a siamese pretraining setting in combination with CTC, yielding improvements compared to the standard ASR pretraining. Tsiamas et al (2023) extended this pretraining in the context of foundation models, while also freezing the text branch for better integration with the text decoder during ST finetuning.…”
Section: Optimal Transportmentioning
confidence: 99%
See 4 more Smart Citations
“…Optimal Transport (OT) (Peyré and Cuturi, 2019), traditionally used in NLP and MT (Chen et al, 2019;Alqahtani et al, 2021), has recently found its way into ST. Zhou et al (2023) used OT to find the alignment between speech and text features to apply Mixup. Le et al (2023) applied OT in a siamese pretraining setting in combination with CTC, yielding improvements compared to the standard ASR pretraining. Tsiamas et al (2023) extended this pretraining in the context of foundation models, while also freezing the text branch for better integration with the text decoder during ST finetuning.…”
Section: Optimal Transportmentioning
confidence: 99%
“…To align a speech representation h s ∈ R n ′ ×d to the text representation h x ∈ R m×d , we are minimizing their Wasserstein loss (Frogner et al, 2015) using Optimal Transport (OT) (Peyré and Cuturi, 2019), as in Le et al (2023); Tsiamas et al (2023). We assume two uniform probability distributions ϕ s , ϕ x , with ϕ s i = 1/n ′ and ϕ x j = 1/m, that define the mass of each position in the speech and text representations.…”
Section: Optimal Transportmentioning
confidence: 99%
See 3 more Smart Citations