2022
DOI: 10.48550/arxiv.2203.07627
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Multilingual Mix: Example Interpolation Improves Multilingual Neural Machine Translation

Abstract: Multilingual neural machine translation models are trained to maximize the likelihood of a mix of examples drawn from multiple language pairs. The dominant inductive bias applied to these models is a shared vocabulary and a shared set of parameters across languages; the inputs and labels corresponding to examples drawn from different language pairs might still reside in distinct subspaces. In this paper, we introduce multilingual crossover encoder-decoder (mXEncDec) to fuse language pairs at an instance level.… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...

Citation Types

0
0
0

Publication Types

Select...

Relationship

0
0

Authors

Journals

citations
Cited by 0 publications
references
References 21 publications
(36 reference statements)
0
0
0
Order By: Relevance

No citations

Set email alert for when this publication receives citations?