Proceedings of the Third Conference on Machine Translation: Research Papers 2018
DOI: 10.18653/v1/w18-6309
|View full text |Cite
|
Sign up to set email alerts
|

A neural interlingua for multilingual machine translation

Abstract: We incorporate an explicit neural interlingua into a multilingual encoder-decoder neural machine translation (NMT) architecture. We demonstrate that our model learns a languageindependent representation by performing direct zero-shot translation (without using pivot translation), and by using the source sentence embeddings to create an English Yelp review classifier that, through the mediation of the neural interlingua, can also classify French and German reviews. Furthermore, we show that, despite using a sma… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

7
87
0

Year Published

2019
2019
2021
2021

Publication Types

Select...
7
2
1

Relationship

0
10

Authors

Journals

citations
Cited by 96 publications
(94 citation statements)
references
References 14 publications
7
87
0
Order By: Relevance
“…In this paper we focus on models that allow the translation between many languages, where we outline the development of a languageindependent representation based on an attention bridge that is shared across all languages. This is in contrast with previous attempts to obtain such a "neural interlingua" (Lu et al, 2018), where the authors have only tested theirs under a one-to-many and many-to-one scenario. In order to do this, we propose an architecture based on shared self-attention for multilingual NMT with language-specific encoders and decoders, that achieves comparable results to the current state-of-the-art architectures and can as well address the task of obtaining language-independent sentence embeddings.…”
Section: Introductionmentioning
confidence: 83%
“…In this paper we focus on models that allow the translation between many languages, where we outline the development of a languageindependent representation based on an attention bridge that is shared across all languages. This is in contrast with previous attempts to obtain such a "neural interlingua" (Lu et al, 2018), where the authors have only tested theirs under a one-to-many and many-to-one scenario. In order to do this, we propose an architecture based on shared self-attention for multilingual NMT with language-specific encoders and decoders, that achieves comparable results to the current state-of-the-art architectures and can as well address the task of obtaining language-independent sentence embeddings.…”
Section: Introductionmentioning
confidence: 83%
“…explore sharing various components in self-attentional (Transformer) models. Lu et al (2018) add a shared "interlingua" layer while using separate encoders and decoders. Zaremoodi et al (2018) utilize recurrent units with multiple blocks together with a trainable routing network.…”
Section: Multilinguality and Zero-shot Performancementioning
confidence: 99%
“…Our pivot adapter (Section 3.2) shares the same motivation with the interlingua component of Lu et al (2018), but is much compact, independent of variable input length, and easy to train offline. The adapter training algorithm is adopted from bilingual word embedding mapping (Xing et al, 2015).…”
Section: Related Workmentioning
confidence: 99%