Proceedings of the 28th International Conference on Computational Linguistics 2020
DOI: 10.18653/v1/2020.coling-main.416
|View full text |Cite
|
Sign up to set email alerts
|

Emergent Communication Pretraining for Few-Shot Machine Translation

Abstract: While state-of-the-art models that rely upon massively multilingual pretrained encoders achieve sample efficiency in downstream applications, they still require abundant amounts of unlabelled text. Nevertheless, most of the world's languages lack such resources. Hence, we investigate a more radical form of unsupervised knowledge transfer in the absence of linguistic data. In particular, for the first time we pretrain neural networks via emergent communication from referential games. Our key assumption is that … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

1
16
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
4
2

Relationship

0
6

Authors

Journals

citations
Cited by 7 publications
(19 citation statements)
references
References 41 publications
1
16
0
Order By: Relevance
“…We also establish the non-triviality of such a transfer performance by comparing to other synthetic and natural source corpora, as well as multiple ablation studies on the EC and downstream transfer setups to help understand the transferability of emergent language. Notably, our method of corpus transfer significantly outperforms directly transferring the trained emergent speaker model (Li et al, 2020b), demonstrating that modeling the emergent language could yield greater usefulness than directly transferring the EC agents.…”
Section: Speakermentioning
confidence: 97%
See 4 more Smart Citations
“…We also establish the non-triviality of such a transfer performance by comparing to other synthetic and natural source corpora, as well as multiple ablation studies on the EC and downstream transfer setups to help understand the transferability of emergent language. Notably, our method of corpus transfer significantly outperforms directly transferring the trained emergent speaker model (Li et al, 2020b), demonstrating that modeling the emergent language could yield greater usefulness than directly transferring the EC agents.…”
Section: Speakermentioning
confidence: 97%
“…A typical setup is the image referential game (Figure 1(a)), where a speaker generates a discrete sequence of tokens based on an input image, a listener is challenged to select the input out of distractors based on the message, and both networks are optimized jointly via game success signals. By studying these games, researchers are interested in the emergence of desirable properties resembling natural language, such as game success generalization Lazaridou & Baroni, 2020) and compositionality (Smith et al, 2003;Kirby et al, 2015;Lazaridou et al, 2018;Li et al, 2020b). However, these properties are mostly defined and analyzed within each individual game framework.…”
Section: Speakermentioning
confidence: 99%
See 3 more Smart Citations