Proceedings of the 2019 Conference of the North 2019
DOI: 10.18653/v1/n19-1188
|View full text |Cite
|
Sign up to set email alerts
|

Learning Unsupervised Multilingual Word Embeddings with Incremental Multilingual Hubs

Abstract: Recent research has discovered that a shared bilingual word embedding space can be induced by projecting monolingual word embedding spaces from two languages using a selflearning paradigm without any bilingual supervision. However, it has also been shown that for distant language pairs such fully unsupervised self-learning methods are unstable and often get stuck in poor local optima due to reduced isomorphism between starting monolingual spaces. In this work, we propose a new robust framework for learning uns… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
17
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
4
2
1

Relationship

1
6

Authors

Journals

citations
Cited by 22 publications
(18 citation statements)
references
References 25 publications
0
17
0
Order By: Relevance
“…Therefore, the follow-up work aimed to improve the robustness of unsupervised CLWE induction by introducing more robust self-learning procedures (Artetxe et al, 2018b;Kementchedjhieva et al, 2018). Besides increased robustness, recent work claims that fully unsupervised projection-based CLWEs can even match or surpass their supervised counterparts Artetxe et al, 2018b;Alvarez-Melis and Jaakkola, 2018;Hoshen and Wolf, 2018;Heyman et al, 2019).…”
Section: Introductionmentioning
confidence: 99%
“…Therefore, the follow-up work aimed to improve the robustness of unsupervised CLWE induction by introducing more robust self-learning procedures (Artetxe et al, 2018b;Kementchedjhieva et al, 2018). Besides increased robustness, recent work claims that fully unsupervised projection-based CLWEs can even match or surpass their supervised counterparts Artetxe et al, 2018b;Alvarez-Melis and Jaakkola, 2018;Hoshen and Wolf, 2018;Heyman et al, 2019).…”
Section: Introductionmentioning
confidence: 99%
“…We solve the above two stages sequentially using known techniques. Our methodology contrasts with the existing unsupervised MWE methods (Alaux et al, 2019;Chen and Cardie, 2018;Heyman et al, 2019), which learn the unsupervised word alignments and the cross-lingual word embedding mappings jointly. Despite its apparent simplicity, we empirically observe that the proposed approach illustrates remarkable generalization ability and robustness.…”
Section: Unsupervised Multilingual Multi-stage Frameworkmentioning
confidence: 99%
“…They obtain the bilingual lexicons using the Gromov-Wasserstein approach (Alvarez-Melis and Jaakkola, 2018) and mapping operators between languages using the RCSLS algorithm (Joulin et al, 2018). Heyman et al (2019) propose to learn the shared multilingual space by incrementally adding languages to it, one in each iteration. Their approach is based on a reformulation of the bilingual self-learning algorithm proposed by Artetxe et al (2018b).…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…Wada et al (2019) instead use a sentence-level neural language model for directly learning multilingual word embeddings and as a result bypassing the need for mapping functions. In the paradigm of aligning pre-trained word embeddings where we focus, Heyman et al (2019) propose a technique that iteratively builds a multilingual space starting from a monolingual space and incrementally incorporating languages to it. Even if this strategy deviates from the traditional TB/MP model, it still preserves the idea of having a pivot language.…”
Section: Related Workmentioning
confidence: 99%