Proceedings of the 5th Workshop on Representation Learning for NLP 2020
DOI: 10.18653/v1/2020.repl4nlp-1.6
|View full text |Cite
|
Sign up to set email alerts
|

Learning Geometric Word Meta-Embeddings

Abstract: We propose a geometric framework for learning meta-embeddings of words from different embedding sources. Our framework transforms the embeddings into a common latent space, where, for example, simple averaging or concatenation of different embeddings (of a given word) is more amenable. The proposed latent space arises from two particular geometric transformations-source embedding specific orthogonal rotations and a common Mahalanobis metric scaling. Empirical results on several word similarity and word analogy… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 9 publications
(7 citation statements)
references
References 13 publications
(29 reference statements)
0
7
0
Order By: Relevance
“…With these alignment transformations on monolingual space, one can obtain a better cross-lingual integration of the vector spaces. Recent work has also found that applying orthogonal transformations to source embeddings facilitates averaging (García et al, 2020;Jawanpuria et al, 2020). We expand on this work by incorporating orthogonal transformations in word prisms, which learn word meta-embeddings for specific downstream tasks.…”
Section: Word Meta-embeddingsmentioning
confidence: 95%
“…With these alignment transformations on monolingual space, one can obtain a better cross-lingual integration of the vector spaces. Recent work has also found that applying orthogonal transformations to source embeddings facilitates averaging (García et al, 2020;Jawanpuria et al, 2020). We expand on this work by incorporating orthogonal transformations in word prisms, which learn word meta-embeddings for specific downstream tasks.…”
Section: Word Meta-embeddingsmentioning
confidence: 95%
“…Although averaging does not increase the dimensionality of the ME space as with concatenation, it does not consistently outperform concatenation, especially when the orthogonality condition does not hold. To overcome this problem, Jawanpuria et al [2020] proposed to first learn orthogonal projection matrices for each source embedding space. They measure the Mahalanobis metric between the projected source embeddings, which is a generalisation of the innerproduct that does not assume the dimensions in the vector space to be uncorrelated.…”
Section: Averagingmentioning
confidence: 99%
“…Given independently trained multiple word representations (aka embeddings) learnt using diverse algorithms and lexical resources, word meta-embedding (ME) learning methods [Yin and Schütze, 2016;Bao and Bollegala, 2018;Bollegala et al, 2018a;Wu et al, 2020;He et al, 2020;Jawanpuria et al, 2020;Coates and Bollegala, 2018] attempt to learn more accurate and wide-coverage word embeddings. The input and output word embeddings to the ME algorithm are referred respectively as the source and meta-embeddings.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…This requires minimal bilingual supervision while still leveraging large amounts of monolingual corpora with very competitive results (Artetxe et al, 2016(Artetxe et al, , 2018. These techniques are used by Doval et al (2018); García-Ferrero et al (2020); Jawanpuria et al (2020); He et al (2020) to generate meta-embeddings. This usually involves mapping all the source embeddings to a common vector space followed by averaging.…”
Section: Papermentioning
confidence: 99%