Proceedings of the 2015 Conference of the North American Chapter of the Association for Computational Linguistics: Human Langua 2015
DOI: 10.3115/v1/n15-1028
|View full text |Cite
|
Sign up to set email alerts
|

Deep Multilingual Correlation for Improved Word Embeddings

Abstract: Word embeddings have been found useful for many NLP tasks, including part-of-speech tagging, named entity recognition, and parsing. Adding multilingual context when learning embeddings can improve their quality, for example via canonical correlation analysis (CCA) on embeddings from two languages. In this paper, we extend this idea to learn deep non-linear transformations of word embeddings of the two languages, using the recently proposed deep canonical correlation analysis. The resulting embeddings, when eva… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
89
0

Year Published

2016
2016
2022
2022

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 107 publications
(90 citation statements)
references
References 31 publications
1
89
0
Order By: Relevance
“…In addition to that, we would like to explore non-linear transformations (Lu et al, 2015) and alternative dictionary induction methods Smith et al, 2017). Finally, we would like to apply our model in the decipherment scenario (Dou et al, 2015).…”
Section: Discussionmentioning
confidence: 99%
“…In addition to that, we would like to explore non-linear transformations (Lu et al, 2015) and alternative dictionary induction methods Smith et al, 2017). Finally, we would like to apply our model in the decipherment scenario (Dou et al, 2015).…”
Section: Discussionmentioning
confidence: 99%
“…There is an expansive body of research on learning multilingual word embeddings (Gouws et al, 2014;Faruqui and Dyer, 2014;Lu et al, 2015;Lauly et al, 2014;Luong et al, 2015). Previous work has shown its effectiveness across a wide range of multilingual transfer tasks including tagging (Kim et al, 2015), syntactic parsing (Xiao and Guo, 2014;Guo et al, 2015;Durrett et al, 2012), and machine translation (Zou et al, 2013;Mikolov et al, 2013b).…”
Section: Multilingual Word Embeddingsmentioning
confidence: 99%
“…Most methods rely on supervision encoded in parallel data, at the document level (Vulić and Moens, 2015), the sentence level (Zou et al, 2013;Chandar A P et al, 2014;Hermann and Blunsom, 2014;Kočiský et al, 2014;Luong et al, 2015;Coulmance et al, 2015;Oshikiri et al, 2016), or the word level (i.e. in the form of seed lexicon) (Gouws and Søgaard, 2015;Wick et al, 2016;Duong et al, 2016;Shi et al, 2015;Mikolov et al, 2013a;Faruqui and Dyer, 2014;Lu et al, 2015;Ammar et al, 2016;Zhang et al, 2016aZhang et al, , 2017Smith et al, 2017).…”
Section: Bilingual Lexicon Inductionmentioning
confidence: 99%
“…a linear map, to connect separately trained word embeddings cross-lingually. Learning such a transformation typically calls for cross-lingual supervision from parallel data (Faruqui and Dyer, 2014;Lu et al, 2015;Smith et al, 2017).…”
Section: Introductionmentioning
confidence: 99%