Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Langua 2016
DOI: 10.18653/v1/n16-1072
|View full text |Cite
|
Sign up to set email alerts
|

Cross-lingual Wikification Using Multilingual Embeddings

Abstract: Cross-lingual Wikification is the task of grounding mentions written in non-English documents to entries in the English Wikipedia. This task involves the problem of comparing textual clues across languages, which requires developing a notion of similarity between text snippets across languages. In this paper, we address this problem by jointly training multilingual embeddings for words and Wikipedia titles. The proposed method can be applied to all languages represented in Wikipedia, including those for which … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
132
0

Year Published

2016
2016
2023
2023

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 97 publications
(133 citation statements)
references
References 17 publications
(20 reference statements)
0
132
0
Order By: Relevance
“…Such embedding models enable us to design NED models that capture the contextual information required to address NED. These models are typically based on conventional word embedding models (e.g., skip-gram (Mikolov et al, 2013)) that assign a fixed embedding to each word and entity (Yamada et al, 2016;Fang et al, 2016;Tsai and Roth, 2016;Cao et al, 2017;Ganea and Hofmann, 2017). In this study, we aim to test the effectiveness of the pretrained contextualized embeddings for NED.…”
Section: Background and Related Workmentioning
confidence: 99%
“…Such embedding models enable us to design NED models that capture the contextual information required to address NED. These models are typically based on conventional word embedding models (e.g., skip-gram (Mikolov et al, 2013)) that assign a fixed embedding to each word and entity (Yamada et al, 2016;Fang et al, 2016;Tsai and Roth, 2016;Cao et al, 2017;Ganea and Hofmann, 2017). In this study, we aim to test the effectiveness of the pretrained contextualized embeddings for NED.…”
Section: Background and Related Workmentioning
confidence: 99%
“…2). SBWES may be used to support many tasks, e.g., computing cross-lingual/multilingual semantic word similarity (Faruqui and Dyer, 2014), learning bilingual word lexicons (Mikolov et al, 2013a;Gouws et al, 2015;, cross-lingual entity linking (Tsai and Roth, 2016), parsing (Guo et al, 2015;Johannsen et al, 2015), machine translation (Zou et al, 2013), or crosslingual information retrieval (Vulić and Moens, 2015;Mitra et al, 2016).…”
Section: Monolingual Vs Bilingualmentioning
confidence: 99%
“…We use the system proposed in Tsai and Roth (2016), which grounds input strings to the intersection of (the title spaces of) the English and the target language Wikipedias. The only requirement is a multilingual Wikipedia dump and it can be applied to all languages in Wikipedia.…”
Section: Cross-lingual Wikifier Featuresmentioning
confidence: 99%
“…The key contribution of this paper is the development of a method that makes use of crosslingual wikification and entity linking (Tsai and Roth, 2016;Moro et al, 2014) to generate language-independent features for NER, and showing how useful this can be for training NER models with no annotation in the target language. Given a mention (sub-string) from a document written in a foreign language, the goal of cross-lingual wikification is to find the cor- Figure 1: An example of a German sentence.…”
Section: Introductionmentioning
confidence: 99%