Proceedings of the 20th SIGNLL Conference on Computational Natural Language Learning 2016
DOI: 10.18653/v1/k16-1026
|View full text |Cite
|
Sign up to set email alerts
|

Entity Disambiguation by Knowledge and Text Jointly Embedding

Abstract: For most entity disambiguation systems, the secret recipes are feature representations for mentions and entities, most of which are based on Bag-of-Words (BoW) representations. Commonly, BoW has several drawbacks: (1) It ignores the intrinsic meaning of words/entities; (2) It often results in high-dimension vector spaces and expensive computation; (3) For different applications, methods of designing handcrafted representations may be quite different, lacking of a general guideline. In this paper, we propose a … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
58
0

Year Published

2016
2016
2023
2023

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 66 publications
(58 citation statements)
references
References 11 publications
0
58
0
Order By: Relevance
“…Such embedding models enable us to design NED models that capture the contextual information required to address NED. These models are typically based on conventional word embedding models (e.g., skip-gram (Mikolov et al, 2013)) that assign a fixed embedding to each word and entity (Yamada et al, 2016;Fang et al, 2016;Tsai and Roth, 2016;Cao et al, 2017;Ganea and Hofmann, 2017). In this study, we aim to test the effectiveness of the pretrained contextualized embeddings for NED.…”
Section: Background and Related Workmentioning
confidence: 99%
“…Such embedding models enable us to design NED models that capture the contextual information required to address NED. These models are typically based on conventional word embedding models (e.g., skip-gram (Mikolov et al, 2013)) that assign a fixed embedding to each word and entity (Yamada et al, 2016;Fang et al, 2016;Tsai and Roth, 2016;Cao et al, 2017;Ganea and Hofmann, 2017). In this study, we aim to test the effectiveness of the pretrained contextualized embeddings for NED.…”
Section: Background and Related Workmentioning
confidence: 99%
“…Although the learning of the embeddings might seem straightforward, as it uses the standard skip-gram model, we see this as an advantage. On one hand, it allows our training to scale efficiently to huge vocabulary of words and concepts without the need for a lot of preprocessing (e.g., removing low frequent words and phrases as in Wang et al (2014); Fang et al (2016)). On the other hand, to learn from the knowledge graph contexts, we propose simple adaption to the skip-gram model (cf.…”
Section: Related Workmentioning
confidence: 99%
“…This is a simpler and more computationally efficient function than the scoring function proposed by previous approaches which learn from knowledge graphs (cf. Fang et al (2016)'s equation 1).…”
Section: Related Workmentioning
confidence: 99%
“…We represent entities on three levels: entity, word and character. Our entity-level representation is similar to work on relation extraction (Wang et al, 2014;, entity linking (Yamada et al, 2016;Fang et al, 2016), and entity typing (Yaghoobzadeh and Schütze, 2015). Our word-level representation with distributional word embeddings is similarly used to represent entities for entity linking and relation extraction (Socher et al, 2013;Wang et al, 2014).…”
Section: Related Workmentioning
confidence: 99%