Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining 2019
DOI: 10.1145/3292500.3330914
|View full text |Cite
|
Sign up to set email alerts
|

Hierarchical Multi-Task Word Embedding Learning for Synonym Prediction

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
22
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 21 publications
(22 citation statements)
references
References 20 publications
0
22
0
Order By: Relevance
“…Our work, distinguishing with them, aims to discover entity synonyms from raw text corpora, which is more practical and challenging for real-world application. Ranking-based and classification-based approaches were proposed to distinguish between candidate entities and the query entity by representing entities with different features, such as co-occurrence statistics [8], textual pattern [6], distributional similarity [9], and semantic types of entities [10], [22]. Recently, Qu et al [7] proposed integrating distributional features and textual patterns to discover the entity synonym with knowledge bases automatically.…”
Section: Entity Synonym Discoverymentioning
confidence: 99%
“…Our work, distinguishing with them, aims to discover entity synonyms from raw text corpora, which is more practical and challenging for real-world application. Ranking-based and classification-based approaches were proposed to distinguish between candidate entities and the query entity by representing entities with different features, such as co-occurrence statistics [8], textual pattern [6], distributional similarity [9], and semantic types of entities [10], [22]. Recently, Qu et al [7] proposed integrating distributional features and textual patterns to discover the entity synonym with knowledge bases automatically.…”
Section: Entity Synonym Discoverymentioning
confidence: 99%
“…Given a pair of entities, our synonym discovery model returns the probability that they are synonymous. We use two types of features for entity pairs 3 : (1) lexical features based on entity surface names (e.g., Jaro-Winkler similarity , token edit distance (Fei et al, 2019), etc), and (2) semantic features based on entity embeddings (e.g., cosine similarity between two entities' SkipGram embeddings). As these feature values have different scales, we use a tree-based boosting model XGBoost (Chen and Guestrin, 2016) to predict whether two entities are synonyms.…”
Section: Proposed Synonym Discovery Modelmentioning
confidence: 99%
“…In comparison, this work aims to develop a method to extract synonym sets directly from raw text corpus. Given a corpus and a term list, one can leverage surface string , co-occurrence statistics (Baroni and Bisi, 2004), textual pattern (Yahya et al, 2014), distributional similarity (Wang et al, 2015), or their combinations (Qu et al, 2017;Fei et al, 2019) to extract synonyms. These methods mostly find synonymous term pairs or a rank list of query entity's synonym, instead of entity synonym sets.…”
Section: Related Workmentioning
confidence: 99%
“…One straightforward approach to obtain synonyms is from pub- Recently, researchers focus on mining synonyms from a raw text corpus, which is more challenging. Two types of approaches are widely exploited, including the pattern based approaches (Nguyen et al, 2017) and the distributional based approaches (Wang et al, 2019a,b;Fei et al, 2019;Zhang et al, 2019). The pattern based approaches lay emphasis on the local contexts, such as "commonly known as".…”
Section: Related Workmentioning
confidence: 99%
“…Two types of approaches are widely exploited to discover synonyms automatically from text corpora, including the distributional based approaches (Wang et al, 2019a,b;Fei et al, 2019) and the pattern based approaches (Nguyen et al, 2017). The distributional based approaches assume that if two terms appear in similar contexts, they are likely to be synonyms.…”
Section: Introductionmentioning
confidence: 99%