Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing 2017
DOI: 10.18653/v1/d17-1022
|View full text |Cite
|
Sign up to set email alerts
|

Hierarchical Embeddings for Hypernymy Detection and Directionality

Abstract: We present a novel neural model HyperVec to learn hierarchical embeddings for hypernymy detection and directionality. While previous embeddings have shown limitations on prototypical hypernyms, HyperVec represents an unsupervised measure where embeddings are learned in a specific order and capture the hypernym-hyponym distributional hierarchy. Moreover, our model is able to generalize over unseen hypernymy pairs, when using only small sets of training data, and by mapping to other languages. Results on benchma… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
31
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 65 publications
(31 citation statements)
references
References 31 publications
0
31
0
Order By: Relevance
“…The recent trend of learning efficient representations for lexical entailment has moved to supervised learning. In particular, pre-trained word embeddings are retrained to distinguish a hypernymy relation from other relations (Vulić and Mrkšić, 2018;Nguyen et al, 2017;Alsuhaibani et al, 2019). Hierarchical structures defined in taxonomies and ontologies (e.g., WordNet (Miller, 1995)) are commonly used for the retraining (Nguyen et al, 2017;Alsuhaibani et al, 2019 the loss functions.…”
Section: Supervised Representation Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…The recent trend of learning efficient representations for lexical entailment has moved to supervised learning. In particular, pre-trained word embeddings are retrained to distinguish a hypernymy relation from other relations (Vulić and Mrkšić, 2018;Nguyen et al, 2017;Alsuhaibani et al, 2019). Hierarchical structures defined in taxonomies and ontologies (e.g., WordNet (Miller, 1995)) are commonly used for the retraining (Nguyen et al, 2017;Alsuhaibani et al, 2019 the loss functions.…”
Section: Supervised Representation Methodsmentioning
confidence: 99%
“…We used the two thresholds introduced in Section 3.3.3 when evaluating on the WBLESS and BIB-LESS datasets. Following Vulić and Mrkšić (2018) and Nguyen et al (2017), we tuned the thresholds with 2% randomly chosen from the datasets and evaluated our method on the remaining 98%. We repeated this procedure 1,000 times and report the average accuracy.…”
Section: Task Overviewmentioning
confidence: 99%
“…Joint learning models. Many joint learning models introduce semantic lexicons or linguistic structures as additional constraints to the representation learning models (Liu et al 2015;Ono, Miwa, and Sasaki 2015;Nguyen et al 2017). Yu and Dredze (2014) integrate word2vec (Mikolov et al 2013) with synonym constraints that the distances between synonym representations shall be close.…”
Section: Word Representation Specializationmentioning
confidence: 99%
“…Both post-processing and joint learning models require high-quality, manually constructed, semantic lexicons or linguistic structures (Nguyen et al 2017;Liu et al 2018;Mrkšić et al 2016;Glavaš and Vulić 2018). It should be noted that structured data on the web contains rich semantic information, which has the advantage that it can be easily extracted (Zhang et al 2013).…”
Section: Semantic Knowledge Constructionmentioning
confidence: 99%
“…A generalization of that second approach may be more promising. Perhaps different features are involved in a feature comparison process depending upon the particular semantic discrimination required in a particular condition 1 (e.g., Baroni et al, 2012;Lenci and Benotto, 2012;Roller et al, 2014;Weeds et al, 2014;Nguyen et al, 2016Nguyen et al, , 2017aShwartz et al, 2016). More specifically, in the Anomalous condition, as is the case for more traditional feature comparison theories, an overall similarity comparison is sufficient to distinguish true from false stimuli.…”
Section: Introductionmentioning
confidence: 99%