Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics 2019
DOI: 10.18653/v1/p19-1568
|View full text |Cite
|
Sign up to set email alerts
|

Zero-shot Word Sense Disambiguation using Sense Definition Embeddings

Abstract: Word Sense Disambiguation (WSD) is a longstanding but open problem in Natural Language Processing (NLP). WSD corpora are typically small in size, owing to an expensive annotation process. Current supervised WSD methods treat senses as discrete labels and also resort to predicting the Most-Frequent-Sense (MFS) for words unseen during training. This leads to poor performance on rare and unseen senses. To overcome this challenge, we propose Extended WSD Incorporating Sense Embeddings (EWISE), a supervised model t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

2
103
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 90 publications
(105 citation statements)
references
References 36 publications
(54 reference statements)
2
103
0
Order By: Relevance
“…In addition, there is significant research into strategies for learning neural representations of entities in knowledge bases and coding systems. Past work has investigated diverse approaches, such as leveraging rich semantic information from knowledge base structure and web-scale annotated corpora (34,97,98), utilizing definitions of word senses (similar to our use of ICF definitions) (99,100), and combining terminologies with targeted selection of training corpora to learn applicationtailored concept representations (101,102). While most of the research on entity representations requires resources not yet available for FSI (e.g., large, annotated corpora; well-developed terminologies; robust and interconnected knowledge graph structure), all present significant opportunities to advance FSI coding technologies as more resources are developed.…”
Section: Alternative Coding Approachesmentioning
confidence: 99%
“…In addition, there is significant research into strategies for learning neural representations of entities in knowledge bases and coding systems. Past work has investigated diverse approaches, such as leveraging rich semantic information from knowledge base structure and web-scale annotated corpora (34,97,98), utilizing definitions of word senses (similar to our use of ICF definitions) (99,100), and combining terminologies with targeted selection of training corpora to learn applicationtailored concept representations (101,102). While most of the research on entity representations requires resources not yet available for FSI (e.g., large, annotated corpora; well-developed terminologies; robust and interconnected knowledge graph structure), all present significant opportunities to advance FSI coding technologies as more resources are developed.…”
Section: Alternative Coding Approachesmentioning
confidence: 99%
“…To overcome the aforementioned shortcomings, coarser sense inventories (Lacerra et al 2020) and automatic data augmentation approaches (Pasini and Navigli 2017;Pasini, Elia, and Navigli 2018;Scarlini, Pasini, and Navigli 2019) have been developed to cover more words, senses and languages. At the same time, dedicated architectures have been built to exploit the definitional information of a knowledge base (Luo et al 2018;Kumar et al 2019).…”
Section: Related Workmentioning
confidence: 99%
“…Among knowledge-based approaches, we took into account the extension of Lesk comprising word embeddings (Basile, Caputo, and Semeraro 2014, Lesk ext +emb),the extended version of UKB with gloss relations (Agirre, de Lacalle, and Soroa 2014, UKB gloss ) and Babelfy (Moro, Raganato, and Navigli 2014). As for supervised systems we considered an SVM-based classifier integrated with word embeddings (Iacobacci, Pilehvar, and Navigli 2016, IMS+emb), the Bi-LSTM with attention and multi-task objective presented in Raganato, Delli Bovi, and Navigli, Bi-LSTM (2017), and the more recent supervised systems leveraging sense definitions, i.e., HCAN (Luo et al 2018) and EWISE (Kumar et al 2019). We also performed a comparison with the two LSTM-based architectures of Yuan et al (2016, LSTM-LP) and context2vec (Melamud, Goldberger, and Dagan 2016) for learning representations of the annotated sentences in the training corpus.…”
Section: Wsd Modelmentioning
confidence: 99%
See 1 more Smart Citation
“…Wang [18] extended the word features into sentence-level features and combined them in a further study. Liu [19] and Kumar [20] focused on the sense of words and studied word embedding. Such reverse thinking provided new insights into disambiguation.…”
Section: State Of the Artmentioning
confidence: 99%