Proceedings of the 15th Conference of the European Chapter of The Association for Computational Linguistics: Volume 1 2017
DOI: 10.18653/v1/e17-1009
|View full text |Cite
|
Sign up to set email alerts
|

Unsupervised Does Not Mean Uninterpretable: The Case for Word Sense Induction and Disambiguation

Abstract: The current trend in NLP is the use of highly opaque models, e.g. neural networks and word embeddings. While these models yield state-of-the-art results on a range of tasks, their drawback is poor interpretability. On the example of word sense induction and disambiguation (WSID), we show that it is possible to develop an interpretable model that matches the state-of-the-art models in accuracy. Namely, we present an unsupervised, knowledge-free WSID approach, which is interpretable at three levels: word sense i… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
22
0

Year Published

2017
2017
2024
2024

Publication Types

Select...
4
4

Relationship

0
8

Authors

Journals

citations
Cited by 26 publications
(22 citation statements)
references
References 36 publications
(59 reference statements)
0
22
0
Order By: Relevance
“…These induced senses do not necessarily correspond to human notions of sense distinctions, or are not easily distinguishable. For this reason, methods have been proposed to improve the interpretability of unsupervised sense representations, either by extracting their hypernyms or their visual representations (i.e., an image illustrating a specific meaning) (Panchenko et al, 2017b) or by mapping the induced senses to external sense inventories (Panchenko, 2016).…”
Section: Interpretabilitymentioning
confidence: 99%
“…These induced senses do not necessarily correspond to human notions of sense distinctions, or are not easily distinguishable. For this reason, methods have been proposed to improve the interpretability of unsupervised sense representations, either by extracting their hypernyms or their visual representations (i.e., an image illustrating a specific meaning) (Panchenko et al, 2017b) or by mapping the induced senses to external sense inventories (Panchenko, 2016).…”
Section: Interpretabilitymentioning
confidence: 99%
“…Babelfy is based on a dense subgraph heuristic for selecting coherent semantic interpretations of the input text. Panchenko et al [49] proposed an unsupervised and knowledge-free word sense induction and disambiguation approach that relies on induced inventories as a pivot for learning sense feature representations.…”
Section: Related Workmentioning
confidence: 99%
“…(3) a combination of cosine and Jaro-Winkler [43] measures is used to calculate the similarity score as follows.…”
Section: Word Sense Disambiguation Of the Conceptsmentioning
confidence: 99%
“…Four criteria are considered for calculating the contribution score: (1) lexical cohesion, (2) co-occurrences cohesion, (3) semantic cohesion and (4) Hierarchical path proximity. Lexical Cohesion: calculates the lexical cohesion of each concept in a message with all the remaining concepts in the same message by measuring Jaro-Winkler [43] similarity between them. This will give us the idea of how much each lexical entity is compatible with the information-content of message.…”
Section: 1 Contribution Score For Semantic Networkmentioning
confidence: 99%