Proceedings of the 15th Conference of the European Chapter of The Association for Computational Linguistics: Volume 1 2017
DOI: 10.18653/v1/e17-1010
|View full text |Cite
|
Sign up to set email alerts
|

Word Sense Disambiguation: A Unified Evaluation Framework and Empirical Comparison

Abstract: Word Sense Disambiguation is a longstanding task in Natural Language Processing, lying at the core of human language understanding. However, the evaluation of automatic systems has been problematic, mainly due to the lack of a reliable evaluation framework. In this paper we develop a unified evaluation framework and analyze the performance of various Word Sense Disambiguation systems in a fair setup. The results show that supervised systems clearly outperform knowledge-based models. Among the supervised system… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

5
298
0
5

Year Published

2017
2017
2019
2019

Publication Types

Select...
5
3

Relationship

3
5

Authors

Journals

citations
Cited by 249 publications
(317 citation statements)
references
References 34 publications
5
298
0
5
Order By: Relevance
“…It is worth noting that RNN-based architectures outperformed classical supervised approaches (Zhong and Ng, 2010;Iacobacci et al, 2016) when dealing with verbs, which are shown to be highly ambiguous (Raganato et al, 2017). The performance on coarse-grained WSD followed the same trend (Table 2).…”
Section: Resultsmentioning
confidence: 64%
See 2 more Smart Citations
“…It is worth noting that RNN-based architectures outperformed classical supervised approaches (Zhong and Ng, 2010;Iacobacci et al, 2016) when dealing with verbs, which are shown to be highly ambiguous (Raganato et al, 2017). The performance on coarse-grained WSD followed the same trend (Table 2).…”
Section: Resultsmentioning
confidence: 64%
“…Architecture Details. To set a level playing field with comparison systems on English all-words WSD, we followed Raganato et al (2017) and, for all our models, we used a layer of word embeddings pre-trained 8 on the English ukWaC corpus (Baroni et al, 2009) as initialization, and kept them fixed during the training process. For all architectures we then employed 2 layers of bidirectional LSTM with 2048 hidden units (1024 units per direction).…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…Gold standard datasets We performed our evaluations using the framework made available by Raganato et al (2017a) on five different allwords datasets, namely: the Senseval-2 (Edmonds and Cotton, 2001), Senseval-3 (Snyder and Palmer, 2004), SemEval-2007(Pradhan et al, 2007, SemEval-2013 and SemEval-2015 (Moro and WSD datasets. We focused on nouns only, given the fact that Wikipedia provides connections between nominal synsets only, and therefore contributes mainly to syntagmatic relations between nouns.…”
Section: Semantic Networkmentioning
confidence: 99%
“…Given that scaling the manual annotation process becomes practically unfeasible when both lexicographic and encyclopedic knowledge is addressed (Schubert, 2006), recent years have witnessed efforts to produce larger sense-annotated corpora automatically (Moro et al, 2014a;Taghipour and Ng, 2015a;Scozzafava et al, 2015;Raganato et al, 2016). Even though these automatic approaches produce noisier corpora, it has been shown that training on them leads to better supervised and semi-supervised models (Taghipour and Ng, 2015b;Raganato et al, 2016;Yuan et al, 2016;Raganato et al, 2017), as well as to effective embedded representations for senses (Iacobacci et al, 2015;Flekova and Gurevych, 2016).…”
Section: Introductionmentioning
confidence: 99%