2016
DOI: 10.48550/arxiv.1606.03568
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Word Sense Disambiguation using a Bidirectional LSTM

Abstract: In this paper we present a clean, yet effective, model for word sense disambiguation. Our approach leverage a bidirectional long short-term memory network which is shared between all words. This enables the model to share statistical strength and to scale well with vocabulary size. The model is trained end-to-end, directly from the raw text to sense labels, and makes effective use of word order. We evaluate our approach on two standard datasets, using identical hyperparameter settings, which are in turn tuned … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
12
0

Year Published

2016
2016
2020
2020

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 9 publications
(13 citation statements)
references
References 7 publications
(8 reference statements)
0
12
0
Order By: Relevance
“…Bi-LSTM (Kågebäck and Salomonsson, 2016) is a baseline for neural models. Bi-LSTM +att.+LEX+P OS (Raganato et al, 2017a) is a multi-task learning framework for WSD, POS tagging, and LEX with self-attention mechanism, which converts WSD to a sequence learning task.…”
Section: Resultsmentioning
confidence: 99%
See 2 more Smart Citations
“…Bi-LSTM (Kågebäck and Salomonsson, 2016) is a baseline for neural models. Bi-LSTM +att.+LEX+P OS (Raganato et al, 2017a) is a multi-task learning framework for WSD, POS tagging, and LEX with self-attention mechanism, which converts WSD to a sequence learning task.…”
Section: Resultsmentioning
confidence: 99%
“…Recent neural-based methods are devoted to dealing with this problem. Kågebäck and Salomonsson (2016) present a supervised classifier based on Bi-LSTM, which shares parameters among all word types except the last layer. Raganato et al (2017a) convert WSD task to a sequence labeling task, thus building a unified model for all polysemous words.…”
Section: Traditionalmentioning
confidence: 99%
See 1 more Smart Citation
“…Drop-Tag (Kågebäck and Salomonsson, 2016) replaces token with a < dropped > tag. The tag is subsequently treated just like any other word in the vocabulary and has a corresponding word embedding that is trained.…”
Section: Token Drop Methodsmentioning
confidence: 99%
“…Some advantages of using word embeddings is the lower dimensionality compared to bag-of-words and that words close in meaning are closer in the word em-bedding space. Very recent work still under preprint on using a special kind of recurrent network named LSTM (Long Short Term Memory) for WSD is recently being made available (Yuan et al, 2016) and with bidirectional LSTM (Kågebäck and Salomonsson, 2016), improving over more traditional supervised learning methods.…”
Section: Related Workmentioning
confidence: 99%