Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) 2016
DOI: 10.18653/v1/p16-1014
|View full text |Cite
|
Sign up to set email alerts
|

Pointing the Unknown Words

Abstract: The problem of rare and unknown words is an important issue that can potentially effect the performance of many NLP systems, including traditional count-based and deep learning models. We propose a novel way to deal with the rare and unseen words for the neural network models using attention. Our model uses two softmax layers in order to predict the next word in conditional language models: one predicts the location of a word in the source sentence, and the other predicts a word in the shortlist vocabulary. At… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
337
0

Year Published

2017
2017
2019
2019

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 411 publications
(337 citation statements)
references
References 14 publications
0
337
0
Order By: Relevance
“…We accomplish this by using an attentionbased copying mechanism (Jia and Liang, 2016;Gulcehre et al, 2016;Gu et al, 2016). At every time step, the decoder may either output a token from the training vocabulary or copy a word from the input sentence.…”
Section: Semantic Parsingmentioning
confidence: 99%
“…We accomplish this by using an attentionbased copying mechanism (Jia and Liang, 2016;Gulcehre et al, 2016;Gu et al, 2016). At every time step, the decoder may either output a token from the training vocabulary or copy a word from the input sentence.…”
Section: Semantic Parsingmentioning
confidence: 99%
“…Particularly, we base our model on the attention mechanism of and the pointer-softmax copying mechanism of Gulcehre et al (2016). In question generation, we can condition our encoder on two different sources of information (compared to the single source in neural machine translation (NMT)): a document that the question should be about and an answer that should fit the generated question.…”
Section: Encoder-decoder Model For Question Generationmentioning
confidence: 99%
“…When formulating questions based on documents, it is common to refer to phrases and entities that appear directly in the text. We therefore incorporate into our decoder a mechanism for copying relevant words from D. We use the pointer-softmax formulation (Gulcehre et al, 2016), which has two output layers: the shortlist softmax and the location softmax. The shortlist softmax places a distribution over words in a predefined output vocabulary.…”
Section: Decodermentioning
confidence: 99%
See 1 more Smart Citation
“…Thus, we expect to decisively prohibit excessive generation. Finally, we evaluate the effectiveness of our method on well-studied ABS benchmark data provided by Rush et al (2015), and evaluated in (Chopra et al, 2016;Nallapati et al, 2016b;Kikuchi et al, 2016;Takase et al, 2016;Ayana et al, 2016;Gulcehre et al, 2016).…”
Section: Introductionmentioning
confidence: 99%