Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing 2016
DOI: 10.18653/v1/d16-1147
|View full text |Cite
|
Sign up to set email alerts
|

Key-Value Memory Networks for Directly Reading Documents

Abstract: Directly reading documents and being able to answer questions from them is an unsolved challenge. To avoid its inherent difficulty, question answering (QA) has been directed towards using Knowledge Bases (KBs) instead, which has proven effective. Unfortunately KBs often suffer from being too restrictive, as the schema cannot support certain types of answers, and too sparse, e.g. Wikipedia contains much more information than Freebase. In this work we introduce a new method, Key-Value Memory Networks, that makes… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
511
0
2

Year Published

2017
2017
2019
2019

Publication Types

Select...
5
5

Relationship

0
10

Authors

Journals

citations
Cited by 781 publications
(542 citation statements)
references
References 17 publications
1
511
0
2
Order By: Relevance
“…We would also like to investigate alternatives to reinforcement learning for implementing sparse attention, e.g. sparsemax (Martins and Astudillo, 2016) and key-value memory networks (Miller et al, 2016) (preliminary investigations with sparsemax were not extremely promising, but we leave this to future work). Resolving these issues can allow attention models to become more scalable, especially in computationally intensive tasks such as document summarization.…”
Section: Resultsmentioning
confidence: 99%
“…We would also like to investigate alternatives to reinforcement learning for implementing sparse attention, e.g. sparsemax (Martins and Astudillo, 2016) and key-value memory networks (Miller et al, 2016) (preliminary investigations with sparsemax were not extremely promising, but we leave this to future work). Resolving these issues can allow attention models to become more scalable, especially in computationally intensive tasks such as document summarization.…”
Section: Resultsmentioning
confidence: 99%
“…Compared IRNs to Memory Networks (MemNN) Sukhbaatar et al, 2015;Miller et al, 2016) and Neural Turing Machines (NTM) (Graves et al, 2014(Graves et al, , 2016, the biggest difference between our model and the existing frameworks is the controller and the use of the shared memory. We follow Shen et al (2017) for using a controller module to dynamically perform a multi-step inference depending on the complexity of the instance.…”
Section: Related Workmentioning
confidence: 98%
“…Memory is an effective way to equip seq2seq systems with external information (Weston et al, 2014;Sukhbaatar et al, 2015;Miller et al, 2016;Kumar et al, 2015). GenQA (Yin et al, 2015) applies a seq2seq model to generate natural answer sentences from a knowledge base, and CoreQA (He et al, 2017b) extends it with copying mechanism (Gu et al, 2016).…”
Section: Related Workmentioning
confidence: 99%