Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) 2018
DOI: 10.18653/v1/p18-1136
|View full text |Cite
|
Sign up to set email alerts
|

Mem2Seq: Effectively Incorporating Knowledge Bases into End-to-End Task-Oriented Dialog Systems

Abstract: End-to-end task-oriented dialog systems usually suffer from the challenge of incorporating knowledge bases. In this paper, we propose a novel yet simple end-toend differentiable model called memoryto-sequence (Mem2Seq) to address this issue. Mem2Seq is the first neural generative model that combines the multihop attention over memories with the idea of pointer network. We empirically show how Mem2Seq controls each generation step, and how its multi-hop attention mechanism helps in learning correlations between… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

2
277
0

Year Published

2018
2018
2019
2019

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 252 publications
(279 citation statements)
references
References 26 publications
2
277
0
Order By: Relevance
“…The Mem2Seq (Madotto et al, 2018) incorporates structured knowledge into the end-to-end task-oriented dialogue. introduces factmatching and knowledge-diffusion to generate meaningful, diverse and natural responses using structured knowledge triplets.…”
Section: Introductionmentioning
confidence: 99%
“…The Mem2Seq (Madotto et al, 2018) incorporates structured knowledge into the end-to-end task-oriented dialogue. introduces factmatching and knowledge-diffusion to generate meaningful, diverse and natural responses using structured knowledge triplets.…”
Section: Introductionmentioning
confidence: 99%
“…The hyper-parameter settings are adopted as the best practice settings for each training set following the Madotto's (Madotto et al, 2018) and best experimental results on baselines SEQ2SEQ and Mem2Seq. Detailed models and their settings are as follows:…”
Section: Baselines and Training Setupmentioning
confidence: 99%
“…Shang et al, 2015), show that training a fully data-driven end-to-end model is a promising way to build domain-agnostic dialogue system. Their models mostly try to use the attention mechanism, including memory networks techniques, to fetch the most similar knowledge (Sukhbaatar et al, 2015), then incorporate grounding knowledge into a seq2seq neural model to generate a suitable re-sponse (Madotto et al, 2018).…”
Section: Introductionmentioning
confidence: 99%
“…Moreover, they didn't utilize user intent during modeling. In [16], the authors used a memory-to-sequence model that uses multi-hop attention over memories to help in learning correlations between memories which results in faster trained model with a stable performance. As for using joint learning to support end-to-end dialogue agent the work introduced by [11] showed state-of-the-art results where they used an attention based RNN for the joint learning of intent detection and slot filling.…”
Section: Related Workmentioning
confidence: 99%