2016
DOI: 10.1609/aaai.v30i1.9883
|View full text |Cite
|
Sign up to set email alerts
|

Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network Models

Abstract: We investigate the task of building open domain, conversational dialogue systems based on large dialogue corpora using generative models. Generative models produce system responses that are autonomously generated word-by-word, opening up the possibility for realistic, flexible interactions. In support of this goal, we extend the recently proposed hierarchical recurrent encoder-decoder neural network to the dialogue domain, and demonstrate that this model is competitive with state-of-the-art neural language mod… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
209
0
1

Year Published

2018
2018
2022
2022

Publication Types

Select...
10

Relationship

0
10

Authors

Journals

citations
Cited by 1,032 publications
(211 citation statements)
references
References 26 publications
1
209
0
1
Order By: Relevance
“…However, these studies, mostly inspired by psychology findings, are either rule-based or limited to small-scale data. Recently, neural models trained on large-scale data have advanced open-domain conversation generation significantly (Ritter, Cherry, and Dolan 2011;Vinyals and Le 2015;Shang, Lu, and Li 2015;Serban et al 2016). Most of these models aim to improve the content quality of conversation generation (Gu et al 2016;Li et al 2016a;Xing et al 2017;Mou et al 2016;Li et al 2016b).…”
Section: Introductionmentioning
confidence: 99%
“…However, these studies, mostly inspired by psychology findings, are either rule-based or limited to small-scale data. Recently, neural models trained on large-scale data have advanced open-domain conversation generation significantly (Ritter, Cherry, and Dolan 2011;Vinyals and Le 2015;Shang, Lu, and Li 2015;Serban et al 2016). Most of these models aim to improve the content quality of conversation generation (Gu et al 2016;Li et al 2016a;Xing et al 2017;Mou et al 2016;Li et al 2016b).…”
Section: Introductionmentioning
confidence: 99%
“…Data-driven conversational models generally fall into two categories: retrieval-based methods (Lowe et al 2015b;2016a;Zhou et al 2016), which select a response from a predefined repository, and generation-based methods (Ritter, Cherry, and Dolan 2011;Serban et al 2016;Vinyals and Le 2015), which employ an encoder-decoder framework where the message is encoded into a vector representation and, then, fed to the decoder to generate the response. The latter is more natural (as it does not require a response repository) yet suffers from generating dull or vague responses and generally needs a great amount of training data.…”
Section: Related Work Conversational Modelsmentioning
confidence: 99%
“…All the proposed models were implemented during the testing, following Li and Jurafsky (2016). We used the bi-directional recurrent neural network with gated recurrent units (Bi-GRU RNN) (Serban et al 2016a) to capture the information along the word sequences. To train the neural conversation models, we followed the hyperparameter settings in (Shang, Lu, and Li 2015;Song et al 2016).…”
Section: Experiments Experimental Setupsmentioning
confidence: 99%