Our system is currently under heavy load due to increased usage. We're actively working on upgrades to improve performance. Thank you for your patience.
Proceedings of the 2015 Conference of the North American Chapter of the Association for Computational Linguistics: Human Langua 2015
DOI: 10.3115/v1/n15-1020
|View full text |Cite
|
Sign up to set email alerts
|

A Neural Network Approach to Context-Sensitive Generation of Conversational Responses

Abstract: We present a novel response generation system that can be trained end to end on large quantities of unstructured Twitter conversations. A neural network architecture is used to address sparsity issues that arise when integrating contextual information into classic statistical models, allowing the system to take into account previous dialog utterances. Our dynamic-context generative models show consistent gains over both context-sensitive and non-context-sensitive Machine Translation and Information Retrieval b… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

2
583
0
9

Year Published

2016
2016
2023
2023

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 690 publications
(594 citation statements)
references
References 23 publications
2
583
0
9
Order By: Relevance
“…It can locally produce perfectly sensible language fragments, but it fails to take the meaning of the broader discourse context into account. Much research effort has consequently focused on designing systems able to keep information from the broader context into memory, and possibly even perform simple forms of reasoning about it (Hermann et al, 2015;Hochreiter and Schmidhuber, 1997;Sordoni et al, 2015;Sukhbaatar et al, 2015;Wang and Cho, 2015, a.o.). In this paper, we introduce the LAMBADA dataset (LAnguage Modeling Broadened to Account for Discourse Aspects).…”
Section: Introductionmentioning
confidence: 99%
“…It can locally produce perfectly sensible language fragments, but it fails to take the meaning of the broader discourse context into account. Much research effort has consequently focused on designing systems able to keep information from the broader context into memory, and possibly even perform simple forms of reasoning about it (Hermann et al, 2015;Hochreiter and Schmidhuber, 1997;Sordoni et al, 2015;Sukhbaatar et al, 2015;Wang and Cho, 2015, a.o.). In this paper, we introduce the LAMBADA dataset (LAnguage Modeling Broadened to Account for Discourse Aspects).…”
Section: Introductionmentioning
confidence: 99%
“…On the other hand, many recent works focus on models based on recurrent neural network language models (RNNLM) [20]. Sordoni et al [15] employ an RNN architecture to generate responses from a social media corpus, and Vinyals et al [16] present a long short-term memory (LSTM) neural network encoderdecoders to generate dialog responses using movie subtitles or IT support line chats. More recently Wen et al [17] demonstrate a more advanced LSTM that able to control a response semantically by considering dialogue act feature.…”
Section: Related Workmentioning
confidence: 99%
“…However, in the case that no response in the database could adequately respond to a given utterance, this approach will fail. Response generation [15]- [17] which has the ability to generate a new responses, is arguably robust in handling user input comparing to the other approach, however this approach sometimes generates unnatural responses that are incomprehensible to the user [9]. There have been a number of works on response generation for data-driven dialog systems.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Partially observed Markov descision processes (POMDPs) have been applied to this task (Young et al, 2013), but they typically require handcrafted features. (Sordoni et al, 2015) used a recurrent encoder-decoder model to perform response generation from questions as input, and training the model using two posts as input and the following response as target. (Serban et al, 2016) presented a dialog system built as a hierarchical recurrent LSTM encoder-decoder, where the dialogue is seen as a sequence of utterances, and each utterance is modelled as a sequence of words.…”
Section: User Channelmentioning
confidence: 99%