Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing 2018
DOI: 10.18653/v1/d18-1431
|View full text |Cite
|
Sign up to set email alerts
|

Generating More Interesting Responses in Neural Conversation Models with Distributional Constraints

Abstract: Neural conversation models tend to generate safe, generic responses for most inputs. This is due to the limitations of likelihoodbased decoding objectives in generation tasks with diverse outputs, such as conversation. To address this challenge, we propose a simple yet effective approach for incorporating side information in the form of distributional constraints over the generated responses. We propose two constraints that help generate more content rich responses that are based on a model of syntax and topic… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

3
80
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
7
2
1

Relationship

0
10

Authors

Journals

citations
Cited by 75 publications
(83 citation statements)
references
References 35 publications
3
80
0
Order By: Relevance
“…The coherence and relevance of a piece of text in a discourse is highly correlated with the perceived quality of the generated text. Previous work has approached generating coherent utterances in conversations through encouraging the model to learn similar distributed representations throughout the conversation (Baheti et al, 2018;Xu et al, 2018;Zhang et al, 2018a). In contrast, we achieve the same goal with a discriminative classifier, which is trained to contrast the true follow-up question (relevant and coherent) against randomly sampled questions (irrelevant) from other conversations and out-of-order questions (uncoherent).…”
Section: Evaluating Question Specificitymentioning
confidence: 99%
“…The coherence and relevance of a piece of text in a discourse is highly correlated with the perceived quality of the generated text. Previous work has approached generating coherent utterances in conversations through encouraging the model to learn similar distributed representations throughout the conversation (Baheti et al, 2018;Xu et al, 2018;Zhang et al, 2018a). In contrast, we achieve the same goal with a discriminative classifier, which is trained to contrast the true follow-up question (relevant and coherent) against randomly sampled questions (irrelevant) from other conversations and out-of-order questions (uncoherent).…”
Section: Evaluating Question Specificitymentioning
confidence: 99%
“…In the dialog domain, we use an LSTM-based sequence-to-sequence (Seq2Seq) model implemented in the OpenNMT framework (Klein et al, 2017). We match the model architecture and training data of Baheti et al (2018). The Seq2Seq model has four layers each in the encoder and decoder, with hidden size 1000, and was trained on a cleaned version of OpenSubtitles (Tiedemann, 2009) to predict the next utterance given the previous one.…”
Section: Open-ended Dialog Taskmentioning
confidence: 99%
“…Such responses understandably bore users, so there has been much research focus on generating more diverse responses (Li et al, 2016a;Xu et al, 2018;Baheti et al, 2018).…”
Section: Diverse Response Generationmentioning
confidence: 99%