Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conferen 2019
DOI: 10.18653/v1/d19-1198
|View full text |Cite
|
Sign up to set email alerts
|

A Discrete CVAE for Response Generation on Short-Text Conversation

Abstract: Neural conversation models such as encoderdecoder models are easy to generate bland and generic responses. Some researchers propose to use the conditional variational autoencoder (CVAE) which maximizes the lower bound on the conditional log-likelihood on a continuous latent variable. With different sampled latent variables, the model is expected to generate diverse responses. Although the CVAEbased models have shown tremendous potential, their improvement of generating highquality responses is still unsatisfac… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
25
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 33 publications
(25 citation statements)
references
References 24 publications
0
25
0
Order By: Relevance
“…They constructed a classifier to predict the label for the utterance without any labels. Gao et al [39] proposed a discrete CVAE model, which introduces a discrete latent variable with an explicit semantic meaning to improve the general CVAE on dialogue generation task. They proposed a two-stage sampling approach to enable efficient diverse variable selection from a large latent space assumed in the dialogue generation task.…”
Section: Conditional-vaementioning
confidence: 99%
“…They constructed a classifier to predict the label for the utterance without any labels. Gao et al [39] proposed a discrete CVAE model, which introduces a discrete latent variable with an explicit semantic meaning to improve the general CVAE on dialogue generation task. They proposed a two-stage sampling approach to enable efficient diverse variable selection from a large latent space assumed in the dialogue generation task.…”
Section: Conditional-vaementioning
confidence: 99%
“…Gu et al (2018) introduced Gaussian mixture prior network, but it is hard to determine the number of mixtures and the optimization is complicated. Gao et al (2019) assumed the response generation is driven by a single word, and connected each latent variable with words in the vocabulary. Nevertheless, the difficulty is how to target the driving word for a specific post-response pair.…”
Section: Related Workmentioning
confidence: 99%
“…To demonstrate the necessity and effectiveness of our proposed mechanism alone, we build it on Seq2Seq and exclude as many other interferences as possible when comparing with the following state-of-the-art baseline models: S2S During testing phase, we take 3 samplings from the prior network to generate each response. DCVAE (Gao et al, 2019): It is a CVAE-based Seq2Seq model trained with discrete latent variables, where the latent variables are connected with words in the vocabulary. To follow their paper, we use their original implementation and pre-train the model with extracted keywords.…”
Section: Baseline Modelsmentioning
confidence: 99%
See 2 more Smart Citations