Proceedings of the 28th International Conference on Computational Linguistics 2020
DOI: 10.18653/v1/2020.coling-main.441
|View full text |Cite
|
Sign up to set email alerts
|

Adversarial Learning on the Latent Space for Diverse Dialog Generation

Abstract: Generating relevant responses in a dialog is challenging, and requires not only proper modeling of context in the conversation, but also being able to generate fluent sentences during inference. In this paper, we propose a two-step framework based on generative adversarial nets for generating conditioned responses. Our model first learns a meaningful representation of sentences by autoencoding, and then learns to map an input query to the response representation, which is in turn decoded as a response sentence… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
6
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
4
3
1

Relationship

1
7

Authors

Journals

citations
Cited by 8 publications
(6 citation statements)
references
References 14 publications
0
6
0
Order By: Relevance
“…1), and the class predicted by the discriminator network, D ( G ( x )) used on Exp2Sim network applied on the experimental data. This loss function is inspired by 41,42 …”
Section: Exp2simgan and Previous Workmentioning
confidence: 99%
“…1), and the class predicted by the discriminator network, D ( G ( x )) used on Exp2Sim network applied on the experimental data. This loss function is inspired by 41,42 …”
Section: Exp2simgan and Previous Workmentioning
confidence: 99%
“…Several studies (Serban et al, 2017a;Zhao et al, 2018;Gao et al, 2019;Cai and Cai, 2022) introduce discrete latent variables to improve the complexity of these distributions. Further studies use more advanced generative models like Generative Adversarial Network (Goodfellow et al, 2020;Gu et al, 2019;Khan et al, 2020) or Normalizing Flows (Rezende and Mohamed, 2015;Luo and Chien, 2021).…”
Section: Related Workmentioning
confidence: 99%
“…where D train is the training data, and each sample x = {x (s) , x (t) }. We also add an auxiliary MSE loss to the objective function as it is found to stabilize GAN training (Khan et al 2020). The overall loss for the GAN is:…”
Section: Training Stage 2: Text-cvaementioning
confidence: 99%
“…We use 128 latent dimensions for the mean and sigma vectors. During training, we use a batch size of 32, learning rate of 1e-4, and Adam optimizer (Kingma and Ba 2015). The sampling temperature is 1.0 for both training and inference.…”
Section: Implementation Detailsmentioning
confidence: 99%