2021
DOI: 10.1007/978-981-15-9323-9_24
|View full text |Cite
|
Sign up to set email alerts
|

A Differentiable Generative Adversarial Network for Open Domain Dialogue

Abstract: This work presents a novel methodology to train open domain neural dialogue systems within the framework of Generative Adversarial Networks with gradient based optimization methods. We avoid the non-differentiability related to textgenerating networks approximating the word vector corresponding to each generated token via a top-k softmax. We show that a weighted average of the word vectors of the most probable tokens computed from the probabilities resulting of the top-k softmax leads to a good approximation o… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
2
0
2

Year Published

2021
2021
2023
2023

Publication Types

Select...
4
1

Relationship

1
4

Authors

Journals

citations
Cited by 5 publications
(4 citation statements)
references
References 12 publications
0
2
0
2
Order By: Relevance
“…[7] proposed a soccer dialogue dataset along with a KG-Copy mechanism for non-goal oriented dialogues which are KG-integrated. In a slightly similar research line, in past years, we also notice the use of variational autoencoders (VAE) [40,17] and generative adversarial networks (GANs) [23,18] in dialogue generation. However, knowledge graph based dialogue generation is not well-explored in these approaches.…”
Section: Related Workmentioning
confidence: 82%
“…[7] proposed a soccer dialogue dataset along with a KG-Copy mechanism for non-goal oriented dialogues which are KG-integrated. In a slightly similar research line, in past years, we also notice the use of variational autoencoders (VAE) [40,17] and generative adversarial networks (GANs) [23,18] in dialogue generation. However, knowledge graph based dialogue generation is not well-explored in these approaches.…”
Section: Related Workmentioning
confidence: 82%
“…Decoding details: Neural dialogue systems have been well known to generate too generic and repetitive. This problem has been tackled with many approaches, such as modifying the loss function [52] or using adversarial training [54,63]. Lately, making use of a proper decoding procedure has proved to be essential for generative models to produce good quality non-generic responses [34,49].…”
Section: Decoding In the Short-term Model And Gpt2 Candidate Rerankingmentioning
confidence: 99%
“…Amaitzeko, lan honekin proposatu berri dugun (López Zorrilla et al, 2019) eta testuarekin era guztiz diferentzialean lan egin dezakeen sare sortzaile aurkarien arkitektura baliozkotzen dugu, elkarrizketa sistema automatikoak euskaraz eraikitzeko aproposa dela egiaztatuz.…”
Section: Ondorioakunclassified
“…Ingelesez corpusaren tamaina handiagoa denez, sareak ere handiagoak dira eta iterazio gehiagotan entrenatu dugu. Zehaztasunak(López Zorrilla et al, 2019) erreferentzian ematen dira.…”
unclassified