The World Wide Web Conference 2019
DOI: 10.1145/3308558.3313415
|View full text |Cite
|
Sign up to set email alerts
|

Improving Neural Response Diversity with Frequency-Aware Cross-Entropy Loss

Abstract: Sequence-to-Sequence (Seq2Seq) models have achieved encouraging performance on the dialogue response generation task. However, existing Seq2Seq-based response generation methods suffer from a low-diversity problem: they frequently generate generic responses, which make the conversation less interesting. In this paper, we address the low-diversity problem by investigating its connection with model over-confidence reflected in predicted distributions. Specifically, we first analyze the influence of the commonly … Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
66
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
6
2
1

Relationship

1
8

Authors

Journals

citations
Cited by 57 publications
(72 citation statements)
references
References 18 publications
0
66
0
Order By: Relevance
“…However, such methods tend to generate non-informative answers such as “Thank you” and “I have no idea”. To generate more informative answers, researches introduced external knowledge [ 21 ] or adjusted the objective function [ 22 ].…”
Section: Related Workmentioning
confidence: 99%
“…However, such methods tend to generate non-informative answers such as “Thank you” and “I have no idea”. To generate more informative answers, researches introduced external knowledge [ 21 ] or adjusted the objective function [ 22 ].…”
Section: Related Workmentioning
confidence: 99%
“…To discourage the generation of frequently occurring words, [68] proposed a frequency aware cross-entropy function (24) whereby they incorporated a weighting mechanism conditioned on the token frequency whereby frequent words will get lower weights and vice-versa. w i is the weight for y t which is calculated based on frequency in the training set.…”
Section: ) Alternative Loss Function Learningmentioning
confidence: 99%
“…therefore, focused on making the responses diverse [10,11]. Our current work is greatly motivated from these prior works, but our focus is on building a multimodal dialogue system.…”
Section: Plos Onementioning
confidence: 99%
“…In [32], the authors proposed a reinforcement learning-based approach which considers a set of responses jointly and generates multiple diverse responses simultaneously. The authors in [11] propose a Frequency-Aware Cross-Entropy (FACE) loss function for generating diverse responses by incorporating a weighting mechanism conditioned on token frequency. In [33], the authors proposed an easyto-extend learning framework named MEMD (Multi-Encoder to Multi-Decoder), in which an auxiliary encoder and an auxiliary decoder are introduced to provide essential training guidance for generating diverse responses.…”
Section: Unimodal Dialogue Systemsmentioning
confidence: 99%