Findings of the Association for Computational Linguistics: EMNLP 2020 2020
DOI: 10.18653/v1/2020.findings-emnlp.70
|View full text |Cite
|
Sign up to set email alerts
|

Group-wise Contrastive Learning for Neural Dialogue Generation

Abstract: Neural dialogue response generation has gained much popularity in recent years. Maximum Likelihood Estimation (MLE) objective is widely adopted in existing dialogue model learning. However, models trained with MLE objective function are plagued by the lowdiversity issue when it comes to the opendomain conversational setting. Inspired by the observation that humans not only learn from the positive signals but also benefit from correcting behaviors of undesirable actions, in this work, we introduce contrastive l… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
20
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 22 publications
(20 citation statements)
references
References 34 publications
0
20
0
Order By: Relevance
“…Cai et al [20] introduced contrastive learning into dialogue generation, where the model explicitly perceives the difference between the well-chosen positive and negative utterances.…”
Section: Multi-head Attentionmentioning
confidence: 99%
“…Cai et al [20] introduced contrastive learning into dialogue generation, where the model explicitly perceives the difference between the well-chosen positive and negative utterances.…”
Section: Multi-head Attentionmentioning
confidence: 99%
“…In our case, the multi-turn dialogue data setup allows us to further utilize the context-response relationship, and conduct hard negative sampling by using context-response matching models. Following (Cai et al, 2020), we consider training a Multi-hop Selector Network (MSN) (Yuan et al, 2019) which provides matching scores between the context and response inputs. Specifically, we construct a dialogue dataset, in which each context input c is paired with one positive response sample x, and multiple randomly sample distrator response samples x j .…”
Section: Improved CL With Hard Negative Samplingmentioning
confidence: 99%
“…For auto-regressive model, we consider Dialog-GPT (Zhang et al, 2019), which is a GPT-2 based model that is specifically designed for dialogue response generation. For dialogue response generation model that uses contrastive learning, we include group-wise contrastive learning (GCL) (Cai et al, 2020), which conduct CL between target dialogue model and a pretrained reference model. PLATO (Bao et al, 2019) is another model that uses transformer-based model architecture while including a discrete latent variable to tackle the oneto-many mapping problem.…”
Section: Baseline Modelsmentioning
confidence: 99%
See 1 more Smart Citation
“…Here, the gains are evaluated using reconstruction loss. Finally, inspired from the contrastive learning paradigm (Cai et al, 2020;Chen et al, 2020a,b;Mitrovic et al, 2020), we propose relationship enhancement to increase similarity between the representations of data within the same group, and differentiate the representations of data between different groups.…”
Section: Introductionmentioning
confidence: 99%