2020
DOI: 10.1609/aaai.v34i05.6244
|View full text |Cite
|
Sign up to set email alerts
|

Learning from Easy to Complex: Adaptive Multi-Curricula Learning for Neural Dialogue Generation

Abstract: Current state-of-the-art neural dialogue systems are mainly data-driven and are trained on human-generated responses. However, due to the subjectivity and open-ended nature of human conversations, the complexity of training dialogues varies greatly. The noise and uneven complexity of query-response pairs impede the learning efficiency and effects of the neural dialogue generation models. What is more, so far, there are no unified dialogue complexity measurements, and the dialogue complexity embodies multiple a… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
25
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
5

Relationship

2
8

Authors

Journals

citations
Cited by 25 publications
(25 citation statements)
references
References 16 publications
0
25
0
Order By: Relevance
“…Emotion regulation is an influential factor in many problems that blind people suffer from [28]. Interventions that target emotion regulation strategies would be useful [29].…”
Section: Emotion Regulation Strategymentioning
confidence: 99%
“…Emotion regulation is an influential factor in many problems that blind people suffer from [28]. Interventions that target emotion regulation strategies would be useful [29].…”
Section: Emotion Regulation Strategymentioning
confidence: 99%
“…Shang et al (2018) likewise design a matching network to calibrate the dialogue model training through instance weighting. Cai et al (2020) investigate curriculum learning to adapt the instance effect on dialogue model training according to the sample complexity. Whereas our proposed framework learns to reweight not only the original training examples but also the augmented examples.…”
Section: Related Workmentioning
confidence: 99%
“…At training step t, a batch of training samples is obtained from the top f (t) portions of the entire sorted training samples. Following Platanios et al ( 2019) and Cai et al (2020), we define the function f (t) as:…”
Section: Curriculum Plausibilitymentioning
confidence: 99%