Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics 2020
DOI: 10.18653/v1/2020.acl-main.564
|View full text |Cite
|
Sign up to set email alerts
|

Data Manipulation: Towards Effective Instance Learning for Neural Dialogue Generation via Learning to Augment and Reweight

Abstract: Current state-of-the-art neural dialogue models learn from human conversations following the data-driven paradigm. As such, a reliable training corpus is the crux of building a robust and well-behaved dialogue model. However, due to the open-ended nature of human conversations, the quality of user-generated training data varies greatly, and effective training samples are typically insufficient while noisy samples frequently appear. This impedes the learning of those data-driven neural dialogue models. Therefor… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
39
0
1

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
3
2

Relationship

0
10

Authors

Journals

citations
Cited by 42 publications
(43 citation statements)
references
References 25 publications
(30 reference statements)
0
39
0
1
Order By: Relevance
“…Second, our framework additionally contains a weighting module to reform the generated utterances. Our work is also inspired by Cai et al (2020), which proposes a framework to augment the IND data, while our framework aims to generate OOD data.…”
Section: Weighting Modulementioning
confidence: 99%
“…Second, our framework additionally contains a weighting module to reform the generated utterances. Our work is also inspired by Cai et al (2020), which proposes a framework to augment the IND data, while our framework aims to generate OOD data.…”
Section: Weighting Modulementioning
confidence: 99%
“…Recently, there is an increased interest on applying data augmentation techniques on sentence-level and sentence-pair natural language processing (NLP) tasks, such as text classification (Wei and Zou, 2019;Xie et al, 2019), natural language inference (Min et al, 2020) and machine translation . Augmentation methods explored for these tasks either create augmented instances by manipulating a few words in the original instance, such as word replacement (Zhang et al, 2015;Wang and Yang, 2015;Cai et al, 2020), random deletion (Wei and Zou, 2019), or word position swap (Ş ahin and Steedman, 2018;Min et al, 2020); or create entirely artificial instances via generative models, such as variational auto encoders (Yoo et al, 2019;Mesbah et al, 2019) or back-translation models (Yu et al, 2018;Iyyer et al, 2018).…”
Section: Introductionmentioning
confidence: 99%
“…The backbone of our model is the transformerbased sequence to sequence model (Vaswani et al, 2017), and most hyper-parameters follow Cai et al (2020). Specifically, the encoder and decoder each contains 6 layers.…”
Section: Implementation Detailsmentioning
confidence: 99%