2020
DOI: 10.48550/arxiv.2012.07004
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

C2C-GenDA: Cluster-to-Cluster Generation for Data Augmentation of Slot Filling

Abstract: Slot filling, a fundamental module of spoken language understanding, often suffers from insufficient quantity and diversity of training data. To remedy this, we propose a novel Cluster-to-Cluster generation framework for Data Augmentation (DA), named C2C-GenDA. It enlarges the training set by reconstructing existing utterances into alternative expressions while keeping semantic. Different from previous DA works that reconstruct utterances one by one independently, C2C-GenDA jointly encodes multiple existing ut… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
3
0

Year Published

2021
2021
2021
2021

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(4 citation statements)
references
References 34 publications
0
3
0
Order By: Relevance
“…We use the original train / validation / test splits provided with each dataset. For Restaurants8k, we randomly split the training set into training (80%) and SNIPS-3 PROTODA 0.881 GCN-RL 0.822 GCN+RL 0.926 (Hou et al, 2020b) and SC-GPT (Peng et al, 2020b) on few-shot intent detection. We allow our learners to train for 5000 iterations.…”
Section: Methodsmentioning
confidence: 99%
See 2 more Smart Citations

Generative Conversational Networks

Papangelis,
Gopalakrishnan,
Padmakumar
et al. 2021
Preprint
“…We use the original train / validation / test splits provided with each dataset. For Restaurants8k, we randomly split the training set into training (80%) and SNIPS-3 PROTODA 0.881 GCN-RL 0.822 GCN+RL 0.926 (Hou et al, 2020b) and SC-GPT (Peng et al, 2020b) on few-shot intent detection. We allow our learners to train for 5000 iterations.…”
Section: Methodsmentioning
confidence: 99%
“…C2C-GenDA (cluster to cluster generation for data augmentation) (Hou et al, 2020b) is a generative data augmentation approach focused on slot filling. This method jointly encodes multiple realisations (i.e.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation

Generative Conversational Networks

Papangelis,
Gopalakrishnan,
Padmakumar
et al. 2021
Preprint
“…In Table 4, we show a comparison with C2C-GenDA (Hou et al, 2020b) and SC-GPT (Peng et al, 2020b) on SNIPS. GCN outperforms C2C-GenDA while SC-GPT performs better than GCN, which is expected since it is based on GPT-2 (instead of distilGPT2) and fine-tuned 400K additional dialogue act -utterance pairs.…”
Section: Intent Detectionmentioning
confidence: 99%