2022
DOI: 10.1007/978-3-031-05933-9_11
|View full text |Cite
|
Sign up to set email alerts
|

ADAM: An Attentional Data Augmentation Method for Extreme Multi-label Text Classification

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
6
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(6 citation statements)
references
References 8 publications
0
6
0
Order By: Relevance
“…Furthermore, Xu and Le (2022) propose a feature generation method using a conditional variational autoencoder (VAE). Unfortunately, due to the label co-occurrence, it is challenging for these prior methods to handle MLTC (Wu et al 2020;Zhang et al 2022a). Instead of previous sample-level augmentation, we create a new pair-level augmentation strategy, which merely augments positive feature-label pairs for the tail-labels.…”
Section: Data Augmentationmentioning
confidence: 99%
See 2 more Smart Citations
“…Furthermore, Xu and Le (2022) propose a feature generation method using a conditional variational autoencoder (VAE). Unfortunately, due to the label co-occurrence, it is challenging for these prior methods to handle MLTC (Wu et al 2020;Zhang et al 2022a). Instead of previous sample-level augmentation, we create a new pair-level augmentation strategy, which merely augments positive feature-label pairs for the tail-labels.…”
Section: Data Augmentationmentioning
confidence: 99%
“…One immediate approach to address the problem is data augmentation (DA) which can compensate the scarce data for tail-labels (Wang et al 2019;Chu et al 2020;Zhang et al 2020Zhang et al , 2022a. Meanwhile, DA has shown its effectiveness in many low-resource data scenarios, such as low-resource NLP (Wei and Zou 2019;Wu et al 2022), and zero/few-shot learning (Schwartz et al 2018;Keshari, Singh, and Vatsa 2020;Xu and Le 2022).…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…We first discuss the generation process, then we provide details on how the generative model is encouraged to generate high-quality data. Generation: Following prior works (Zhang et al, 2020), to generate synthetic data we employ GPT-2 (Radford et al, 2019) model. GPT-2 is a transformer-based language model pre-trained on 40 GB of textual data.…”
Section: Data Augmentationmentioning
confidence: 99%
“…One of the barriers to this task is the lack of labeled data. Inspired by the recent advances in the application of pre-trained language models to augment training data for low-resources tasks (Zhang et al, 2020;Yang et al, 2020;Peng et al, 2020;Kumar et al, 2020;Anaby-Tavor et al, 2020), we propose to employ the GPT-2 model to overcome the data scarcity of OSD. To address this limitation, we propose a novel model in which the OSD training data are augmented with the synthetic samples generated by a transformer-based language model.…”
Section: Introductionmentioning
confidence: 99%