Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing 2021
DOI: 10.18653/v1/2021.emnlp-main.252
|View full text |Cite
|
Sign up to set email alerts
|

Re-embedding Difficult Samples via Mutual Information Constrained Semantically Oversampling for Imbalanced Text Classification

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
0
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3
3
1

Relationship

0
7

Authors

Journals

citations
Cited by 7 publications
(4 citation statements)
references
References 25 publications
0
0
0
Order By: Relevance
“…h . This goal can be finished with the Mutual Information Neural Estimator (MINE) (Belghazi et al 2018;Hjelm et al 2019;Tian et al 2021;Mroueh et al 2021). Specifically, we use a discriminator T ω a h with parameters ω a h to maximize the mutual information between V a h and V + h , and the loss function can be defined as follows:…”
Section: Gumbel-attack Expertmentioning
confidence: 99%
“…h . This goal can be finished with the Mutual Information Neural Estimator (MINE) (Belghazi et al 2018;Hjelm et al 2019;Tian et al 2021;Mroueh et al 2021). Specifically, we use a discriminator T ω a h with parameters ω a h to maximize the mutual information between V a h and V + h , and the loss function can be defined as follows:…”
Section: Gumbel-attack Expertmentioning
confidence: 99%
“…However, evaluating on naturally imbalanced data provides evidence of a method's real-world effectiveness. Some recent studies combine both types of evaluation (e.g., Tian et al, 2021;Subramanian et al, 2021;Jang et al, 2021). Many NLP tasks require treating a large, often heterogenous catch-all class that contains all instances that are not of interest to the task, while the remaining (minority) classes are approximately same-sized.…”
Section: Controlled Vs Real-world Class Imbalancementioning
confidence: 99%
“…MISO (Tian et al, 2021) generates new instances by transforming the representations of minority class instances that are located nearby majority class instances. They learn a mapping from minority instance vectors to "disentangled" representations, making use of mutual information estimators (Belghazi et al, 2018) to push these representations away from the majority class and closer to the minority class.…”
Section: Data Augmentationmentioning
confidence: 99%
“…By combining the MI with the k-means algorithm, a clustering-based sampling method, global data distribution weighted synthetic oversampling (GDDSYN) outperforms some other existing methods [24]. Another work on text classification also introduced a successful MI-constrained oversampling mechanism (MISO) that safely and robustly re-embeds challenging samples [135]. The MI-based SMOTE is also widely applied to multiple aspects with various classifiers, including the MI classifier [136], the KNN classifier, and the decision tree classifier [137].…”
Section: Information Theorymentioning
confidence: 99%