“…Statisticbased methods statistics distribution of dataset to generate more controllable perturbations [2]. Gradient-based perturbation generation is another approach which based on the gradient of model's prediction loss, usually combined with gradient ascent method based on confused classes [8], adjusting method [13], and attacking methods like the Fast Gradient Sign Method (FGSM) and Project Gradient Descent (PGD) [19] [21]. GAN-based methods improve generator and discriminator by adversarial training, and the quality of sample generation is thus enhanced [14].…”
Section: Related Work 21 Adversarial Training In Image Classificationmentioning
confidence: 99%
“…From the perspective of robustness, [19] introduce compound perturbations for adversarial training. At the perspective of accuracy, works like [2,8,17] take random or gradient-based perturbation to enhance the tail classes training. GAN-based method [14] generates new samples to alleviate the learning problem of tail classes.…”
“…Existing works [8,14,19] perturb data on various generation methods, which enhances the randomness of the data and the model learning of the classification boundary.…”
Section: Inofrmation Levelmentioning
confidence: 99%
“…Adversarial training is a new direction to alleviate the side-effect resulted by the long-tailed distribution. It is achieved by perturbing the input data or features [8,14] for model [19], leading to an effect of data augmentation.…”
Section: Introductionmentioning
confidence: 99%
“…The main difference between existing adversarial training methods lies in their ways to generate and add perturbations. Commonlyused perturbation generation methods include stochastic normal distribution [17], statistical information [2], gradient [5,8,13,19,21], Generative Adversarial Networks (GAN) based methods [14]; while methods to introduce perturbations include data perturbation [5,8,13,14,19,21] and feature perturbation [2,17]. It is worth mentioning that existing methods usually make a trade-off between the model performance and robustness.…”
Adversarial training is originated in image classification to address the problem of adversarial attacks, where an invisible perturbation in an image leads to a significant change in model decision. It recently has been observed to be effective in alleviating the long-tailed classification problem, where an imbalanced size of classes makes the model has much lower performance on small classes. However, existing methods typically focus on the methods to generate perturbations for data, while the contributions of different perturbations to long-tailed classification have not been well analyzed. To this end, this paper presents an investigation on the perturbation generation and incorporation components of existing adversarial training methods and proposes a taxonomy that defines these methods using three levels of components, in terms of information, methodology, and optimization. This taxonomy may serve as a design paradigm where an adversarial training algorithm can be created by combining different components in the taxonomy. A comparative study is conducted to verify the influence of each component in long-tailed classification. Experimental results on two benchmarking datasets show that a combination of statistical perturbations and hybrid optimization achieves a promising performance, and the gradientbased method typically improves the performance of both the head and tail classes. More importantly, it is verified that a reasonable combination of the components in our taxonomy may create an algorithm that outperforms the state-of-the-art.
CCS CONCEPTS• Computing methodologies → Machine learning approaches.
“…Statisticbased methods statistics distribution of dataset to generate more controllable perturbations [2]. Gradient-based perturbation generation is another approach which based on the gradient of model's prediction loss, usually combined with gradient ascent method based on confused classes [8], adjusting method [13], and attacking methods like the Fast Gradient Sign Method (FGSM) and Project Gradient Descent (PGD) [19] [21]. GAN-based methods improve generator and discriminator by adversarial training, and the quality of sample generation is thus enhanced [14].…”
Section: Related Work 21 Adversarial Training In Image Classificationmentioning
confidence: 99%
“…From the perspective of robustness, [19] introduce compound perturbations for adversarial training. At the perspective of accuracy, works like [2,8,17] take random or gradient-based perturbation to enhance the tail classes training. GAN-based method [14] generates new samples to alleviate the learning problem of tail classes.…”
“…Existing works [8,14,19] perturb data on various generation methods, which enhances the randomness of the data and the model learning of the classification boundary.…”
Section: Inofrmation Levelmentioning
confidence: 99%
“…Adversarial training is a new direction to alleviate the side-effect resulted by the long-tailed distribution. It is achieved by perturbing the input data or features [8,14] for model [19], leading to an effect of data augmentation.…”
Section: Introductionmentioning
confidence: 99%
“…The main difference between existing adversarial training methods lies in their ways to generate and add perturbations. Commonlyused perturbation generation methods include stochastic normal distribution [17], statistical information [2], gradient [5,8,13,19,21], Generative Adversarial Networks (GAN) based methods [14]; while methods to introduce perturbations include data perturbation [5,8,13,14,19,21] and feature perturbation [2,17]. It is worth mentioning that existing methods usually make a trade-off between the model performance and robustness.…”
Adversarial training is originated in image classification to address the problem of adversarial attacks, where an invisible perturbation in an image leads to a significant change in model decision. It recently has been observed to be effective in alleviating the long-tailed classification problem, where an imbalanced size of classes makes the model has much lower performance on small classes. However, existing methods typically focus on the methods to generate perturbations for data, while the contributions of different perturbations to long-tailed classification have not been well analyzed. To this end, this paper presents an investigation on the perturbation generation and incorporation components of existing adversarial training methods and proposes a taxonomy that defines these methods using three levels of components, in terms of information, methodology, and optimization. This taxonomy may serve as a design paradigm where an adversarial training algorithm can be created by combining different components in the taxonomy. A comparative study is conducted to verify the influence of each component in long-tailed classification. Experimental results on two benchmarking datasets show that a combination of statistical perturbations and hybrid optimization achieves a promising performance, and the gradientbased method typically improves the performance of both the head and tail classes. More importantly, it is verified that a reasonable combination of the components in our taxonomy may create an algorithm that outperforms the state-of-the-art.
CCS CONCEPTS• Computing methodologies → Machine learning approaches.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.