2020
DOI: 10.1609/aaai.v34i04.6166
|View full text |Cite
|
Sign up to set email alerts
|

Self-Paced Robust Learning for Leveraging Clean Labels in Noisy Data

Abstract: The success of training accurate models strongly depends on the availability of a sufficient collection of precisely labeled data. However, real-world datasets contain erroneously labeled data samples that substantially hinder the performance of machine learning models. Meanwhile, well-labeled data is usually expensive to obtain and only a limited amount is available for training. In this paper, we consider the problem of training a robust model by using large-scale noisy data in conjunction with a small set o… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
4
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
3
3
2

Relationship

0
8

Authors

Journals

citations
Cited by 14 publications
(4 citation statements)
references
References 22 publications
(20 reference statements)
0
4
0
Order By: Relevance
“…SPL is a heuristic that dynamically creates a curriculum based on the losses of the model after each epoch, so as to incorporate the easier samples first. SPL and its variants have been observed to be resilient to noise both in theory (Meng et al, 2016) and practice (Jiang et al, 2018;Zhang et al, 2020), though prior works focus mostly on clean accuracy under unrealizeable label noise; in contrast, we measure targeted misclassification accuracy under adversarially selected, realizeable noise distributions.…”
Section: Related Workmentioning
confidence: 99%
“…SPL is a heuristic that dynamically creates a curriculum based on the losses of the model after each epoch, so as to incorporate the easier samples first. SPL and its variants have been observed to be resilient to noise both in theory (Meng et al, 2016) and practice (Jiang et al, 2018;Zhang et al, 2020), though prior works focus mostly on clean accuracy under unrealizeable label noise; in contrast, we measure targeted misclassification accuracy under adversarially selected, realizeable noise distributions.…”
Section: Related Workmentioning
confidence: 99%
“…Real-world data tend to be massive in quantity, but with quite a few unreliable noisy data that can lead to decreased generalization performance. Many studies have tried to address this, with some degree of success (Wu and Liu 2007;Zhai et al 2020;Zhang et al 2020). However, most of these studies only consider the impact of noisy data on accuracy, rather than on AUC.…”
Section: Introductionmentioning
confidence: 99%
“…Thus, SPL is an effective method for handling noisy data. Many experimental and theoretical analyses have proved its robustness (Meng, Zhao, and Jiang 2017;Liu, Ma, and Meng 2018;Zhang et al 2020). However, existing SPL methods are limited to pointwise learning, while AUC maximization is a pairwise learning problem.…”
Section: Introductionmentioning
confidence: 99%
“…Hence, deep learning suffers from noisy labels that are corrupted from ground-truth labels and are incorrectly labeled. Due to the increasing need to handle noisy-label problems in a massive dataset, learning with noisy labels (LNL) has received much attention in recent years [6,9,10,18,25,[34][35][36]40].…”
Section: Introductionmentioning
confidence: 99%