2020
DOI: 10.48550/arxiv.2012.01768
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Beyond Cats and Dogs: Semi-supervised Classification of fuzzy labels with overclustering

Lars Schmarje,
Johannes Brünger,
Monty Santarossa
et al.

Abstract: A long-standing issue with deep learning is the need for large and consistently labeled datasets. Although the current research in semi-supervised learning can decrease the required amount of annotated data by a factor of 10 or even more, this line of research still uses distinct classes like cats and dogs. However, in the real-world we often encounter problems where different experts have different opinions, thus producing fuzzy labels. We propose a novel framework for handling semi-supervised classifications… Show more

Help me understand this report
View published versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
0
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
2
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(3 citation statements)
references
References 30 publications
(63 reference statements)
0
0
0
Order By: Relevance
“…Schmarje L. et al [23] proposed a semi-supervised classification framework for processing fuzzy labels based on the idea of over-clustering. In this paper, a new loss algorithm using inverse cross-entropy was proposed to maximize mutual information.…”
Section: Label Classification Algorithmmentioning
confidence: 99%
“…Schmarje L. et al [23] proposed a semi-supervised classification framework for processing fuzzy labels based on the idea of over-clustering. In this paper, a new loss algorithm using inverse cross-entropy was proposed to maximize mutual information.…”
Section: Label Classification Algorithmmentioning
confidence: 99%
“…Despite the importance and practical relevance of learning from imprecise data in a variety of settings, so far research has mainly focused on specific tasks (in particular, semi-supervised learning and superset learning), while research on more general representations has been more limited. Furthermore, while several general-purpose algorithmic techniques (including generalized risk minimization [22,30,43,46,49,72], instance-based methods [4,7,51,79,82] or pseudo label-based learning [52,58,80]) have been developed to address these learning tasks, their theoretical and empirical properties have not yet been widely studied [11,15,56,59].…”
Section: Introductionmentioning
confidence: 99%
“…They have attracted interest in both the theoretical and application-oriented literature due to their flexibility as well as for the rich connections with convex analysis [12], optimization [2,35] and statistics [37]. Also in the context of machine learning (ML), credal sets and related models have recently attracted interest as a way to model weak supervision information in a variety of learning settings, including self-supervised learning [22,21], learning from noisy data [23,24], and learning from imprecise data [11,15,22], a general family of settings that encompasses, among others, semi-supervised learning, superset learning [17,26] and fuzzy label learning [10,32]. In all of these settings the idea is to model the weak supervision by means of credal sets, that are assumed to represent the partial or noisy information available to the annotating agent that produced the data: this general framework for studying weakly supervised learning is called credal learning.…”
Section: Introductionmentioning
confidence: 99%