2022
DOI: 10.48550/arxiv.2206.08704
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Maximum Class Separation as Inductive Bias in One Matrix

Abstract: Maximizing the separation between classes constitutes a well-known inductive bias in machine learning and a pillar of many traditional algorithms. By default, deep networks are not equipped with this inductive bias and therefore many alternative solutions have been proposed through differential optimization. Current approaches tend to optimize classification and separation jointly: aligning inputs with class vectors and separating class vectors angularly. This paper proposes a simple alternative: encoding maxi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 21 publications
0
1
0
Order By: Relevance
“…Since directly using the unhinged loss will lead to volatile effects, which is mainly reflected in the rapid increase of feature norms and the imbalance between class prototypes when training DNNs with the stochastic gradient method, as shown in Figure 19. Inspired by recent works (Zhou et al, 2022c;Kasarla et al, 2022;Yang et al, 2022) that use the Neural Collapse structure as an inductive bias (also called prototyping-anchored learning, PAL), we fix prototypes W as a simplex ETF during training, i.e.,…”
Section: The Unhinged Loss With Prototype-anchored Learningmentioning
confidence: 99%
“…Since directly using the unhinged loss will lead to volatile effects, which is mainly reflected in the rapid increase of feature norms and the imbalance between class prototypes when training DNNs with the stochastic gradient method, as shown in Figure 19. Inspired by recent works (Zhou et al, 2022c;Kasarla et al, 2022;Yang et al, 2022) that use the Neural Collapse structure as an inductive bias (also called prototyping-anchored learning, PAL), we fix prototypes W as a simplex ETF during training, i.e.,…”
Section: The Unhinged Loss With Prototype-anchored Learningmentioning
confidence: 99%