2020
DOI: 10.1609/aaai.v34i07.6906
|View full text |Cite
|
Sign up to set email alerts
|

Mis-Classified Vector Guided Softmax Loss for Face Recognition

Abstract: Face recognition has witnessed significant progress due to the advances of deep convolutional neural networks (CNNs), the central task of which is how to improve the feature discrimination. To this end, several margin-based (e.g., angular, additive and additive angular margins) softmax loss functions have been proposed to increase the feature margin between different classes. However, despite great achievements have been made, they mainly suffer from three issues: 1) Obviously, they ignore the importance of in… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
93
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
5
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 132 publications
(93 citation statements)
references
References 27 publications
0
93
0
Order By: Relevance
“…It is intuitive that the mis-classified samples contribute more to the improvement of the identity discriminability [16]. Given this, we proposed the additive supervision softmax loss (ASsoftmax) to make full use of the prior knowledge of the misclassified samples.…”
Section: Additive Supervision Softmax Lossmentioning
confidence: 99%
See 1 more Smart Citation
“…It is intuitive that the mis-classified samples contribute more to the improvement of the identity discriminability [16]. Given this, we proposed the additive supervision softmax loss (ASsoftmax) to make full use of the prior knowledge of the misclassified samples.…”
Section: Additive Supervision Softmax Lossmentioning
confidence: 99%
“…Moreover, the end-to-end loss like TE2E [14] and GE2E [15] have been proposed to training the speaker model in an end-to-end fashion. It is worth noting that aforementioned loss functions did not pour much attention on the hard samples, which are beneficial for learning a distinctive representation [16,17].…”
Section: Introductionmentioning
confidence: 99%
“…. , w K } and the feature x of the last fully connected layer are usually normalized and their magnitudes are replaced as a scale parameter s (Wang et al, 2017;Deng et al, 2019;Wang et al, 2019b). In consequence, given an input feature vector x with its ground truth label y, the original softmax loss Eq.…”
Section: Preliminary Knowledgementioning
confidence: 99%
“…where cos(θ w k ,x ) = w T k x is the cosine similarity and θ w k ,x is the angle between w k and x. As pointed out by a great many studies (Liu et al, 2016;Wang et al, 2018b;Deng et al, 2019;Wang et al, 2019b), the learned features with softmax loss are prone to be separable, rather than to be discriminative for face recognition.…”
Section: Preliminary Knowledgementioning
confidence: 99%
See 1 more Smart Citation