2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) 2016
DOI: 10.1109/cvpr.2016.434
|View full text |Cite
|
Sign up to set email alerts
|

Deep Metric Learning via Lifted Structured Feature Embedding

Abstract: Learning the distance metric between pairs of examples is of great importance for learning and visual recognition. With the remarkable success from the state of the art convolutional neural networks, recent works [1, 31] have shown promising results on discriminatively training the networks to learn semantic feature embeddings where similar examples are mapped close to each other and dissimilar examples are mapped farther apart. In this paper, we describe an algorithm for taking full advantage of the training … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

11
1,493
0
1

Year Published

2017
2017
2023
2023

Publication Types

Select...
5
2
1
1

Relationship

0
9

Authors

Journals

citations
Cited by 1,430 publications
(1,544 citation statements)
references
References 28 publications
(72 reference statements)
11
1,493
0
1
Order By: Relevance
“…Recently, there have been a lot of methods to add the verification information to the CNN for face verification task, such as contrastive loss [3], triplet loss [1], Lifted structured embedding [6]. The CNN trained with verification information can adjust the parameters end-to-end, so that the features generated from these CNN have greater discriminant power than those from normal networks which just use the cross entropy loss for objective.…”
Section: The Proposed Loss Functionmentioning
confidence: 99%
See 1 more Smart Citation
“…Recently, there have been a lot of methods to add the verification information to the CNN for face verification task, such as contrastive loss [3], triplet loss [1], Lifted structured embedding [6]. The CNN trained with verification information can adjust the parameters end-to-end, so that the features generated from these CNN have greater discriminant power than those from normal networks which just use the cross entropy loss for objective.…”
Section: The Proposed Loss Functionmentioning
confidence: 99%
“…Triplet loss [1] proposed online and offline methods for selecting training pairs, and each anchor uses semi-hard sample as its corresponding negative sample. Although Lifted structured embedding [6] does not need to pair the samples in a complex method, if the batchsize is N, a high cost O(N 2 ) is entailed. The research community still does not have reasonable ways to pair samples.…”
Section: The Proposed Loss Functionmentioning
confidence: 99%
“…Before each CSS layer, convolutional activations are normalized to have a L 2 norm [43]. To learn the network, we employed the Caltech-101 dataset [13] excluding testing image pairs used in experiments.…”
Section: Experimental Settingsmentioning
confidence: 99%
“…Comparatively, stochastic triple embedding [24] learns the representation in a triplet manner, and joint Bayesian model (JBM) starts from a different assumption to obtain success in face verification task [25]. Magnet loss (MNL) [13] and lifted structured feature embedding [26] are two latest well-designed methods for enhanced local discrimination modeling. In ASV scenarios, NDA has proved to be a strong substitute for LDA [18].…”
Section: Distance Metric Learningmentioning
confidence: 99%