2014 IEEE Conference on Computer Vision and Pattern Recognition 2014
DOI: 10.1109/cvpr.2014.180
|View full text |Cite
|
Sign up to set email alerts
|

Learning Fine-Grained Image Similarity with Deep Ranking

Abstract: Abstract

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

2
882
1
1

Year Published

2016
2016
2022
2022

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 1,139 publications
(886 citation statements)
references
References 18 publications
2
882
1
1
Order By: Relevance
“…Previous works employ standard classification [8], pairwise-loss [24] or Triplet-loss [165], [17] CNN models for fine-tuning. The introduction of Faster R-CNN to instance retrieval is a promising starting point towards more accurate object localization [17].…”
Section: Towards Generic Instance Retrievalmentioning
confidence: 99%
“…Previous works employ standard classification [8], pairwise-loss [24] or Triplet-loss [165], [17] CNN models for fine-tuning. The introduction of Faster R-CNN to instance retrieval is a promising starting point towards more accurate object localization [17].…”
Section: Towards Generic Instance Retrievalmentioning
confidence: 99%
“…The first approach consists of submitting pairs of training images (X, Y ) to the network, telling it if they are similar or not [19]. The second approach is to use triplets of images (A, B, C) telling the network that d(f (A), f(B)) should be smaller than d(f (A), f(C)) (where d is a distance function and f the embedding function) [38]. Since we are interested in making a ranking system, the order of proximity is what is important to us and the second approach then better suited.…”
Section: Fine-tuning the Networkmentioning
confidence: 99%
“…For example, triplet loss is widely used in many deep verification [33,29] or ranking [40,46,36,31] networks. It is adopted here to enforce the ranking between a query sketch and a pair of positive and negative photos.…”
Section: Shortcuts and Layer Fusion In Deep Learningmentioning
confidence: 99%
“…It is adopted here to enforce the ranking between a query sketch and a pair of positive and negative photos. In the vast majority of cases [40,46,36,31] Euclidean distancebased, or other first-order energy functions are used in the loss formulation. They are first order in the sense that only element-wise comparisons are made, making it sensitive to feature misalignment and meaning that no cross-feature correlation can be exploited in the similarity.…”
Section: Shortcuts and Layer Fusion In Deep Learningmentioning
confidence: 99%