2018 24th International Conference on Pattern Recognition (ICPR) 2018
DOI: 10.1109/icpr.2018.8545492
|View full text |Cite
|
Sign up to set email alerts
|

Learning an Order Preserving Image Similarity through Deep Ranking

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2019
2019
2021
2021

Publication Types

Select...
4
1

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(3 citation statements)
references
References 15 publications
0
3
0
Order By: Relevance
“…The literature in [23] proposed a Two-channel network, in which the two-channel images (image pairs) are used as the CNN input, and the similarity learning was carried out by extracting the two-channel fusion features. In addition, to better describe image feature information for image similarity learning, variant network structure based on the convolutional neural network has been continuously applied, such as Sim-Net [24], NCC-net [25] and Deep Quadlet network [26].…”
Section: Introductionmentioning
confidence: 99%
“…The literature in [23] proposed a Two-channel network, in which the two-channel images (image pairs) are used as the CNN input, and the similarity learning was carried out by extracting the two-channel fusion features. In addition, to better describe image feature information for image similarity learning, variant network structure based on the convolutional neural network has been continuously applied, such as Sim-Net [24], NCC-net [25] and Deep Quadlet network [26].…”
Section: Introductionmentioning
confidence: 99%
“…the query I q than any negative candidate I n r in feature space, the triplet ranking loss [76,77,78,57,79] as in Eq. (3.6) is adopted for metric learning.…”
Section: Training Pipelinementioning
confidence: 99%
“…For accurate indexing, a positive reference image X p r is expected to be closer to the query X q than any negative reference X n r in the feature space. Thus, the triplet ranking loss [76,78,57,79] in Eq. (4.11) is used to train the image representation f θ , where [y] + = max(y, 0) ensures non-negative output and m is the empirical margin.…”
Section: Loss Functionmentioning
confidence: 99%