2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2020
DOI: 10.1109/cvpr42600.2020.00642
|View full text |Cite
|
Sign up to set email alerts
|

Cross-Batch Memory for Embedding Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

2
162
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 206 publications
(164 citation statements)
references
References 38 publications
2
162
0
Order By: Relevance
“…Moreover, in those experiments, we use the simpler data augmentation for faster convergence since our focus is analysing the key components, instead of comparing with existing methods. Specifically, we crop a random size of 224×224 [68] is marked with '*' because it exploits information across mini-batch tasks. The '-' denotes the corresponding results are not reported in the original paper.…”
Section: Training and Optimisation Settingsmentioning
confidence: 99%
See 1 more Smart Citation
“…Moreover, in those experiments, we use the simpler data augmentation for faster convergence since our focus is analysing the key components, instead of comparing with existing methods. Specifically, we crop a random size of 224×224 [68] is marked with '*' because it exploits information across mini-batch tasks. The '-' denotes the corresponding results are not reported in the original paper.…”
Section: Training and Optimisation Settingsmentioning
confidence: 99%
“…Additionally, SoftMax norm and SoftTriple [39] are theoretically non-scalable to extremely large dataset because they use multiple proxies to represent one class. XBM [68] exploits extra information across mini-batch tasks. Some other methods, e.g., Margin [71], Divide & Conquer [43], FastAP [4] and MIC [41], use ResNet-50 [13] as the backbone network.…”
Section: Comparison With Recent Baselinesmentioning
confidence: 99%
“…Therefore, by minimizing this loss, we force the network to assign high similarity values on the positive pairs, i.e., the images originating from the same cell, and low similarity values on the negative pairs, i.e., the images that come from different cells. To utilize more negative samples during the loss calculation, we employ a crossbatch memory bank [45,48]. Additionally, to eliminate the bias from the cells that contain many images, in each training epoch, we sample one image pair from each cell.…”
Section: Training Processmentioning
confidence: 99%
“…Recent work in deep metric learning has introduced a number of training objectives with state of the art performance on computer vision tasks (Kim et al, 2020;Wang et al, 2019). Unfortunately, many of these objectives scale linearly with the number K of classes considered due to a costly linear projection onto R K .…”
Section: Scalable Deep Metric Learning Lossesmentioning
confidence: 99%