2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2019
DOI: 10.1109/cvpr.2019.00065
|View full text |Cite
|
Sign up to set email alerts
|

Bag of Tricks for Image Classification with Convolutional Neural Networks

Abstract: Much of the recent progress made in image classification research can be credited to training procedure refinements, such as changes in data augmentations and optimization methods. In the literature, however, most refinements are either briefly mentioned as implementation details or only visible in source code. In this paper, we will examine a collection of such refinements and empirically evaluate their impact on the final model accuracy through ablation study. We will show that, by combining these refinement… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

6
536
0
4

Year Published

2019
2019
2023
2023

Publication Types

Select...
9
1

Relationship

0
10

Authors

Journals

citations
Cited by 1,102 publications
(546 citation statements)
references
References 28 publications
6
536
0
4
Order By: Relevance
“…One set of techniques that is extremely useful in practice are the tweaks to the ResNet architecture described in [46]. These approaches are used by default in XResNet.…”
Section: Layers and Architecturesmentioning
confidence: 99%
“…One set of techniques that is extremely useful in practice are the tweaks to the ResNet architecture described in [46]. These approaches are used by default in XResNet.…”
Section: Layers and Architecturesmentioning
confidence: 99%
“…While in the shortcut part, in order to make sure that the feature vectors of two parts have the same size, the convolution layer is used to change the dimensionality of input data. We perform batch normalization (BN) [44] right after the convolutions to prevent the gradient dispersion problem. ReLU [45] is performed after the adding to the shortcut.…”
Section: Block 2: 3d-resnext Spectral-spatial Feature Learningmentioning
confidence: 99%
“…The distortion correction and up-sampling were complementary, as their fusion significantly improved both. As He et al [27] proposes, the learning rate based on cosine attenuation should be used for image learning with high detail requirements. The learning rate was initially set to 10 −2 , and then decreased according to a fixed schedule.…”
Section: The Implementation Processmentioning
confidence: 99%