2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) 2016
DOI: 10.1109/cvpr.2016.38
|View full text |Cite
|
Sign up to set email alerts
|

TI-POOLING: Transformation-Invariant Pooling for Feature Learning in Convolutional Neural Networks

Abstract: In this paper we present a deep neural network topology that incorporates a simple to implement transformationinvariant pooling operator (TI-POOLING). This operator is able to efficiently handle prior knowledge on nuisance variations in the data, such as rotation or scale changes. Most current methods usually make use of dataset augmentation to address this issue, but this requires larger number of model parameters and more training data, and results in significantly increased training time and larger chance o… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
201
0
5

Year Published

2017
2017
2019
2019

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 229 publications
(206 citation statements)
references
References 23 publications
0
201
0
5
Order By: Relevance
“…Despite the effectiveness of data augmentation, the main drawback is that learning all possible transformations usually requires a large number of network parameters, which significantly increases the training cost and the risk of overfitting. Most recently, TI-Pooling [18] alleviates the drawback by using parallel network architectures for the transformation set and applying the transformation invariant pooling operator on the outputs before the top layer. Nevertheless, with a builtin data augmentation, TI-Pooling requires significantly more training and testing computational cost than a standard CNN.…”
Section: B Learning Feature Representationsmentioning
confidence: 99%
See 1 more Smart Citation
“…Despite the effectiveness of data augmentation, the main drawback is that learning all possible transformations usually requires a large number of network parameters, which significantly increases the training cost and the risk of overfitting. Most recently, TI-Pooling [18] alleviates the drawback by using parallel network architectures for the transformation set and applying the transformation invariant pooling operator on the outputs before the top layer. Nevertheless, with a builtin data augmentation, TI-Pooling requires significantly more training and testing computational cost than a standard CNN.…”
Section: B Learning Feature Representationsmentioning
confidence: 99%
“…The learning weight decay is set as 0.00005, and the learning rate is reduced to half per 25 epochs. The state-of-the-art STN [19], TI-Pooling [18], ResNet [13] and ORNs [2] are exploited for comparison. Among them, STN is more robust to spatial transformation than the baseline CNNs, due to a spatial transform layer prior to the first convolution layer.…”
Section: A Mnistmentioning
confidence: 99%
“…This can be as simple as averaging the features (Anselmi et al 2016), max-pooling (Laptev et al 2016;Cohen and Welling 2016) or simply by exploiting the group symmetry [such as ignoring the gradient 'sign' in Dalal and Triggs (2005) for vertical flip invariance].…”
Section: Related Workmentioning
confidence: 99%
“…Data augmentation [6] is the most popular technique to mitigate the effects from rotations. Despite the simplicity, it often leads to larger amount of model parameters and is prone to under-or over-fitting [7]. Another drawback of data augmentation is its black-box nature, where it is completely unknown on how the network handles various transformation.…”
Section: Introductionmentioning
confidence: 99%