2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2019
DOI: 10.1109/cvpr.2019.00497
|View full text |Cite
|
Sign up to set email alerts
|

HetConv: Heterogeneous Kernel-Based Convolutions for Deep CNNs

Abstract: We present a novel deep learning architecture in which the convolution operation leverages heterogeneous kernels. The proposed HetConv (Heterogeneous Kernel-Based Convolution) reduces the computation (FLOPs) and the number of parameters as compared to standard convolution operation while still maintaining representational efficiency. To show the effectiveness of our proposed convolution, we present extensive experimental results on the standard convolutional neural network (CNN) architectures such as VGG [30] … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

3
73
0

Year Published

2019
2019
2021
2021

Publication Types

Select...
5
2
2

Relationship

0
9

Authors

Journals

citations
Cited by 106 publications
(76 citation statements)
references
References 30 publications
(79 reference statements)
3
73
0
Order By: Relevance
“…Pruning methods, such as DSD [15] and ThiNet [31], focus on reducing the redundancy in the model parameters by eliminating the least significant weight or connections in CNNs. Besides, Het-Conv [36] propose to replace the vanilla convolution filters with heterogeneous convolution filters that are in different sizes. However, all of these methods ignore the redundancy on the spatial dimension of feature maps, which is addressed by the proposed OctConv, making OctConv orthogonal and complementary to these previous methods.…”
Section: Related Workmentioning
confidence: 99%
“…Pruning methods, such as DSD [15] and ThiNet [31], focus on reducing the redundancy in the model parameters by eliminating the least significant weight or connections in CNNs. Besides, Het-Conv [36] propose to replace the vanilla convolution filters with heterogeneous convolution filters that are in different sizes. However, all of these methods ignore the redundancy on the spatial dimension of feature maps, which is addressed by the proposed OctConv, making OctConv orthogonal and complementary to these previous methods.…”
Section: Related Workmentioning
confidence: 99%
“…Specifically, the heterogeneous architecture comprises heterogeneous convolutions. Here heterogeneous convolutions with P = 2 [45] are used in the IEEB, where P represents part. The heterogeneous convolutions with P = 2 denote that each standard convolution of 3×3 and each convolution of 1×1 are connected.…”
Section: Network Analysismentioning
confidence: 99%
“…The IEEB uses hierarchical LR features and residual learning techniques to enhance the memory ability of shallow layers for improving the SR performance. By incorporating the heterogeneous architecture proposed in [45] into the IEEB, the amounts of parameters and memory consumption for the IEEB are significantly reduced, so as the training time. Then, the RB fuses the extracted global and local features to transform low-frequency features (i.e., the LR features) into high-frequency features (i.e., the HR features) via residual learning and sub-pixel convolution methods.…”
Section: Introductionmentioning
confidence: 99%
“…The LSTM neural network is a special recurrent neural network (RNN), which introduces a weighted connection with memory and feedback functions. Compared with the feedforward neural network, LSTM can avoid gradient explosion and gradient disappearance, so LSTM can achieve continuous learning for longer time series [42]. The LSTM hidden layer structure is shown in Figure 2.…”
Section: Long Short-term Memorymentioning
confidence: 99%