2018 25th IEEE International Conference on Image Processing (ICIP) 2018
DOI: 10.1109/icip.2018.8451123
|View full text |Cite
|
Sign up to set email alerts
|

Online Filter Clustering and Pruning for Efficient Convnets

Abstract: Pruning filters is an effective method for accelerating deep neural networks (DNNs), but most existing approaches prune filters on a pre-trained network directly which limits in acceleration. Although each filter has its own effect in DNNs, but if two filters are same with each other, we could prune one safely. In this paper, we add an extra cluster loss term in the loss function which can force filters in each cluster to be similar online. After training, we keep one filter in each cluster and prune others an… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
9
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
5
3
2

Relationship

0
10

Authors

Journals

citations
Cited by 20 publications
(9 citation statements)
references
References 26 publications
0
9
0
Order By: Relevance
“…In this section, we evaluate our proposed filter pruning method on CIFAR10 and CIFAR100 benchmarks with the single-branch VGGNet-16 and multi-branch ResNet-56 and Resnet-110 networks. We compare our method with the previous methods, such as [6], [9], [12], [13], [27], [31], [32].…”
Section: Methodsmentioning
confidence: 99%
“…In this section, we evaluate our proposed filter pruning method on CIFAR10 and CIFAR100 benchmarks with the single-branch VGGNet-16 and multi-branch ResNet-56 and Resnet-110 networks. We compare our method with the previous methods, such as [6], [9], [12], [13], [27], [31], [32].…”
Section: Methodsmentioning
confidence: 99%
“…The filters with a high level of similarity can be replaced so that even if these filters are removed, the remaining filters can still replace their functions. Zhou et al 14 measured the similarity between filters using clustering and He et al 15 used the geometric median. Duan et al 16 adopted the Pearson correlation coefficient to measure the similarity between output feature maps, then decide which filters to remove based on the similarity information.…”
Section: Related Workmentioning
confidence: 99%
“…Knowledge Distillation Layers: For the proposed method, we select the intermediate features from ResNets [45] and MobileNetV2 [46] Networks with the following spatial sizes [H, W ]: [56, 56], [28,28], [14,14] and [7,7], analyzing L = 4 levels of depth. We assume that both Teacher and Student architectures share the same spatial sizes (in Width and Height, not in Channel dimension) at some points in their architectures.…”
Section: B Implementation Detailsmentioning
confidence: 99%