2019
DOI: 10.48550/arxiv.1911.08630
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

CUP: Cluster Pruning for Compressing Deep Neural Networks

Abstract: We propose Cluster Pruning (CUP) for compressing and accelerating deep neural networks. Our approach prunes similar filters by clustering them based on features derived from both the incoming and outgoing weight connections. With CUP, we overcome two limitations of prior work-(1) nonuniform pruning: CUP can efficiently determine the ideal number of filters to prune in each layer of a neural network. This is in contrast to prior methods that either prune all layers uniformly or otherwise use resource-intensive … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
12
0

Year Published

2020
2020
2021
2021

Publication Types

Select...
3
1

Relationship

2
2

Authors

Journals

citations
Cited by 4 publications
(12 citation statements)
references
References 9 publications
0
12
0
Order By: Relevance
“…Olah et al [40] highlights many examples of similar neurons in InceptionV1 and visualizes which concepts are detected by such neurons; however, the examples are manually curated by the authors. Identifying neurons that discover similar concepts also has practical benefits: in the neural network compression community, several methods [14,15,21,25,56] leverage potential neuron redundancies to generate compressed models while maintaining prediction accuracy. Even though these methods can measure neurons' similarity, there is limited work in interpreting their semantic similarity.…”
Section: Semantic Similarity Of Neuronsmentioning
confidence: 99%
See 2 more Smart Citations
“…Olah et al [40] highlights many examples of similar neurons in InceptionV1 and visualizes which concepts are detected by such neurons; however, the examples are manually curated by the authors. Identifying neurons that discover similar concepts also has practical benefits: in the neural network compression community, several methods [14,15,21,25,56] leverage potential neuron redundancies to generate compressed models while maintaining prediction accuracy. Even though these methods can measure neurons' similarity, there is limited work in interpreting their semantic similarity.…”
Section: Semantic Similarity Of Neuronsmentioning
confidence: 99%
“…Existing research on DNN interpretability tends to focus on inspecting individual neurons [23,37,42]. While helpful, neuron-level inspection cannot easily reveal how clusters of neurons may detect the same concept, even though it is common for multiple neurons to detect similar features [15,21,25,56]. As a result, users can easily miss higher-order interactions that explain how DNNs operate.…”
Section: Design Challengesmentioning
confidence: 99%
See 1 more Smart Citation
“…(3) Similarity measurement. These approaches use various strategies, such as geometric median [15] and clustering [51,6], to identify the most replaceable filters, or those functionally share the most similarity with others.…”
Section: Channel Pruningmentioning
confidence: 99%
“…We compare the performance of our approach with several recent channel pruning methods, namely, minimum weight (MW) [24], Taylor expansion [35], average percentage of zero activation neurons (APoZ) [17], soft filter pruning (SFP) [14], discrimination-aware channel pruning (DCP) [52], neuron importance score propagation (NISP) [47], slimmable neural networks (SNN) [46], autopruner (AP) [30], generative adversarial learning (GAL) [27], geometric median (GM) [15], transformable architecture search (TAS) [5], cluster pruning (CUP) [6], ABC [26], trained rank pruning (TRP) [45], soft channel pruning (SCP) [19], and high-hank (HRank) [25].…”
Section: Experiments Settingsmentioning
confidence: 99%