2021
DOI: 10.3390/info12070264
|View full text |Cite
|
Sign up to set email alerts
|

Combine-Net: An Improved Filter Pruning Algorithm

Abstract: The powerful performance of deep learning is evident to all. With the deepening of research, neural networks have become more complex and not easily generalized to resource-constrained devices. The emergence of a series of model compression algorithms makes artificial intelligence on edge possible. Among them, structured model pruning is widely utilized because of its versatility. Structured pruning prunes the neural network itself and discards some relatively unimportant structures to compress the model’s siz… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(3 citation statements)
references
References 13 publications
(31 reference statements)
0
3
0
Order By: Relevance
“…Some of them play a feeble role in the target detection process, and the cumulative impact of these parameters on the feature map is negligible, and removing these parameters has little impact on the accuracy of target detection; therefore, the parameters between model layers need to be further compressed and optimized. Model pruning is a widely used model compression technique, and from the perspective of pruning granularity, pruning methods can be classified as structured and unstructured pruning (Wang et al, 2021 ), Filter Pruning via Geometric Median (FPGM) (He et al, 2019 ) is a structured weight pruning. The essence of the algorithm is to identify the geometric median close filters present in the network and achieve the purpose of streamlining the weights to accelerate inference by eliminating the redundant filters and their associated input-output relations.…”
Section: Methodsmentioning
confidence: 99%
“…Some of them play a feeble role in the target detection process, and the cumulative impact of these parameters on the feature map is negligible, and removing these parameters has little impact on the accuracy of target detection; therefore, the parameters between model layers need to be further compressed and optimized. Model pruning is a widely used model compression technique, and from the perspective of pruning granularity, pruning methods can be classified as structured and unstructured pruning (Wang et al, 2021 ), Filter Pruning via Geometric Median (FPGM) (He et al, 2019 ) is a structured weight pruning. The essence of the algorithm is to identify the geometric median close filters present in the network and achieve the purpose of streamlining the weights to accelerate inference by eliminating the redundant filters and their associated input-output relations.…”
Section: Methodsmentioning
confidence: 99%
“…In the unstructured pruning, 45 unimportant or least important connections and neurons in the pretrained model are removed depending on the value of the weight magnitudes. Weight pruning optimizes the DL model by removing the unimportant weights from the neural network.…”
Section: Model Compression Strategies For Edge Computingmentioning
confidence: 99%
“…Representing the model weight with lower bit‐width parameters reduces the inference latency and storage. A few quantization approaches 45 are k‐means clustering along with Huffman encoding, binary quantization, and 1‐bit quantization to represent a 32‐bit number to a 1‐bit integer.…”
Section: Model Compression Strategies For Edge Computingmentioning
confidence: 99%