2020
DOI: 10.3390/s20216033
|View full text |Cite
|
Sign up to set email alerts
|

Compressing Deep Networks by Neuron Agglomerative Clustering

Abstract: In recent years, deep learning models have achieved remarkable successes in various applications, such as pattern recognition, computer vision, and signal processing. However, high-performance deep architectures are often accompanied by a large storage space and long computational time, which make it difficult to fully exploit many deep neural networks (DNNs), especially in scenarios in which computing resources are limited. In this paper, to tackle this problem, we introduce a method for compressing the struc… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(2 citation statements)
references
References 32 publications
(31 reference statements)
0
2
0
Order By: Relevance
“…This approach takes into account that several connections may share the same weight value, and then fine-tunes those shared weights. In the case of feedforward structures, this strategy was already successfully employed to minimize the complexity of NN models [46], [69]- [71]. In this paper, we use the same method as in [46], but modify it for the recurrent layers as well.…”
Section: B Weights Clusteringmentioning
confidence: 99%
“…This approach takes into account that several connections may share the same weight value, and then fine-tunes those shared weights. In the case of feedforward structures, this strategy was already successfully employed to minimize the complexity of NN models [46], [69]- [71]. In this paper, we use the same method as in [46], but modify it for the recurrent layers as well.…”
Section: B Weights Clusteringmentioning
confidence: 99%
“…However, the existing research work does not consider the heterogeneous capabilities of IoT devices, dynamic changes of environmental conditions, and is difficult to achieve real-time adaptive decision-making under the diversified environment configuration and high computational complexity of problem solving. It is worth noting that, the above work is orthogonal to the compression and acceleration methods that use weight pruning [25,26], quantization [27,28] and low-precision inference [29,30] to reduce the computational cost of DNN models. At the same time, these two technologies are used to accelerate the DNN inference.…”
Section: Introductionmentioning
confidence: 99%