2019 16th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS) 2019
DOI: 10.1109/avss.2019.8909826
|View full text |Cite
|
Sign up to set email alerts
|

Information theory based pruning for CNN compression and its application to image classification and action recognition

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
5
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
2
2

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(5 citation statements)
references
References 17 publications
0
5
0
Order By: Relevance
“…In [55], quantization and pruning are used for compressing the model, consequently including the two main techniques applied in the proposed approach. In [52], they apply a reduction of parameters using covariance and correlation criteria to convolutional and fully connected layers. Many works just apply their methods to either convolutional or fully connected layers exclusively, so it is difficult to compare the method proposed with the same standards.…”
Section: Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…In [55], quantization and pruning are used for compressing the model, consequently including the two main techniques applied in the proposed approach. In [52], they apply a reduction of parameters using covariance and correlation criteria to convolutional and fully connected layers. Many works just apply their methods to either convolutional or fully connected layers exclusively, so it is difficult to compare the method proposed with the same standards.…”
Section: Resultsmentioning
confidence: 99%
“…This way, they get the relationships of similarity among filters and only leave without pruning those closer to the centroid of the cluster. Also, in [52] they propose a criteria based on the covariance and correlation of filters in convolutional and fully connected layers and successfully compressed different state-of-the-art models. Even though nowadays most typical criteria chosen by researchers are still the absolute value of weights, l1 and l2 norms [53].…”
Section: Network Pruningmentioning
confidence: 99%
“…(4) A face mask detection machine learning architecture is developed. (5) Easily reproducible open-source benchmarking templates are delivered that only use publicly available vision libraries. It is important to note that for the first time such a high number of hardware platforms, frameworks, and IC/OD models have been benchmarked and compared, not only on model latency performance but the full video pipeline.…”
Section: (3) a Comparison Between Raspberry Pi 4 Intel Neuralmentioning
confidence: 99%
“…Deploying efficiently AI applications on edge devices poses various challenges like discussed in [4], specifically constraints around compute, memory, and power consumption. To tackle these, quantization and weight pruning [5] are two popular techniques that normally trade a slight reduction in accuracy for performance gains. In quantization, the neural network weights and/or the feature maps are expressed by using shorter data types, such as FP16, INT16, or INT8 instead of FP32 [6]; this leads to a lower memory footprint as well as to a lower latency as the computation cost is reduced and the SIMD instructions can be used to calculate more operations per instruction.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation