2019 57th Annual Allerton Conference on Communication, Control, and Computing (Allerton) 2019
DOI: 10.1109/allerton.2019.8919683
|View full text |Cite
|
Sign up to set email alerts
|

pSConv: A Pre-defined Sparse Kernel Based Convolution for Deep CNNs

Abstract: The high demand for computational and storage resources severely impedes the deployment of deep convolutional neural networks (CNNs) in limited resource devices. Recent CNN architectures have proposed reduced complexity versions (e.g,. SuffleNet and MobileNet) but at the cost of modest decreases in accuracy. This paper proposes pSConv, a pre-defined sparse 2D kernel based convolution, which promises significant improvements in the trade-off between complexity and accuracy for both CNN training and inference. T… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
4
3
1

Relationship

4
4

Authors

Journals

citations
Cited by 9 publications
(6 citation statements)
references
References 21 publications
0
6
0
Order By: Relevance
“…Our model can be extended to fullyconnected layers with f l i and f l o as the number of input and output features respectively. In particular, for an ANN, the total number of FLOPS for layer l, denoted F l AN N , is shown in row 1 of Table III [49], [50]. The formula can be easily adjusted for an SNN in which the number of FLOPs at layer l is a function of the average spiking activity at the layer (ζ l ) denoted as F l SN N in Table III.…”
Section: B Reduction In Flops and Compute Energymentioning
confidence: 99%
“…Our model can be extended to fullyconnected layers with f l i and f l o as the number of input and output features respectively. In particular, for an ANN, the total number of FLOPS for layer l, denoted F l AN N , is shown in row 1 of Table III [49], [50]. The formula can be easily adjusted for an SNN in which the number of FLOPs at layer l is a function of the average spiking activity at the layer (ζ l ) denoted as F l SN N in Table III.…”
Section: B Reduction In Flops and Compute Energymentioning
confidence: 99%
“…Behavior cloning, a method of teaching a machine learning algorithm how to complete a task, was implemented for the training of this model. This method was chosen because of its ease of implementation and its ability to "yield optimal results efficiently" [15]. Behavior cloning allows for pre-collected data from a human expert to be treated as a ground truth dataset.…”
Section: Trainingmentioning
confidence: 99%
“…This section first describes pSConv, a form of pre-defined sparse kernel based convolution that we initially proposed in [1]. It then describe how we introduce periodicity to this framework to reduce the overhead of managing sparse matrix representations.…”
Section: Pre-defined Sparsitymentioning
confidence: 99%
“…This paper proposes pre-defined sparse convolutions to improve energy and storage efficiency during both training and inference. We refer to this approach as pSConv and presented initial simulation results that show negligible performance degradation compared to fully-connected baseline models in [1]. However, as mentioned earlier, unstructured forms of pSConv may not lead to energy reductions due to the overhead of managing their sparse matrix representations.…”
Section: Introductionmentioning
confidence: 99%