2017
DOI: 10.1049/el.2017.2219
|View full text |Cite
|
Sign up to set email alerts
|

Quantisation and pooling method for low‐inference‐latency spiking neural networks

Abstract: Spiking neural network (SNN) that converted from conventional deep neural network (DNN) has shown great potential as a solution for fast and efficient recognition. A layer-wise quantisation method based on retraining is proposed to quantise the activation of DNN, which reduces the number of time steps required by converted SNN to achieve minimal accuracy loss. Pooling function is incorporated into convolutional layers to reduce at most 20% of spiking neurons. The converted SNNs achieved 99.15% accuracy on MNIS… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 12 publications
(3 citation statements)
references
References 4 publications
0
3
0
Order By: Relevance
“…When applying NPTD , if the pruning thresholds are individually assigned to all the neurons, search space becomes prohibitively large and the memories to store the pruning thresholds should be large as well. Referred from the previous literatures ( Bengio et al, 2006 ; Lin et al, 2017 ; Wang et al, 2019 ), where layer-wise search algorithms are utilized to find optimum design points such as bit-widths of quantization or approximation parameters, pruning thresholds of the NPTD are searched per layer in this work.…”
Section: Neuron Pruning In Temporal Domains ( Nptd )mentioning
confidence: 99%
“…When applying NPTD , if the pruning thresholds are individually assigned to all the neurons, search space becomes prohibitively large and the memories to store the pruning thresholds should be large as well. Referred from the previous literatures ( Bengio et al, 2006 ; Lin et al, 2017 ; Wang et al, 2019 ), where layer-wise search algorithms are utilized to find optimum design points such as bit-widths of quantization or approximation parameters, pruning thresholds of the NPTD are searched per layer in this work.…”
Section: Neuron Pruning In Temporal Domains ( Nptd )mentioning
confidence: 99%
“…However, even in the spike domain, the MaxPooling tends to produce higher classification accuracy (Rueckauer et al, 2017 ) than the aforementioned alternatives. When implementing the spiking MaxPooling, researchers are drawn to several popular approaches: rate-based spike accumulation (Hu and Pfeiffer, 2016 ; Chen et al, 2018 ; Kim et al, 2020 ), time-to-first-spike (Masquelier and Thorpe, 2007 ; Zhao et al, 2014 ; Li J. et al, 2017 ; Mozafari et al, 2019 ), and lateral inhibition or temporal winner-take-all (Orchard et al, 2015 ; Lin et al, 2017 ).…”
Section: Sub-sampling By Pooling Operationmentioning
confidence: 99%
“…To improve the real-time performance of SNN, [8] proposed two optimization methods to normalize the network weights, namely model-based normalization 205 and data-based normalization, so that the neuron activations were sufficiently small to prevent from overestimating output activations. Retraining based layer-wise quantization method to quantize the neuron activation and pooling layer incorporation to reduce the number requirement of neurons were pro-210 posed in [25], the authors reported that these methods can build hardware-friendly SNNs with ultra-low-inference latency.…”
Section: Inference Latency 185mentioning
confidence: 99%