2019
DOI: 10.1109/jetcas.2019.2950093
|View full text |Cite
|
Sign up to set email alerts
|

EBPC: Extended Bit-Plane Compression for Deep Neural Network Inference and Training Accelerators

Abstract: In the wake of the success of convolutional neural networks in image classification, object recognition, speech recognition, etc., the demand for deploying these compute-intensive ML models on embedded and mobile systems with tight power and energy constraints at low cost, as well as for boosting throughput in data centers, is growing rapidly. This has sparked a surge of research into specialized hardware accelerators. Their performance is typically limited by I/O bandwidth, power consumption is dominated by I… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
21
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
2
1

Relationship

1
6

Authors

Journals

citations
Cited by 31 publications
(21 citation statements)
references
References 41 publications
0
21
0
Order By: Relevance
“…Activations extracted by ReLU provide a known sparsity [10], thus several sparsity-based coding formats are designed to compress feature maps: Coordinate (COO) [25], Bitmap [26], Run-Length Coding [27] etc.. Several methods based on these coding formats describe their hardware architecture to reduce computational cost: Cnvlutin [11], SCNN [12], Eyeriss [13], EIE [14], etc.. Cambricon-SE [28] tries to use Huffman and LZW to improve the compression ratio of feature maps.…”
Section: A Coding Format and Compression Encodersmentioning
confidence: 99%
See 1 more Smart Citation
“…Activations extracted by ReLU provide a known sparsity [10], thus several sparsity-based coding formats are designed to compress feature maps: Coordinate (COO) [25], Bitmap [26], Run-Length Coding [27] etc.. Several methods based on these coding formats describe their hardware architecture to reduce computational cost: Cnvlutin [11], SCNN [12], Eyeriss [13], EIE [14], etc.. Cambricon-SE [28] tries to use Huffman and LZW to improve the compression ratio of feature maps.…”
Section: A Coding Format and Compression Encodersmentioning
confidence: 99%
“…As a result, the large data of feature maps on CNN have to be repeatedly transferred between on-chip and off-chip memory, which greatly increases the power consumption and the data transfer bandwidth. Therefore, it is a topic worth exploring to reduce this power consumption and latency by compressing feature maps to improve the performance of CNN models on specialized hardware [10].…”
Section: Introductionmentioning
confidence: 99%
“…The main drawback of compression representation learning based approaches is that they alter the DNN model and then require a retraining phase. EBPC [6] is a hardware-friendly and lossless compression scheme for the feature maps present within CNNs. However it is limited on the compression of the feature maps although the model parameters/weights are responsible for a major fraction of the overall memory/communication traffic (see Fig.…”
Section: Related Workmentioning
confidence: 99%
“…Current DNN models rely on millions or even billions of parameters, thus exacerbating the role played by the communication and memory sub-systems for moving such high data volume from the main memory into the accelerator and then to its many PEs. Thus, the performance and energy figures of a DNN accelerator are severely affected by the communication and memory sub-system [6], [7]. Fig.…”
Section: Introductionmentioning
confidence: 99%
“…However, transmitting the data comes at a high energy cost, introduces privacy concerns, requires expensive infrastructure, and has high latency. Alternatively, the challenge of analyzing the data near the sensor can be tackled by a combination of algorithmic optimizations to allow working with reduced arithmetic precision, and hardware acceleration [3], [4] with various techniques to maximize energy efficiency, such as minimizing off-accelerator data transfers [5]- [7].…”
Section: Introductionmentioning
confidence: 99%