2024
DOI: 10.3390/electronics13030644
|View full text |Cite
|
Sign up to set email alerts
|

AE-Qdrop: Towards Accurate and Efficient Low-Bit Post-Training Quantization for A Convolutional Neural Network

Jixing Li,
Gang Chen,
Min Jin
et al.

Abstract: Blockwise reconstruction with adaptive rounding helps achieve acceptable 4-bit post-training quantization accuracy. However, adaptive rounding is time intensive, and the optimization space of weight elements is constrained to a binary set, thus limiting the performance of quantized models. The optimality of block-wise reconstruction requires that subsequent network blocks remain unquantized. To address this, we propose a two-stage post-training quantization scheme, AE-Qdrop, encompassing block-wise reconstruct… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 33 publications
0
1
0
Order By: Relevance
“…Deep convolutional neural networks (DCNNs) have realized great success in many computer vision tasks, such as object identification [1][2][3], object detection [4][5][6] and image segmentation [7][8][9]. The design of DCNNs is becoming increasingly intricate, which is accompanied by a simultaneous enhancement in their performance.…”
Section: Introductionmentioning
confidence: 99%
“…Deep convolutional neural networks (DCNNs) have realized great success in many computer vision tasks, such as object identification [1][2][3], object detection [4][5][6] and image segmentation [7][8][9]. The design of DCNNs is becoming increasingly intricate, which is accompanied by a simultaneous enhancement in their performance.…”
Section: Introductionmentioning
confidence: 99%