IECON 2019 - 45th Annual Conference of the IEEE Industrial Electronics Society 2019
DOI: 10.1109/iecon.2019.8927153
|View full text |Cite
|
Sign up to set email alerts
|

Deep Learning Optimization for Edge Devices: Analysis of Training Quantization Parameters

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
3
3
2

Relationship

1
7

Authors

Journals

citations
Cited by 12 publications
(3 citation statements)
references
References 16 publications
0
3
0
Order By: Relevance
“…Successful use of DL algorithms in healthcare [1]- [3], ophthalmology [4]- [6], developmental disorders [7]- [9], in autonomous robots and vehicles [10]- [12], in image processing classification and detection [13], [14], in speech and audio processing [15], [16], cyber-security [17], [18], and many more indicate the reach of DL algorithms in our daily lives. Easier access to high-performance compute nodes using cloud computing ecosystems, high-throughput AI accelerators to enhance performance, and access to big-data scale datasets and storage enables deep learning providers to research, test, and operate ML algorithms at scale in small edge devices [19], smartphones [20], and AI-based web-services using Application Programming Interfaces (APIs) for wider exposure to any applications.…”
Section: Introductionmentioning
confidence: 99%
“…Successful use of DL algorithms in healthcare [1]- [3], ophthalmology [4]- [6], developmental disorders [7]- [9], in autonomous robots and vehicles [10]- [12], in image processing classification and detection [13], [14], in speech and audio processing [15], [16], cyber-security [17], [18], and many more indicate the reach of DL algorithms in our daily lives. Easier access to high-performance compute nodes using cloud computing ecosystems, high-throughput AI accelerators to enhance performance, and access to big-data scale datasets and storage enables deep learning providers to research, test, and operate ML algorithms at scale in small edge devices [19], smartphones [20], and AI-based web-services using Application Programming Interfaces (APIs) for wider exposure to any applications.…”
Section: Introductionmentioning
confidence: 99%
“…However, the training and inference stages of deep learning neural networks are limited by the space of the memory and a variety of factors, including programming complexity and even the reliability of the system. The process of quantization has become increasingly popular due to the significant impact on performance and minimal accuracy loss [220]. Furthermore, to ensure that local SE decisions are synchronized with minimum time delay, dynamic models that are trained online will be indispensable.…”
Section: B Deployment Of Advanced ML Solutionsmentioning
confidence: 99%
“…However, the training and inference stages of deep learning neural networks are limited by the space of the memory and a variety of factors, including programming complexity and even reliability of the system. On the whole the process of quantization has become increasingly popular due to the significant impact on performance and minimal accuracy loss [147]. Furthermore, to ensure that local SE decisions are synchronized with minimum time delay, dynamic models that are trained online will be indispensable.…”
Section: F Quantization-aware Model Trainingmentioning
confidence: 99%