2021
DOI: 10.48550/arxiv.2104.10645
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Measuring what Really Matters: Optimizing Neural Networks for TinyML

Lennart Heim,
Andreas Biri,
Zhongnan Qu
et al.

Abstract: With the surge of inexpensive computational and memory resources, neural networks (NNs) have experienced an unprecedented growth in architectural and computational complexity. Introducing NNs to resource-constrained devices enables cost-efficient deployments, widespread availability, and the preservation of sensitive data.This work addresses the challenges of bringing Machine Learning to microcontroller units (MCUs), where we focus on the ubiquitous ARM Cortex-M architecture. The detailed effects and trade-off… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
8
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 7 publications
(8 citation statements)
references
References 27 publications
0
8
0
Order By: Relevance
“…Although quantization is still an open research topic, it has become a standard compression method for TinyMLrelated applications. It allows the consumption of less flash memory and RAM whilst maintaining almost the same accuracy of the original model [28]. Moreover, compressing the network from 32-bit to 8-bit results in a significantly faster processing, shorter inference time, and lower power consumption.…”
Section: Methodsmentioning
confidence: 99%
“…Although quantization is still an open research topic, it has become a standard compression method for TinyMLrelated applications. It allows the consumption of less flash memory and RAM whilst maintaining almost the same accuracy of the original model [28]. Moreover, compressing the network from 32-bit to 8-bit results in a significantly faster processing, shorter inference time, and lower power consumption.…”
Section: Methodsmentioning
confidence: 99%
“…Table 1: Comparative overview of the digital hardware capabilities Software [26], [25] [27] MCUs [17], [18], [21] [22] FPGA [13], [14], [15] [19], [23] ASIC [14], [20 It can be seen from the table that while software, microcontrollers, FPGAs, and Application Specific Integrated Circuits (ASICs) can all implement ANNs, FPGAs offer distinct advantages in terms of performance, efficiency, flexibility, resource utilization, scalability, real-time capabilities, and price.…”
Section: The Digital Implementation Approachmentioning
confidence: 99%
“…The core implementation of TinyML relies on ML model quantization, which reduces its numerical precision and size. Hence, implementing TinyML in environments with limited resources presents several ongoing challenges -The low computational capabilities and storage capacities of smaller devices restrict the complexity of the models that can be deployed [351]. This constraint can adversely affect the efficacy and precision of TinyML-based applications.…”
Section: Tiny Machine Learningmentioning
confidence: 99%