2020
DOI: 10.1109/twc.2019.2946140
|View full text |Cite
|
Sign up to set email alerts
|

Edge AI: On-Demand Accelerating Deep Neural Network Inference via Edge Computing

Abstract: As a key technology of enabling Artificial Intelligence (AI) applications in 5G era, Deep Neural Networks (DNNs) have quickly attracted widespread attention. However, it is challenging to run computation-intensive DNN-based tasks on mobile devices due to the limited computation resources. What's worse, traditional cloud-assisted DNN inference is heavily hindered by the significant wide-area network latency, leading to poor real-time performance as well as low quality of user experience. To address these challe… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
218
0
1

Year Published

2020
2020
2024
2024

Publication Types

Select...
7
2
1

Relationship

2
8

Authors

Journals

citations
Cited by 582 publications
(244 citation statements)
references
References 39 publications
0
218
0
1
Order By: Relevance
“…Specifically, the collaborative learning process in PerFit mainly consists of the following three stages as depicted in r Offloading stage: When the edge is trustworthy (e.g., edge gateway at home), the IoT device user can offload its whole learning model and data samples to the edge for fast computation. Otherwise, the device user will carry out model partitioning by keeping the input layers and its data samples locally on its device and offloading the remaining model layers to the edge for device-edge collaborative computing [14].…”
Section: Cloud-edge Framework For Personalized Federated Learningmentioning
confidence: 99%
“…Specifically, the collaborative learning process in PerFit mainly consists of the following three stages as depicted in r Offloading stage: When the edge is trustworthy (e.g., edge gateway at home), the IoT device user can offload its whole learning model and data samples to the edge for fast computation. Otherwise, the device user will carry out model partitioning by keeping the input layers and its data samples locally on its device and offloading the remaining model layers to the edge for device-edge collaborative computing [14].…”
Section: Cloud-edge Framework For Personalized Federated Learningmentioning
confidence: 99%
“…Once appeared, it received a tremendous amount of interest. In terms of edge AI algorithms, in [22] [23], Li et al proposed an Edgent framework, which leverages deviceedge synergy computing to accelerate training and inference of deep neural network (DNN). And the authors in [24] design a momentum federated learning (MFL) framework to accelerate federated learning by momentum gradient descent.…”
Section: B Edge Intelligence For Iotmentioning
confidence: 99%
“…We call such a scenario as the Non-MEC system, and will compare the two systems in the simulation section. 6 2) CP's Profit Maximization in Stage I: To maximize the CP's profit regarding the revenue sharing ratio, we need to first analyze the total profit of the CP. Recall that p is the price of per content request paid to the CP, s CD is the CP's delivery cost of per content request, and I · (|I RE | − B|I AG |) is the extra requesters' demands that exceed the agents' total serving capacities.…”
Section: Stackelberg Equilibrium Definitionmentioning
confidence: 99%