Proceedings of the Workshop on Distributed Infrastructures for Deep Learning 2019
DOI: 10.1145/3366622.3368147
|View full text |Cite
|
Sign up to set email alerts
|

Challenges and Opportunities of DNN Model Execution Caching

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
0
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
1

Relationship

1
4

Authors

Journals

citations
Cited by 6 publications
(1 citation statement)
references
References 12 publications
0
0
0
Order By: Relevance
“…An interesting research question is then how to manage a large number of deep learning models given dynamic workload to satisfy performance objectives while lowering resource costs [91]. One promising direction is to design deep learning-specific caching algorithms to manage main memory resources and minimize the performance impact of the model cold start problem [87]. Formulating the model management problem as a caching problem allows leveraging rich literature on caching; however, questions such as how to incorporate the relatively slow PCIe transfer to GPUs and effectively use virtual GPU memory remain unsolved.…”
Section: Iot and Ai Are Becoming The Main Applicationsmentioning
confidence: 99%
“…An interesting research question is then how to manage a large number of deep learning models given dynamic workload to satisfy performance objectives while lowering resource costs [91]. One promising direction is to design deep learning-specific caching algorithms to manage main memory resources and minimize the performance impact of the model cold start problem [87]. Formulating the model management problem as a caching problem allows leveraging rich literature on caching; however, questions such as how to incorporate the relatively slow PCIe transfer to GPUs and effectively use virtual GPU memory remain unsolved.…”
Section: Iot and Ai Are Becoming The Main Applicationsmentioning
confidence: 99%