2018 51st Annual IEEE/ACM International Symposium on Microarchitecture (MICRO) 2018
DOI: 10.1109/micro.2018.00038
|View full text |Cite
|
Sign up to set email alerts
|

In-Register Parameter Caching for Dynamic Neural Nets with Virtual Persistent Processor Specialization

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
8
1

Relationship

0
9

Authors

Journals

citations
Cited by 16 publications
(5 citation statements)
references
References 48 publications
0
5
0
Order By: Relevance
“…• Hardware Accelerators: Data-flow execution models using GPUs [46,37], FPGAs [101,107,120,64,60] and ASICs [42,4,21,115] are more efficient choices for CNNs than traditional CPUs. Among these, FPGAs are more flexible compared to ASICs and more efficient than GPUs.…”
Section: Power-efficient Cnnsmentioning
confidence: 99%
“…• Hardware Accelerators: Data-flow execution models using GPUs [46,37], FPGAs [101,107,120,64,60] and ASICs [42,4,21,115] are more efficient choices for CNNs than traditional CPUs. Among these, FPGAs are more flexible compared to ASICs and more efficient than GPUs.…”
Section: Power-efficient Cnnsmentioning
confidence: 99%
“…GPUrdma [65] proposed a matrix-vector product persistent kernel holding a constant matrix in shared memory. Khorasani et al [67] use persistent threads to keep parameters in cache. Zhu et al [68] proposed a sparse persistent implementation of recurrent neural networks.…”
Section: Related Workmentioning
confidence: 99%
“…Training [57]. Cross-layer approaches related to our work include high- [52] and low- [29] level code generation techniques, and also memory management [23] and memory partitioning techniques [25], [35], [53]. There have been some recent works on SIMD and in particular looking at AVX extensions.…”
Section: Verma Et Al Present a Workload Characterization Of Mlperfmentioning
confidence: 99%