2015 International Conference on Parallel Architecture and Compilation (PACT) 2015
DOI: 10.1109/pact.2015.21
|View full text |Cite
|
Sign up to set email alerts
|

Towards General-Purpose Neural Network Computing

Abstract: Abstract-Machine learning is becoming pervasive; decades of research in neural network computation is now being leveraged to learn patterns in data and perform computations that are difficult to express using standard programming approaches. Recent work has demonstrated that custom hardware accelerators for neural network processing can outperform software implementations in both performance and power consumption. However, there is neither an agreed-upon interface to neural network accelerators nor a consensus… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 17 publications
(4 citation statements)
references
References 31 publications
0
4
0
Order By: Relevance
“…Accelerator design for neural networks has become a major line of computer architecture research in recent years. A handful of prior work explored the design space of neural network acceleration, which can be categorized into ASICs [15], [16], [18]- [22], [26], [27], [30], [34], [37], [38], [41], [42], FPGA implementations [17], [28], [35], [36], [43], using unconventional devices for acceleration [29], [33], [40], and dataflow optimizations [16], [23]- [25], [31], [32], [39]. Most of these studies have focused on accelerator design and optimization of merely one specific type of convolutional as the most computeintensive operation in deep convolutional neural networks.…”
Section: -Gmentioning
confidence: 99%
“…Accelerator design for neural networks has become a major line of computer architecture research in recent years. A handful of prior work explored the design space of neural network acceleration, which can be categorized into ASICs [15], [16], [18]- [22], [26], [27], [30], [34], [37], [38], [41], [42], FPGA implementations [17], [28], [35], [36], [43], using unconventional devices for acceleration [29], [33], [40], and dataflow optimizations [16], [23]- [25], [31], [32], [39]. Most of these studies have focused on accelerator design and optimization of merely one specific type of convolutional as the most computeintensive operation in deep convolutional neural networks.…”
Section: -Gmentioning
confidence: 99%
“…The study of [88] proposes an approach of k-modular redundancy which is created by k-fold replicating of hidden neurons and dividing the weights to k compared with the voting mechanism of the conventional methods. The authors evaluated the method for three applications on the NN accelerator: Black-Scholes, RSA, and Sobel.…”
Section: ) Resilience Enhancement By Redundancymentioning
confidence: 99%
“…Research in this area traditionally has employed HLS tools as "compiler" intermediaries between high-level implementations and the actual hardware design [48], which suffer from unpredictable resource usage. Recently exploratory work has been done using Chisel as the high-level implementation language and the low-level description language [49]. While this still is in its initial stages, we have begun to explore and develop such a solution for the purposes of Bragg peak detection [50].…”
Section: Lightweight Ai Capability For Future Detector Systemsmentioning
confidence: 99%