2019
DOI: 10.1109/access.2019.2920885
|View full text |Cite
|
Sign up to set email alerts
|

A Novel Systolic Parallel Hardware Architecture for the FPGA Acceleration of Feedforward Neural Networks

Abstract: New chips for machine learning applications appear, they are tuned for a specific topology, being efficient by using highly parallel designs at the cost of high power or large complex devices. However, the computational demands of deep neural networks require flexible and efficient hardware architectures able to fit different applications, neural network types, number of inputs, outputs, layers, and units in each layer, making the migration from software to hardware easy. This paper describes novel hardware im… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
21
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 51 publications
(21 citation statements)
references
References 43 publications
0
21
0
Order By: Relevance
“…There are several feed-forward neural network methods, that is radial basis function NN (RBFNN), extreme learning machine (ELM), kernel extreme learning machine (KELM), and perceptron [20]. Figure 1 shows the design of the FFNN algorithm with circles in a network that forms neurons in an artificial neural network [28]. − Feed backward neural network (FBNN)…”
Section: Neural Networkmentioning
confidence: 99%
“…There are several feed-forward neural network methods, that is radial basis function NN (RBFNN), extreme learning machine (ELM), kernel extreme learning machine (KELM), and perceptron [20]. Figure 1 shows the design of the FFNN algorithm with circles in a network that forms neurons in an artificial neural network [28]. − Feed backward neural network (FBNN)…”
Section: Neural Networkmentioning
confidence: 99%
“…It is further divided into two sub-components: five systolic arrays implementing the RNN cells and Integration and Prediction unit that combines the outputs of the RNN cells, which makes a prediction of the correctness of the currently fetched instruction. The systolic array architecture [29] implements the vector matrix multiplications of the RNN cells. In total, five systolic array units are used where each one contains eight multiplication-andaccumulation (MAC) elements.…”
Section: B Implementation Of the Detectormentioning
confidence: 99%
“…Artificial neural networks (ANNs) are an important area of artificial intelligence (AI) used to perform several tasks, such as classification [1][2][3][4], pattern recognition [5][6][7][8], communications [9,10], control systems [11,12], prediction [13,14], among others. An ANN models a biological neural network employing a collection of nodes called artificial neurons, connected by edges to transmit signals like the synapses in a brain; during its transmission, the signal value changes according to the weight of the edges, adjusted by a learning process.…”
Section: Introductionmentioning
confidence: 99%
“…An attractive solution is the development of hardware neuronal networks (HNN) in Field-Programmable Gate Arrays (FPGAs) [15][16][17][18][19][20][21]. In this regard, the FPGA-based implementation of AFs in HNN is one of the challenges for embedded system design according to recent studies; this is because the AF implementations require low hardware resources and low power consumption [1,2,5,12,[22][23][24][25]. Currently, the most common non-linear functions for ANNs are the Sigmoid [11,[26][27][28][29][30][31][32] ans TanhAFs [22,32,33].…”
Section: Introductionmentioning
confidence: 99%