2020
DOI: 10.1109/access.2020.3040849
|View full text |Cite
|
Sign up to set email alerts
|

Speculative Backpropagation for CNN Parallel Training

Abstract: The parallel learning in neural networks can greatly shorten the training time. Its prior efforts were mostly limited to distributing inputs to multiple computing engines. It is because the gradient descent algorithm in the neural network training is inherently sequential. This paper proposes a novel CNN parallel training method for image recognition. It overcomes the sequential property of the gradient descent and enables the parallel training with the speculative backpropagation. We found that the Softmax an… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
10
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 15 publications
(10 citation statements)
references
References 24 publications
(35 reference statements)
0
10
0
Order By: Relevance
“…Techniques which use speculative approaches [49], spiking neural network concepts [50,51] and memory use optimization [52] have also been proposed. Kim and Ko [53] and (separately) Ma, Lewis and Kleijn [54] have proposed techniques which train neural networks while specifically avoiding backpropagation altogether.…”
Section: Gradient Descent and Machine Learningmentioning
confidence: 99%
“…Techniques which use speculative approaches [49], spiking neural network concepts [50,51] and memory use optimization [52] have also been proposed. Kim and Ko [53] and (separately) Ma, Lewis and Kleijn [54] have proposed techniques which train neural networks while specifically avoiding backpropagation altogether.…”
Section: Gradient Descent and Machine Learningmentioning
confidence: 99%
“…The specific process of using CNN to train the SSH inversion model can be described as follows: (1) First, the training data are input into the network in batches for forward propagation; (2) Then, the weights of network are modified by using back propagation learning [28]; (3) The iterative training is carried out continuously and stops after 10000.…”
Section: ( ) ( || ( ) || ) K X Y Expmentioning
confidence: 99%
“…As a relatively mature technique, a variety of approaches have been developed, including those that have been shown to enhance performance speed [77], techniques that attempt to increase accuracy by introducing noise [78] and some that seek to create attack resilience [79,80]. Other techniques have incorporated federated learning and momentum [81] and used evolutionary algorithms [82], speculative approaches [83] and spiking neural network concepts [84,85]. Yet other techniques have focused on supporting deep networks [86], memory use optimization [87], bias factors [88,89] and initial condition sensitivity [90].…”
Section: Gradient Descentmentioning
confidence: 99%