1990
DOI: 10.1007/3-540-52255-7_32
|View full text |Cite
|
Sign up to set email alerts
|

Acceleration techniques for the backpropagation algorithm

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
52
0
3

Year Published

1993
1993
2017
2017

Publication Types

Select...
6
1
1

Relationship

1
7

Authors

Journals

citations
Cited by 113 publications
(55 citation statements)
references
References 4 publications
0
52
0
3
Order By: Relevance
“…Since the basic version of BP is sensitivity towards learning rate and momentum factor [15], several improvements were suggested by researchers: 1) a fast BP algorithm, called Quickpro was proposed in [66,67]; 2) a delta-bar technique and an acceleration technique was suggested for tuning BP learning rate η in [68] and in [69] respectively; and 3) a variant of BP, called resilient propagator (Rprop) was proposed in [70].…”
Section: Conventional Optimization Approachesmentioning
confidence: 99%
“…Since the basic version of BP is sensitivity towards learning rate and momentum factor [15], several improvements were suggested by researchers: 1) a fast BP algorithm, called Quickpro was proposed in [66,67]; 2) a delta-bar technique and an acceleration technique was suggested for tuning BP learning rate η in [68] and in [69] respectively; and 3) a variant of BP, called resilient propagator (Rprop) was proposed in [70].…”
Section: Conventional Optimization Approachesmentioning
confidence: 99%
“…We can then consider such penalties as ravines that are parallel to some axes. The so-called adaptive step size technique [9], which was originally proposed for accelerating the optimization procedure in neural networks learning, can then be exploited for optimization involving such penalties. Note that for a ravine in the objective function parallel to an axis, use of an appropriate individual step size is equivalent to re-scaling the ravine.…”
Section: A Simple Approach For Optimization Involving L 1 Penaltiesmentioning
confidence: 99%
“…The following strategies are usually suggested: (i) start with a small learning rate and increase it exponentially, if suc-cessive epochs reduce the error, or rapidly decrease it, if a significant error increase occurs [3,25], (ii) start with a small learning rate and increase it, if successive epochs keep gradient direction fairly constant, or rapidly decrease it, if the direction of the gradient varies greatly at each epoch [6], (iii) for each weight, an individual learning rate is given, which increases if the successive changes in the weights are in the same direction and decreases otherwise [10,15,17,22], and (iv) use a closed formula to calculate a common learning rate for all the weights at each iteration [9,12,16] or a different learning rate for each weight [7,13]. Note that all the above-mentioned strategies employ heuristic parameters in an attempt to enforce the decrease of the learning error at each iteration and to secure the converge of the training algorithm.…”
Section: Deterministic Learning Rate Adaptationmentioning
confidence: 99%