1995
DOI: 10.1016/0893-6080(94)00100-z
|View full text |Cite
|
Sign up to set email alerts
|

The generalized proportional-integral-derivative (PID) gradient descent back propagation algorithm

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
2
0

Year Published

1996
1996
2022
2022

Publication Types

Select...
3
3
1

Relationship

0
7

Authors

Journals

citations
Cited by 15 publications
(2 citation statements)
references
References 10 publications
0
2
0
Order By: Relevance
“…The use of PID laws in various nonlinear optimization problems have revealed a different and interesting application field besides their conventional use for the minimization of error signals in control systems. For example, it has been presented that the problems related to reaching the global minimum can be solved by providing better convergence speed by using PID actions in the conventional gradient descent algorithm [2]. In another study, a three term backpropagation algorithm was developed by adding a proportional action to the standard backpropagation algorithm, and it was shown to be successful in terms of convergence rate and local minimums for artificial neural networks training [3].…”
Section: Introductionmentioning
confidence: 99%
“…The use of PID laws in various nonlinear optimization problems have revealed a different and interesting application field besides their conventional use for the minimization of error signals in control systems. For example, it has been presented that the problems related to reaching the global minimum can be solved by providing better convergence speed by using PID actions in the conventional gradient descent algorithm [2]. In another study, a three term backpropagation algorithm was developed by adding a proportional action to the standard backpropagation algorithm, and it was shown to be successful in terms of convergence rate and local minimums for artificial neural networks training [3].…”
Section: Introductionmentioning
confidence: 99%
“…There are a number of ways to decrease the training time. Several improvements have been made to the rule used to adjust the weights to increase the rate of descent down the error surface or to decrease a network's likelihood of being trapped in a local minima [5]. For example, the conjugate gradient method can in some cases decrease the number of iterations by an order of magnitude [6].…”
Section: Introductionmentioning
confidence: 99%