2005
DOI: 10.1109/tnn.2005.844903
|View full text |Cite
|
Sign up to set email alerts
|

Deterministic Convergence of an Online Gradient Method for BP Neural Networks

Abstract: Online gradient methods are widely used for training feedforward neural networks. We prove in this paper a convergence theorem for an online gradient method with variable step size for backward propagation (BP) neural networks with a hidden layer. Unlike most of the convergence results that are of probabilistic and nonmonotone nature, the convergence result that we establish here has a deterministic and monotone nature.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
50
1

Year Published

2006
2006
2017
2017

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 139 publications
(52 citation statements)
references
References 17 publications
1
50
1
Order By: Relevance
“…The number of layers of perceptrons is made up of the number of layers in a NN. FFNNs could be used to map any function from input to output and they are known as gradient based learning algorithms (Steepest Decent Method) which is the supreme algorithm used in FFNNs [6][7][8][9][10]. However, since FFNNs model is based on human thought processes, it is essential to appreciate how the human brain functions on a basic level [11,12].…”
Section: Taxonomy Of Fnnmentioning
confidence: 99%
See 1 more Smart Citation
“…The number of layers of perceptrons is made up of the number of layers in a NN. FFNNs could be used to map any function from input to output and they are known as gradient based learning algorithms (Steepest Decent Method) which is the supreme algorithm used in FFNNs [6][7][8][9][10]. However, since FFNNs model is based on human thought processes, it is essential to appreciate how the human brain functions on a basic level [11,12].…”
Section: Taxonomy Of Fnnmentioning
confidence: 99%
“…More so, there's no need to use more than one HL for many real-world issues and problems that require two HLs are seldom bump into. Differences between the numbers of HLs are summarized in table 1 below: [7] Again, deciding on the number of hidden neurons in layers is a very critical part in determining the overall neural network architecture. Although they do not directly relate with the external environment these layers have an incredible impact on the ultimate output.…”
Section: Fig 4 Conduit Structure Of Ffnnmentioning
confidence: 99%
“…On the other hand, the fact that { } n V is monotonically non-increasing sequence is not necessary to achieve (20) in principle.…”
Section: Preliminariesmentioning
confidence: 99%
“…Unfortunately, this gives that the learning goes faster in the beginning and slows down in the late stage. The convergence analysis of learning algorithm with deterministic (non-stochastic) nature has been given in [17][18][19][20][21][22]. In contrast to the stochastic approach, several of these results allow to employ a constant learning rate [19,23].…”
Section: Introductionmentioning
confidence: 99%
“…These problems are inherent to the basic learning rule of FNN that are based on GD optimization methods [15,30]. The convergence properties of such algorithms are discussed in [5,7,12,15,16,18], and [22]. Learning algorithms based on GD includes real-time recurrent learning (RTRL), ordered derivative learning and so on [1].…”
Section: Introductionmentioning
confidence: 99%