2017
DOI: 10.18517/ijaseit.7.5.2972
|View full text |Cite
|
Sign up to set email alerts
|

An Optimized Back Propagation Learning Algorithm with Adaptive Learning Rate

Abstract: Back Propagation (BP) is commonly used algorithm that optimize the performance of network for training multilayer feed-forward artificial neural networks. However, BP is inherently slow in learning and it sometimes gets trapped at local minima. These problems occur mailnly due to a constant and non-optimum learning rate (a fixed step size) in which the fixed value of learning rate is set to an initial starting value before training patterns for an input layer and an output layer. This fixed learning rate often… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
6
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
6
2
1

Relationship

0
9

Authors

Journals

citations
Cited by 12 publications
(6 citation statements)
references
References 14 publications
0
6
0
Order By: Relevance
“…However, there are also the researchers who use architectures with the number of neurons in the hidden layer equal to the number of neurons in the input layer [4], [5]. Some studies on the prediction of time series data using NNBP architecture with one hidden layer has been widely done namely classification of Australian credit card [6], diabetic detection [7], identification for a single-shaft gas turbine [8], particle swarm optimization [9], measuring the severity of osteoarthritis [10] [9] with architecture 40-20-1, and obtained a coefficient of correlation value of 0.990. The average architecture they used is that the number of neurons in the hidden layer is smaller than the number of neurons in the input layer.…”
Section: Introductionmentioning
confidence: 99%
“…However, there are also the researchers who use architectures with the number of neurons in the hidden layer equal to the number of neurons in the input layer [4], [5]. Some studies on the prediction of time series data using NNBP architecture with one hidden layer has been widely done namely classification of Australian credit card [6], diabetic detection [7], identification for a single-shaft gas turbine [8], particle swarm optimization [9], measuring the severity of osteoarthritis [10] [9] with architecture 40-20-1, and obtained a coefficient of correlation value of 0.990. The average architecture they used is that the number of neurons in the hidden layer is smaller than the number of neurons in the input layer.…”
Section: Introductionmentioning
confidence: 99%
“…This is because even if the output unit saturates the corresponding decent gradient takes a small value, even if the output error is large, which will result in no significant progress in the weight adjustment. The second disadvantage of this method is the difficulty in choosing a proper learning rate ƞ to achieve fast learning while maintaining the learning procedure stable [ 35 ]. These problems contribute to the lack of an inability to apply conventional BP to a wide number of applications.…”
Section: The Proposed Approachmentioning
confidence: 99%
“…The Mini-batch gradient descent is a compromise between the two approaches. This gradient descent method performs an update for each mini-batch of training data; therefore, the entire training data set is not used, but only a certain part of it, see (3) [16], [17]. Unlike the stochastic gradient descent, this method is not as sensitive to changes in hyper-parameters, especially the learning rate used during optimization, which leads to a reduction in fluctuations and thus to a more stable convergence [14].…”
Section: Deep Feedforward Networkmentioning
confidence: 99%
“…Back-propagation (BP) is a commonly used algorithm for training the multilayer feed forward NN [16]. There are various learning parameters, such as the learning rate, momentum or activation function which can improve the BP learning algorithm.…”
Section: A Back-propagationmentioning
confidence: 99%