2021
DOI: 10.3390/math9172176
|View full text |Cite
|
Sign up to set email alerts
|

Adaptive Levenberg–Marquardt Algorithm: A New Optimization Strategy for Levenberg–Marquardt Neural Networks

Abstract: Engineering data are often highly nonlinear and contain high-frequency noise, so the Levenberg–Marquardt (LM) algorithm may not converge when a neural network optimized by the algorithm is trained with engineering data. In this work, we analyzed the reasons for the LM neural network’s poor convergence commonly associated with the LM algorithm. Specifically, the effects of different activation functions such as Sigmoid, Tanh, Rectified Linear Unit (RELU) and Parametric Rectified Linear Unit (PRLU) were evaluate… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
3
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
8
1
1

Relationship

0
10

Authors

Journals

citations
Cited by 26 publications
(9 citation statements)
references
References 24 publications
0
3
0
Order By: Relevance
“…This typically resulted in fitting errors during the first 1–2 min of anodization ( h PAAO < 200 nm) as shown in Figure 3a . Moreover, the LM algorithm is sensitive to the initial guess value of the fit parameters and may converge to a bad local minimum [ 33 ] resulting in a wrong thickness value.…”
Section: Resultsmentioning
confidence: 99%
“…This typically resulted in fitting errors during the first 1–2 min of anodization ( h PAAO < 200 nm) as shown in Figure 3a . Moreover, the LM algorithm is sensitive to the initial guess value of the fit parameters and may converge to a bad local minimum [ 33 ] resulting in a wrong thickness value.…”
Section: Resultsmentioning
confidence: 99%
“…The weights and biases are adjusted to minimize the MSE of the training dataset only. The neural network is implemented in the MATLAB version 2023.1 deep learning toolbox [19] in agreement to the Levenberg-Marquardt algorithm [20,21]. The algorithm is used many times, each of them referred to as an epoch) on the training dataset, measuring the MSE for each training, validation, and testing set.…”
Section: Neural Network Definition and Trainingmentioning
confidence: 99%
“…Trust-region algorithms provide more robustness compared to line-search methods. The SCG based training algorithm addresses the drawback of the line-search method by incorporating the trust-region method, similar to the approach utilized in the Levenberg-Marquardt method [47].…”
Section: Scaled Conjugate Gradient (Scg) Based Training Algorithmmentioning
confidence: 99%