2023
DOI: 10.1109/tai.2022.3208223
|View full text |Cite
|
Sign up to set email alerts
|

AdaInject: Injection-Based Adaptive Gradient Descent Optimizers for Convolutional Neural Networks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
8
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
4
1
1

Relationship

0
6

Authors

Journals

citations
Cited by 8 publications
(8 citation statements)
references
References 29 publications
0
8
0
Order By: Relevance
“…As detailed in the literature (e.g. [15]), an ideal parameter optimization method should follow the rules depicted in Figure 4, where Θ is the parameters tensor, delta is the difference of the parameters between two training iterations (see Eq. 21), g are the gradients.…”
Section: Discussion On Adam Variantsmentioning
confidence: 99%
See 1 more Smart Citation
“…As detailed in the literature (e.g. [15]), an ideal parameter optimization method should follow the rules depicted in Figure 4, where Θ is the parameters tensor, delta is the difference of the parameters between two training iterations (see Eq. 21), g are the gradients.…”
Section: Discussion On Adam Variantsmentioning
confidence: 99%
“…Angular Injection (AI) optimizer is based on AngularGrad [39] and injection [15]. It generates a score to control the step size based on the gradient angular information of previous iterations.…”
Section: B: Mind Optimizermentioning
confidence: 99%
“…RAdam [21] computes the variance of the adaptive learning rate and uses it to provide stability in the training process through corrections in the learning formula. AdaInject [22] controls parameter updates to minimize oscillations closer to the minimum. Recently, there has been research on PNM [23] and AdaPNM [23], aiming to replace the traditional momentum with the Positive-Negative momentum approach.…”
Section: Related Work a Sgd-based Optimization Methodsmentioning
confidence: 99%
“…There were made experiments of minimization of Rastrigin and Rosenbrock functions (https://github.com/jettify/pytorchoptimizer), where AdaBound achieved the highest accuracy, converging in the neighborhood of global minimum. But such approach is too complex for optimization and there exists much simple method, which is called AdamInject [64]. It reduces time consumption, preserving convergence rate.…”
Section: Adam-type Algorithmsmentioning
confidence: 99%