2022
DOI: 10.3934/fods.2021037
|View full text |Cite
|
Sign up to set email alerts
|

Smart Gradient - An adaptive technique for improving gradient estimation

Abstract: <p style='text-indent:20px;'>Computing the gradient of a function provides fundamental information about its behavior. This information is essential for several applications and algorithms across various fields. One common application that requires gradients are optimization techniques such as stochastic gradient descent, Newton's method and trust region methods. However, these methods usually require a numerical computation of the gradient at every iteration of the method which is prone to numerical err… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
8
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
3

Relationship

5
3

Authors

Journals

citations
Cited by 10 publications
(8 citation statements)
references
References 8 publications
(17 reference statements)
0
8
0
Order By: Relevance
“…Latent Gaussian models are a broad class of models such as Gaussian processes for time series, spatial and spatio-temporal data or clustered random effects models for data with a grouping structure. The formulation given in (1) can be seen as any model where each one of the f k (.) terms can be written in matrix form as A A A k u u u k .…”
Section: Latent Gaussian Modelmentioning
confidence: 99%
See 2 more Smart Citations
“…Latent Gaussian models are a broad class of models such as Gaussian processes for time series, spatial and spatio-temporal data or clustered random effects models for data with a grouping structure. The formulation given in (1) can be seen as any model where each one of the f k (.) terms can be written in matrix form as A A A k u u u k .…”
Section: Latent Gaussian Modelmentioning
confidence: 99%
“…that includes the linear predictors as the first n elements of X X X and the m model components as defined in (1). This augmentation results in a singular covariance matrix of X X X since η η η is a deterministic function of the other elements of X X X .…”
Section: Classic Model Formulation Of Inlamentioning
confidence: 99%
See 1 more Smart Citation
“…Often the directional derivatives are computed according to this canonical basis, however, noncanonical bases can just as well be used. INLA makes use of knowledge from previous iterations to choose directional derivatives exhibiting more robust numerical properties and hence faster overall convergence, see [16] for details. Independently of the choice of basis, the directional derivatives are computed for each component of θ and each time entail one or two function evaluations of f .…”
Section: Parallelization Of Function Evaluationsmentioning
confidence: 99%
“…The estimated gradient ∇π(θ θ θ|y y y) at each iteration for this algorithm and the estimated Hessian ∇ 2 π(θ θ θ|y y y) of π(θ θ θ|y y y) at θ θ θ * are approximated using numerical differentiation methods boosted by the Smart Gradient and Hessian Techniques [40]. We need to correct for the constraints at every configuration of the hyperparameter θ θ θ. INLA computes uncorrected mean x x x * un and uncorrected covariance Σ Σ Σ * un using sparse solvers, then it uses kriging technique to find x x x * and Σ Σ Σ * .…”
Section: Appendicesmentioning
confidence: 99%