2021
DOI: 10.48550/arxiv.2112.03754
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

A Continuous-time Stochastic Gradient Descent Method for Continuous Data

Abstract: Optimization problems with continuous data appear in, e.g., robust machine learning, functional data analysis, and variational inference. Here, the target function is given as an integral over a family of (continuously) indexed target functions -integrated with respect to a probability measure. Such problems can often be solved by stochastic optimization methods: performing optimization steps with respect to the indexed target function with randomly switched indices.In this work, we study a continuous-time var… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
2
2

Relationship

1
3

Authors

Journals

citations
Cited by 4 publications
(6 citation statements)
references
References 26 publications
0
6
0
Order By: Relevance
“…We follow [52] in using Latin Hybercube Sampling [60], a quasi-random approach for space filling sampling, to obtain the collocation points for training. In our experiments, we re-sample the collocation points in every epoch to get a better coverage of the sampling domain; see also [29]. As mentioned above, the differential operator A and any derivatives in the boundary condition are evaluated using automatic differentiation [4].…”
Section: Physics-informed Neural Networkmentioning
confidence: 99%
“…We follow [52] in using Latin Hybercube Sampling [60], a quasi-random approach for space filling sampling, to obtain the collocation points for training. In our experiments, we re-sample the collocation points in every epoch to get a better coverage of the sampling domain; see also [29]. As mentioned above, the differential operator A and any derivatives in the boundary condition are evaluated using automatic differentiation [4].…”
Section: Physics-informed Neural Networkmentioning
confidence: 99%
“…Stochastic gradient descent is often computationally advantageous compared to normal gradient descent [26] due to computational efficiency as well as being able to escape local minimisers in non-convex optimisation problems [13,34]. As mentioned earlier, the theory employed in this work is based on the continuous-time analysis of stochastic gradient descent by Latz [23] that was further generalised in [21], but is somewhat orthogonal to the diffusion-based continuous-time analysis of SGD of, e.g. [25].…”
Section: Literature Overviewmentioning
confidence: 99%
“…In a setting where the differential inclusions are actually differential equations and sufficiently smooth, one can sometimes show that ( x(t)) t≥0 → (x(t)) t≥0 in a weak sense, as λ → 0. We refer to [20,23] for results of this type and a general perspective on stochastic approximation in continuous time.…”
Section: Problem Setting and Motivationmentioning
confidence: 99%
“…Indeed, we aim to show that the stochastic approximations can approximate the deterministic dynamics at any accuracy. Results of this type have been discussed in [20,23] considering smooth stochastic optimisation and in [17] considering Markov chain Monte Carlo. The theoretical foundation is given by Kushner's perturbed test function theory [21,22].…”
Section: Approximation Propertiesmentioning
confidence: 99%