The platform will undergo maintenance on Sep 14 at about 7:45 AM EST and will be unavailable for approximately 2 hours.
2019
DOI: 10.1007/s10208-018-09409-5
|View full text |Cite
|
Sign up to set email alerts
|

Stochastic Subgradient Method Converges on Tame Functions

Abstract: This work considers the question: what convergence guarantees does the stochastic subgradient method have in the absence of smoothness and convexity? We prove that the stochastic subgradient method, on any semialgebraic locally Lipschitz function, produces limit points that are all first-order stationary. More generally, our result applies to any function with a Whitney stratifiable graph. In particular, this work endows the stochastic subgradient method, and its proximal extension, with rigorous convergence g… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

5
263
0
1

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
1

Relationship

1
5

Authors

Journals

citations
Cited by 165 publications
(269 citation statements)
references
References 40 publications
(84 reference statements)
5
263
0
1
Order By: Relevance
“…Note also that virtually all deep network architectures used in applications are actually definable, see e.g. [24]. (b) Despite our efforts we do not see any means to obtain Corollary 6 easily.…”
Section: Deep Neural Network and Nonsmooth Backpropagationmentioning
confidence: 98%
See 2 more Smart Citations
“…Note also that virtually all deep network architectures used in applications are actually definable, see e.g. [24]. (b) Despite our efforts we do not see any means to obtain Corollary 6 easily.…”
Section: Deep Neural Network and Nonsmooth Backpropagationmentioning
confidence: 98%
“…Proof : Using the chain rule characterization, all proofs boil down to providing a chain rule with the Clarke subdifferential for each of the above mentioned situation. We refer to [48] for convex, Clarke and prox regular functions, [24] for tame functions.…”
Section: Corollary 3 (Integrability and Clarke Subdifferential)mentioning
confidence: 99%
See 1 more Smart Citation
“…Our current work sits within the broader scope of analyzing subgradient and proximal methods for weakly convex problems [9, 11, 13-16, 25, 26]; see also the recent survey [12]. In particular, the paper [9] proves a global sublinear rate of convergence, in terms of a natural stationarity measure, of a (stochastic) subgradient method on any weakly convex function. In contrast, here we are interested in subgradient methods that are locally linearly convergent under the additional sharpness assumption.…”
Section: Introductionmentioning
confidence: 99%
“…Thus, one might hope to prove convergence results for a GS algorithm (with predetermined stepsizes rather than line searches) that parallel convergence theory for stochastic gradient methods. Recent work by Davis, Drusvyatskiy, Kakade and Lee [DDKL18] gives convergence results for stochastic subgradient methods on a broad class of problems.…”
Section: Discussionmentioning
confidence: 99%