2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2016
DOI: 10.1109/icassp.2016.7472156
|View full text |Cite
|
Sign up to set email alerts
|

Adaptive sequential optimization with applications to machine learning

Abstract: A framework is introduced for solving a sequence of slowly changing optimization problems, including those arising in regression and classification applications, using optimization algorithms such as stochastic gradient descent (SGD). The optimization problems change slowly in the sense that the minimizers change at either a fixed or bounded rate. A method based on estimates of the change in the minimizers and properties of the optimization algorithm is introduced for adaptively selecting the number of samples… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
30
0

Year Published

2016
2016
2019
2019

Publication Types

Select...
3
2

Relationship

3
2

Authors

Journals

citations
Cited by 8 publications
(30 citation statements)
references
References 27 publications
0
30
0
Order By: Relevance
“…Since {ρ i } are random variables, if we combine them by taking their maximum, any particular one step estimateρ i that is large will pull up the overall estimateρ n . This would driveρ n → diam(X ), as n → ∞, resulting in aρ n that is trivial in the limit of large n. 1 Note that a choice of t n that is no greater than 1/ √ n − 1 works here.…”
Section: ) Combining Onementioning
confidence: 98%
See 4 more Smart Citations
“…Since {ρ i } are random variables, if we combine them by taking their maximum, any particular one step estimateρ i that is large will pull up the overall estimateρ n . This would driveρ n → diam(X ), as n → ∞, resulting in aρ n that is trivial in the limit of large n. 1 Note that a choice of t n that is no greater than 1/ √ n − 1 works here.…”
Section: ) Combining Onementioning
confidence: 98%
“…To apply cross-validation to our framework, we run C parallel versions of our approach and at time n we generate C different choices for the number of samples K (i) n . We then choose K n = max{K (1) n , . .…”
Section: B Cross Validationmentioning
confidence: 99%
See 3 more Smart Citations