2010
DOI: 10.1007/s10107-010-0420-4
|View full text |Cite
|
Sign up to set email alerts
|

Pegasos: primal estimated sub-gradient solver for SVM

Abstract: We describe and analyze a simple and effective stochastic sub-gradient descent algorithm for solving the optimization problem cast by Support Vector Machines (SVM). We prove that the number of iterations required to obtain a solution of accuracy isÕ(1/ ), where each iteration operates on a single training example. In contrast, previous analyses of stochastic gradient descent methods for SVMs require Ω(1/ 2 ) iterations. As in previously devised SVM solvers, the number of iterations also scales linearly with 1/… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

6
1,154
0
7

Year Published

2010
2010
2021
2021

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 1,243 publications
(1,167 citation statements)
references
References 27 publications
(47 reference statements)
6
1,154
0
7
Order By: Relevance
“…It can be checked easily that the above chosen u and v, together with α satisfy the KKT condition (17).…”
Section: Algorithm 4 the Practical Second-order Working Set Selectionmentioning
confidence: 99%
See 1 more Smart Citation
“…It can be checked easily that the above chosen u and v, together with α satisfy the KKT condition (17).…”
Section: Algorithm 4 the Practical Second-order Working Set Selectionmentioning
confidence: 99%
“…The existing algorithms can be broadly categorized into two categories: the cutting-plane methods [11,5,12,13,25], and subgradient methods [3,17]. For example, in [17], Shalev-Shwartz et al described and analyzed a simple and effective stochastic sub-gradient descent algorithm and prove that the number of iterations required to obtain a solution of accuracy is O(1/ ). Generally speaking, without counting the loading time, these recent advances on linear classification have shown that training one million instances takes only a few seconds [22].…”
Section: Introductionmentioning
confidence: 99%
“…We used a primal projected sub-gradient algorithm [19], and tuned the regularization constants on the validation data. The mixture model took typically less than 10 global iterations to converge.…”
Section: Comparison Of Modelsmentioning
confidence: 99%
“…This corresponds to a ranking SVM [11]. One simple strategy to minimize this objective is to use a primal sub-gradient method [19], which is the approach we use in this paper.…”
Section: Introductionmentioning
confidence: 99%
“…However, the problem with bottom-q sketches, is that the samples lose their alignment. In applications of large-scale machine learning this alignment is needed in order to efficiently construct a dot-product for use with a linear support vector machine (SVM) 4 such as LIBLINEAR [9] or Pegasos [19]. Using the alignment of kˆminwise, it was shown how to construct such a dot-product in [13] based on this scheme.…”
Section: Introductionmentioning
confidence: 99%