Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence 2019
DOI: 10.24963/ijcai.2019/328
|View full text |Cite
|
Sign up to set email alerts
|

Scalable Semi-Supervised SVM via Triply Stochastic Gradients

Abstract: Semi-supervised learning (SSL) plays an increasingly important role in the big data era because a large number of unlabeled samples can be used effectively to improve the performance of the classifier. Semi-supervised support vector machine (S 3 VM) is one of the most appealing methods for SSL, but scaling up S 3 VM for kernel learning is still an open problem. Recently, a doubly stochastic gradient (DSG) algorithm has been proposed to achieve efficient and scalable training for kernel methods. However, the al… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
5
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
5
2

Relationship

2
5

Authors

Journals

citations
Cited by 11 publications
(5 citation statements)
references
References 1 publication
0
5
0
Order By: Relevance
“…Since our proposed method contains three sources of randomness, we denote our method as Triply Stochastic Gradient descent for SU Classification (TSGSU). Theoretically, we give a new theoretically analysis based on the framework in Geng et al (2019), Dai et al (2014) and prove that our method can converge to the stationary point at the rate of O( 1 √ T ) after T iterations. Our experiments on various benchmark datasets and high-dimensional datasets not only demonstrate the scalability but also show the efficiency of TSGSU compared with existing learning algorithms while retaining similar generalization performance.…”
Section: Introductionmentioning
confidence: 83%
See 2 more Smart Citations
“…Since our proposed method contains three sources of randomness, we denote our method as Triply Stochastic Gradient descent for SU Classification (TSGSU). Theoretically, we give a new theoretically analysis based on the framework in Geng et al (2019), Dai et al (2014) and prove that our method can converge to the stationary point at the rate of O( 1 √ T ) after T iterations. Our experiments on various benchmark datasets and high-dimensional datasets not only demonstrate the scalability but also show the efficiency of TSGSU compared with existing learning algorithms while retaining similar generalization performance.…”
Section: Introductionmentioning
confidence: 83%
“…It uses pseudo-random number generators to generate the random features on the fly, which highly reduces the memory requirement of RFF. Due to its superior performance, DSG has been successfully applied to scale up kernelbased algorithms in many applications, such as (Gu et al 2018b;Li et al 2017;Rahimi and Recht 2009;Le et al 2013;Shi et al 2019;Geng et al 2019). The theoretical analysis of Dai et al (2014), Gu et al (2018b), Li et al (2017), Shi et al (2019) are all based on the assumption that the objective functions of these problems are convex.…”
Section: Kernel Approximationmentioning
confidence: 99%
See 1 more Smart Citation
“…However, solving the S 3 V M problem is very challenging for its nonconvexity and computational cost [23]. It is still an open problem to scale up S 3 V M for large-scale applications.…”
Section: Label Recommendationmentioning
confidence: 99%
“…However, RFF method needs to save large amounts of random features. Instead of saving all the random features, Dai et al, (2014) proposed DSG algorithm to use pseudorandom number generators to generate the random features on-the-fly, which has been widely used (Shi et al 2019;Geng et al 2019;Li et al 2017). Our method can be viewed as an extension of (Shi et al 2019).…”
Section: Kernel Approximationmentioning
confidence: 99%