2017 IEEE International Conference on Data Mining (ICDM) 2017
DOI: 10.1109/icdm.2017.18
|View full text |Cite
|
Sign up to set email alerts
|

Large Scale Kernel Methods for Online AUC Maximization

Abstract: Learning for maximizing AUC performance is an important research problem in Machine Learning and Artificial Intelligence. Unlike traditional batch learning methods for maximizing AUC which often suffer from poor scalability, recent years have witnessed some emerging studies that attempt to maximize AUC by single-pass online learning approaches. Despite their encouraging results reported, the existing online AUC maximization algorithms often adopt simple online gradient descent approaches that fail to exploit t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
29
0

Year Published

2018
2018
2020
2020

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 22 publications
(29 citation statements)
references
References 30 publications
(40 reference statements)
0
29
0
Order By: Relevance
“…In addition to these, we implement two mini-batch stochastic gradient algorithms for large scale CTR prediction problems: MB-PHL is a mini-batch gradient descent algorithm which uses PHL. A variant of this approach is also proposed in the recent work of [12]. MB-PSL is another mini-batch gradient method that uses PSL.…”
Section: Methodsmentioning
confidence: 99%
See 2 more Smart Citations
“…In addition to these, we implement two mini-batch stochastic gradient algorithms for large scale CTR prediction problems: MB-PHL is a mini-batch gradient descent algorithm which uses PHL. A variant of this approach is also proposed in the recent work of [12]. MB-PSL is another mini-batch gradient method that uses PSL.…”
Section: Methodsmentioning
confidence: 99%
“…Later, [8] use the pairwise squared loss function, which eliminates the need for buffering previous instances; [9] propose adaptive gradient/subgradient methods which can also handle sparse inputs, while [10], [11] consider the nonlinear AUC maximization problem using kernel and multiple-kernel methods. Most recently, [12] focuses on scalable kernel methods.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…For massive datasets, however, the size of the budget needs to be large to reduce the variance of the model and to achieve an acceptable accuracy, which in turns increases the training time complexity. The work [8] attempts to address the scalability problem of kernelized online AUC maximization by learning a mini-batch linear classifier on an embedded feature space. The authors explore both Nyström approximation and random Fourier features to construct an embedding in an online setting.…”
Section: Related Workmentioning
confidence: 99%
“…AUC optimization has been extensively studied over the past decades [6], [26], [27], [28]. Most of the algorithms were designed for learning classifiers for classification problems.…”
Section: Auc Optimizationmentioning
confidence: 99%