2014 IEEE International Conference on Big Data (Big Data) 2014
DOI: 10.1109/bigdata.2014.7004354
|View full text |Cite
|
Sign up to set email alerts
|

Boosting Stochastic Newton Descent for Bigdata large scale classification

Abstract: Efficient Bigdata classification requires low cost learning methods. A standard approach involves Stochastic Gradient Descent algorithm (SGD) for the minimization of the Hinge Loss in the primal space. Although complexity of Stochastic Gradient Descent is linear with the number of samples these method suffers from slow convergence. In order to cope with this issue, we propose here a Boosting Stochastic Newton Descent (BSND) method for minimization of any calibrated loss in the primal space. BSND approximates t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
2
0

Year Published

2017
2017
2017
2017

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(2 citation statements)
references
References 22 publications
(19 reference statements)
0
2
0
Order By: Relevance
“…Gradient based methods are computationally more efficient compared to other families of online learning algorithms, i.e., for a sequence of M -dimensional feature vectors {x t } t≥0 , where x t ∈ Ê M , the computational complexity is only in the order of O(M ). However, their convergence rates remain significantly slow when achieving an optimal solution, since no statistics other than the gradient is used [3], [16], [19]. Even though gradient based methods suffer from this convergence issue, they are extensively used in big data applications due to such low computational demand [20].…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…Gradient based methods are computationally more efficient compared to other families of online learning algorithms, i.e., for a sequence of M -dimensional feature vectors {x t } t≥0 , where x t ∈ Ê M , the computational complexity is only in the order of O(M ). However, their convergence rates remain significantly slow when achieving an optimal solution, since no statistics other than the gradient is used [3], [16], [19]. Even though gradient based methods suffer from this convergence issue, they are extensively used in big data applications due to such low computational demand [20].…”
Section: Introductionmentioning
confidence: 99%
“…T ECHNOLOGICAL developments in recent years have substantially increased the amount of data gathered from real life systems [1], [2], [3], [4]. There exists a significant data flow through the recently arising applications such as large-scale sensor networks, information sensing mobile devices and web based social networks [5], [6], [7].…”
Section: Introductionmentioning
confidence: 99%