2015
DOI: 10.1007/978-3-319-26123-2_24
|View full text |Cite
|
Sign up to set email alerts
|

Distributed Coordinate Descent for L1-regularized Logistic Regression

Abstract: Solving logistic regression with L1-regularization in distributed settings is an important problem. This problem arises when training dataset is very large and cannot fit the memory of a single machine. We present d-GLMNET, a new algorithm solving logistic regression with L1-regularization in the distributed settings. We empirically show that it is superior over distributed online learning via truncated gradient.

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2015
2015
2023
2023

Publication Types

Select...
3
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(6 citation statements)
references
References 7 publications
0
6
0
Order By: Relevance
“…These methods have attracted considerable attention in the past few years, and include SCD [69], RCDM [49], UCDC [59], ICD [77], PCDM [60], SPCDM [14], SPDC [86], APCG [37], RCD [44], APPROX [15], QUARTZ [55] and ALPHA [53]. Recent advances on mini-batch and distributed variants can be found in [38], [90], [58], [16], [79], [25], [43] and [41]. Other related work includes [47,13,1,89,17,76].…”
Section: Stochastic Dual Coordinate Ascent With Adaptive Probabilitie...mentioning
confidence: 99%
“…These methods have attracted considerable attention in the past few years, and include SCD [69], RCDM [49], UCDC [59], ICD [77], PCDM [60], SPCDM [14], SPDC [86], APCG [37], RCD [44], APPROX [15], QUARTZ [55] and ALPHA [53]. Recent advances on mini-batch and distributed variants can be found in [38], [90], [58], [16], [79], [25], [43] and [41]. Other related work includes [47,13,1,89,17,76].…”
Section: Stochastic Dual Coordinate Ascent With Adaptive Probabilitie...mentioning
confidence: 99%
“…Training algorithms that can be distributed across multiple machines have been the subject of a significant amount of research. Distributed techniques based on stochastic gradient descent have been proposed (see [17] and [18]) as well as methods based on coordinate descent/ascent (see [7], [19], [20], [21] and [22]). These distributed learning algorithms typically involve each machine (or worker) performing a number of optimization steps to approximately minimize the global objective function using the local data that it has available.…”
Section: Distributed Stochastic Learningmentioning
confidence: 99%
“…The convergence behavior of the distributed SCD algorithm can be improved by optimizing the aggregation step. Existing work has considered both averaging and adding of updates [24], introducing an aggregation parameter that can be set freely [25] and even performing a line search method to explicitly optimize the aggregation parameter [21]. We propose a new method to optimize aggregation for distributed ridge regression whereby an optimal value of an aggregation parameter is precisely computed in a distributed manner.…”
Section: B Adaptive Aggregationmentioning
confidence: 99%
“…Inspired by GLMNET and [34], the work of [3,18] introduced the idea of a block-diagonal Hessian upper approximation in the distributed L 1 context. The later work of [29] specialized this approach to sparse logistic regression.…”
Section: Related Workmentioning
confidence: 99%
“…If hypothetically each of our quadratic subproblems G σ k (∆α [k] ) as defined in (2) were to be minimized exactly, the resulting steps could be interpreted as block-wise Newton-type steps on each coordinate block k, where the Newton-subproblem is modified to also contain the L 1 -regularizer [18,34,23]. While [18] allows a fixed accuracy for these subproblems-but not arbitrary approximation quality Θ as in our framework-the work of [29,34,31] assumes that the quadratic subproblems are solved exactly. Therefore, these methods are not able to freely trade off communication and computation.…”
Section: Related Workmentioning
confidence: 99%