2017
DOI: 10.1134/s1054661817020122
|View full text |Cite
|
Sign up to set email alerts
|

Distributed coordinate descent for generalized linear models with regularization

Abstract: Generalized linear model with L 1 and L 2 regularization is a widely used technique for solving classification, class probability estimation and regression problems.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
8
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 7 publications
(8 citation statements)
references
References 23 publications
(41 reference statements)
0
8
0
Order By: Relevance
“…Also, they do not allow the re-use of arbitrary local solvers. On the theoretical side, the convergence rate results provided by Mahajan et al (2017); Trofimov and Genkin (2016); and Yuan et al (2012) are not explicit convergence rates but only asymptotic, as the quadratic upper bounds are not explicitly controlled for safety as with our σ .…”
Section: Related Workmentioning
confidence: 82%
See 1 more Smart Citation
“…Also, they do not allow the re-use of arbitrary local solvers. On the theoretical side, the convergence rate results provided by Mahajan et al (2017); Trofimov and Genkin (2016); and Yuan et al (2012) are not explicit convergence rates but only asymptotic, as the quadratic upper bounds are not explicitly controlled for safety as with our σ .…”
Section: Related Workmentioning
confidence: 82%
“…If hypothetically each of our quadratic subproblems G σ k (∆α [k] ) as defined in (10) were to be minimized exactly, the resulting steps could be interpreted as block-wise Newton-type steps on each coordinate block k, where the Newton-subproblem is modified to also contain the L 1 -regularizer (Mahajan et al, 2017;Yuan et al, 2012;Qu et al, 2016). While Mahajan et al (2017) allows a fixed accuracy for these subproblems, but not arbitrary approximation quality Θ as in our framework, the works of Trofimov and Genkin (2016); Yuan et al (2012);and Yen et al (2015) assume that the quadratic subproblems are solved exactly. Therefore, these methods are not able to freely trade off communication and computation.…”
Section: Related Workmentioning
confidence: 99%
“…The GLM fits generalized linear models to the data by maximizing the log-likelihood. The GLM is a flexible, robust and highly interpretable model provide useful approach to modeling classification [79].…”
Section: Generalized Linear Model (Glm)mentioning
confidence: 99%
“…Line-search vs Trust-region. Line-search techniques are a popular way to guarantee convergence and they have recently been explored in distributed settings, e.g., (Hsieh et al, 2016;Lee & Chang, 2017;Trofimov & Genkin, 2017;Mahajan et al, 2017;Lee et al, 2018). Our trust-region approach has clear advantages compared to line-search methods: i) a line-search method assumes a fixed auxiliary model -which may be an arbitrarily bad approximation of the true objective-that is used to find an acceptable step size.…”
Section: Related Workmentioning
confidence: 99%