2013
DOI: 10.1007/978-3-642-40994-3_6
|View full text |Cite
|
Sign up to set email alerts
|

Bundle CDN: A Highly Parallelized Approach for Large-Scale ℓ1-Regularized Logistic Regression

Abstract: Abstract. Parallel coordinate descent algorithms emerge with the growing demand of large-scale optimization. In general, previous algorithms are usually limited by their divergence under high degree of parallelism (DOP), or need data pre-process to avoid divergence. To better exploit parallelism, we propose a coordinate descent based parallel algorithm without needing of data pre-process, termed as Bundle Coordinate Descent Newton (BCDN), and apply it to large-scale ℓ1-regularized logistic regression. BCDN fir… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2015
2015
2019
2019

Publication Types

Select...
2
1
1

Relationship

1
3

Authors

Journals

citations
Cited by 4 publications
(5 citation statements)
references
References 8 publications
0
5
0
Order By: Relevance
“…Experimental results show that the proposed PCDN algorithm performs favorably against the other methods. As a feature-based parallel algorithm, the proposed PCDN solver performs well for sparse datasets with more features as shown by the results on the rcv1 and news20 datasets, which are very sparse (training data sparsity, defined by the ratio of zero elements in design matrix X and explained in Table 2, is 99.85% and 99.97%, respectively) with a large number of features (47,236 and 1,355,191). In such cases, the PCDN algorithm performs well against the TRON method.…”
Section: Methodsmentioning
confidence: 90%
See 3 more Smart Citations
“…Experimental results show that the proposed PCDN algorithm performs favorably against the other methods. As a feature-based parallel algorithm, the proposed PCDN solver performs well for sparse datasets with more features as shown by the results on the rcv1 and news20 datasets, which are very sparse (training data sparsity, defined by the ratio of zero elements in design matrix X and explained in Table 2, is 99.85% and 99.97%, respectively) with a large number of features (47,236 and 1,355,191). In such cases, the PCDN algorithm performs well against the TRON method.…”
Section: Methodsmentioning
confidence: 90%
“…Proof. (1) We first prove that E B t [ λ(B t )] is monotonically increasing with respect to P and E B t [ λ(B t )] is constant with respect to P , if λ i is constant or…”
Section: Discussionmentioning
confidence: 99%
See 2 more Smart Citations
“…Some past works have considered using diagonal blocks as the approximation of the Hessian. For logistic regression, Bian et al (2013) consider diagonal elements of the Hessian to solve several one-variable sub-problems in parallel. Mahajan et al (2017) study a more general setting in which using diagonal blocks is a special case.…”
Section: Diagonal Gauss-newton Matrix Approximationmentioning
confidence: 99%