2017
DOI: 10.48550/arxiv.1708.07821
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

$\ell_1$ Regression using Lewis Weights Preconditioning and Stochastic Gradient Descent

Abstract: We present preconditioned stochastic gradient descent (SGD) algorithms for the ℓ 1 minimization problem min x Ax − b 1 in the overdetermined case, where there are far more constraints than variables. Specifically, we have A ∈ R n×d for n ≫ d. Commonly known as the Least Absolute Deviations problem, ℓ 1 regression can be used to solve many important combinatorial problems, such as minimum cut and shortest path. SGD-based algorithms are appealing for their simplicity and practical efficiency. Our primary insight… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
4
0

Year Published

2020
2020
2020
2020

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(4 citation statements)
references
References 14 publications
(20 reference statements)
0
4
0
Order By: Relevance
“…On the other hand, the complexity of first-order methods usually has dependences on various parameters of the input matrix A, which can be unbounded in the worst case. Fortunately, recent developments in ℓ 1 regression [32] show how to precondition the matrix A by simply doing an ℓ 1 Lewis weights sampling, and then rotating the matrix appropriately. By carefully combining this preconditioning procedure with Accelerated Gradient Descent, we obtain an algorithm for (1+ε)-approximate ℓ 1 regression with communication complexity O(sd 3 L/ε) in the coordinator model, which shows it is indeed possible to improve the ε dependence for ℓ 1 regression.…”
Section: Linear Regressionmentioning
confidence: 99%
See 3 more Smart Citations
“…On the other hand, the complexity of first-order methods usually has dependences on various parameters of the input matrix A, which can be unbounded in the worst case. Fortunately, recent developments in ℓ 1 regression [32] show how to precondition the matrix A by simply doing an ℓ 1 Lewis weights sampling, and then rotating the matrix appropriately. By carefully combining this preconditioning procedure with Accelerated Gradient Descent, we obtain an algorithm for (1+ε)-approximate ℓ 1 regression with communication complexity O(sd 3 L/ε) in the coordinator model, which shows it is indeed possible to improve the ε dependence for ℓ 1 regression.…”
Section: Linear Regressionmentioning
confidence: 99%
“…), we use Lemma 29 in [32], which states that with constant probability, the leverage scores of SA satisfy τ i (SA) = O(d/N ) for all i. Since leverage scores are invariant under change of basis (see, e.g., Section 2.4 in [65]), we have for all i,…”
Section: Use the Protocol Inmentioning
confidence: 99%
See 2 more Smart Citations