2010
DOI: 10.1137/090759574
|View full text |Cite
|
Sign up to set email alerts
|

Trading Accuracy for Sparsity in Optimization Problems with Sparsity Constraints

Abstract: Abstract. We study the problem of minimizing the expected loss of a linear predictor while constraining its sparsity, i.e., bounding the number of features used by the predictor. While the resulting optimization problem is generally NP-hard, several approximation algorithms are considered. We analyze the performance of these algorithms, focusing on the characterization of the trade-off between accuracy and sparsity of the learned predictor in different scenarios. Key words. sparsity, linear prediction AMS subj… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

5
126
0

Year Published

2013
2013
2022
2022

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 123 publications
(134 citation statements)
references
References 30 publications
(61 reference statements)
5
126
0
Order By: Relevance
“…It is pointed out in [46] that in many engineering applications, researchers are interested in an approximate solution of problem (2) as a linear combination of a few elements from a given system D of elements. There is an increasing interest in building such sparse approximate solutions using different greedy-type algorithms (see, for instance, [1,[3][4][5]11,12,17,19,[22][23][24]34,45,46]). We refer the reader to the papers [1] and [23] for concise surveys of recent results on greedy algorithms from the point of view of convex optimization and signal processing.…”
Section: Introductionmentioning
confidence: 99%
See 2 more Smart Citations
“…It is pointed out in [46] that in many engineering applications, researchers are interested in an approximate solution of problem (2) as a linear combination of a few elements from a given system D of elements. There is an increasing interest in building such sparse approximate solutions using different greedy-type algorithms (see, for instance, [1,[3][4][5]11,12,17,19,[22][23][24]34,45,46]). We refer the reader to the papers [1] and [23] for concise surveys of recent results on greedy algorithms from the point of view of convex optimization and signal processing.…”
Section: Introductionmentioning
confidence: 99%
“…It is used for exact recovery of sparse signals and for approximation of signals by sparse ones. An analog of the WCGA in convex optimization was introduced in [34] under the name Fully Corrective Forward Greedy Selection. The WRGA is the approximation theory analog of the classical Frank-Wolfe algorithm, introduced in [21] and studied in many papers (see, for instance, [12,14,17,18,22,23]).…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…Existing literatures [15], [16] indicates that in a classical classification environment, when 1 regularizer is applied, there will always be a trade-off between the enforced sparsity and the accuracy obtained. However, depending on the dataset itself, the overall performance of an 1 regularized classifier can sometimes beat an 2 regularized classifier [17] in terms of both bias and variance.…”
Section: Why 1 Regularizer For Data Miner?mentioning
confidence: 99%
“…A popular method to solve this problem is through decision tree learning (such as CART [6] and C4.5 [8]), which has an important advantage for handling heterogenous data with case when different features come from different sources.…”
Section: Tree-based Modelsmentioning
confidence: 99%