2012
DOI: 10.1137/120865094
|View full text |Cite
|
Sign up to set email alerts
|

Agnostic Learning of Monomials by Halfspaces Is Hard

Abstract: We prove the following strong hardness result for learning: Given a distribution of labeled examples from the hypercube such that there exists a monomial consistent with (1 − ǫ) of the examples, it is NP-hard to find a halfspace that is correct on (1/2 + ǫ) of the examples, for arbitrary constants ǫ > 0. In learning theory terms, weak agnostic learning of monomials is hard, even if one is allowed to output a hypothesis from the much bigger concept class of halfspaces. This hardness result subsumes a long line … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

3
83
0

Year Published

2012
2012
2020
2020

Publication Types

Select...
4
3
2

Relationship

1
8

Authors

Journals

citations
Cited by 96 publications
(86 citation statements)
references
References 63 publications
3
83
0
Order By: Relevance
“…On the contrary, the use of a nonconvex loss leads to NP-problems, which cannot be exactly solved for sample sets whose cardinality exceeds few tens of data (e.g. n > 30) (Anthony 2001;Feldman et al 2009), but for which approximate solutions can be eventually found (Lawler and Wood 1966;Yuille and Rangarajan 2003). As a matter of fact, if one has to cope with a non-convex loss, a convex relaxation is often used in order to reformulate the problem so to make it computationally tractable.…”
Section: The Supervised Learning Frameworkmentioning
confidence: 99%
“…On the contrary, the use of a nonconvex loss leads to NP-problems, which cannot be exactly solved for sample sets whose cardinality exceeds few tens of data (e.g. n > 30) (Anthony 2001;Feldman et al 2009), but for which approximate solutions can be eventually found (Lawler and Wood 1966;Yuille and Rangarajan 2003). As a matter of fact, if one has to cope with a non-convex loss, a convex relaxation is often used in order to reformulate the problem so to make it computationally tractable.…”
Section: The Supervised Learning Frameworkmentioning
confidence: 99%
“…Typically, the α-error is regarded much worse than the β-error. Unfortunately, empirical risk minimization in the case of the hypothesis class of DNFs is NP-hard because this is already the case for the subclass of monomials (Feldman et al, 2012). Thus, heuristics come into play.…”
Section: Threshold Heuristic For Learning Dnfsmentioning
confidence: 99%
“…It was first introduced in [21] for proving hardness results in hypergraph coloring and subsequently utilized for other applications in [25] [26] and [16].…”
Section: Np-hardness Reductionsmentioning
confidence: 99%
“…A key ingredient in our reductions is the smooth version of Label Cover which enables us to devise a more sophisticated decoding procedure which can be combined with the dictatorship test. It is pertinent to note that a couple of the (few) previous results using smooth versions of Label Cover, on hardness of learning intersection of halfspaces [26] and monomials [16], have also used analysis based on versions of the Central Limit Theorem. This work suggests the possibility of obtaining optimal NP-hardness results for other combinatorial optimization problems using reductions based on smooth versions of Label Cover.…”
Section: Introductionmentioning
confidence: 99%