2008
DOI: 10.1016/j.jcss.2007.04.011
|View full text |Cite
|
Sign up to set email alerts
|

The complexity of properly learning simple concept classes

Abstract: We consider the complexity of properly learning concept classes, i.e. when the learner must output a hypothesis of the same form as the unknown concept. We present the following new upper and lower bounds on well-known concept classes:• We show that unless NP = RP, there is no polynomial-time PAC learning algorithm for DNF formulas where the hypothesis is an OR-of-thresholds. Note that as special cases, we show that neither DNF nor OR-of-thresholds are properly learnable unless NP = RP. Previous hardness resul… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
31
0

Year Published

2009
2009
2020
2020

Publication Types

Select...
5
2

Relationship

1
6

Authors

Journals

citations
Cited by 35 publications
(35 citation statements)
references
References 31 publications
0
31
0
Order By: Relevance
“…Valiant's original definition required that the learning algorithm output a DNF expression but this restriction was later relaxed to any efficiently-evaluatable hypothesis with the stricter version being referred to as proper learning. All these variants of the DNF learning question remained open until a recent result by Alekhnovich et al that established NP-hardness of the hardest variant: proper learning from random examples only [2]. Building on their proof, we resolve one more of Valiant's questions: Theorem 1.2 (Informal).…”
mentioning
confidence: 84%
See 1 more Smart Citation
“…Valiant's original definition required that the learning algorithm output a DNF expression but this restriction was later relaxed to any efficiently-evaluatable hypothesis with the stricter version being referred to as proper learning. All these variants of the DNF learning question remained open until a recent result by Alekhnovich et al that established NP-hardness of the hardest variant: proper learning from random examples only [2]. Building on their proof, we resolve one more of Valiant's questions: Theorem 1.2 (Informal).…”
mentioning
confidence: 84%
“…These results were strengthened by Nock et al [29] who proved similar hardness even when learning by formulas of size k α n β (where α 2 and β is any constant). Finally, Alekhnovich et al removed any bounds on the size of the hypothesis (other than those naturally imposed by the polynomial running time of the learning algorithm) [2]. Angluin and Kharitonov prove that if non-uniform one-way functions exist then MQs do not help predicting DNF formulae [5].…”
Section: Relation To Other Workmentioning
confidence: 99%
“…However, a result such as Theorem 1 was not known for learning intersections of (two) halfspaces. Blum and Rivest [8] showed that it is NP-hard to learn an intersection of two halfspaces with an intersection of two halfspaces, and Alekhnovich, Braverman, Feldman, Klivans and Pitassi [1] proved a similar result even when the hypothesis is an intersection of halfspaces for any constant . Both the results are only NP-hardness results and do not prove APX-hardness for the underlying optimization problem.…”
Section: Previous Workmentioning
confidence: 96%
“…Note that the imperfect completeness is necessary in the above theorem, since via linear programming, one can always efficiently find a halfspace that correctly classifies all the points, if one exists. The theorem is optimal, since one can easily classify 1 2 fraction of the data points correctly, by taking an arbitrary halfspace or its complement as a hypothesis. From the learning theory perspective, such an optimal hardness result is especially satisfying, since if one could efficiently find ( 1 2 + ε)-consistent hypothesis (i.e.…”
Section: Previous Workmentioning
confidence: 99%
“…While there has been significant recent progress on the problem, including that random DNF are learnable under the uniform distribution [12,24,25], virtually nothing is known about their learnability in the worst case, even in the classical noiseless PAC model. In fact, there are serious impediments to learning DNF, including hardness results for their proper learnability [1]. The previous best algorithm for learning s-term DNF from random examples is due to Verbeurgt [27] and runs in quasipolynomial time O(n log s ), where is the error rate.…”
Section: Implications To Related Problemsmentioning
confidence: 99%