Proceedings of the Twelfth Annual Conference on Computational Learning Theory 1999
DOI: 10.1145/307400.307405
|View full text |Cite
|
Sign up to set email alerts
|

The robustness of the p -norm algorithms

Abstract: We consider two on-line learning frameworks: binary classification through linear threshold functions and linear regression. We study a family of on-line algorithms, called p-norm algorithms, introduced by Grove, Littlestone and Schuurmans in the context of deterministic binary classification. We show how to adapt these algorithms for use in the regression setting, and prove worst-case bounds on the square loss, using a technique from Kivinen and Warmuth. As pointed out by Grove, et al., these algorithms can b… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

5
133
0

Year Published

2001
2001
2008
2008

Publication Types

Select...
5
4

Relationship

1
8

Authors

Journals

citations
Cited by 86 publications
(138 citation statements)
references
References 18 publications
5
133
0
Order By: Relevance
“…Grove, Littlestone, and Schuurmans (1997) have generalized this by giving for each p > 2 a classification algorithm that has a loss bound in terms of on the p-norm of the instances and the q-norm of the correct weight vector, where 1/ p + 1/q = 1. This result for general norm pairs has also been extended to linear regression (Gentile & Littlestone, 1999).…”
Section: Let Us Define Lossmentioning
confidence: 75%
See 1 more Smart Citation
“…Grove, Littlestone, and Schuurmans (1997) have generalized this by giving for each p > 2 a classification algorithm that has a loss bound in terms of on the p-norm of the instances and the q-norm of the correct weight vector, where 1/ p + 1/q = 1. This result for general norm pairs has also been extended to linear regression (Gentile & Littlestone, 1999).…”
Section: Let Us Define Lossmentioning
confidence: 75%
“…In work parallel to this, the general additive update (9) in the context of linear classification, i.e., with a thresholded transfer function, has recently been developed and analyzed by Grove, Littlestone, and Schuurmans (1997) with methods and results very similar to ours; see also Gentile and Littlestone (1999). Gentile and Warmuth (1999) have shown how the notion of matching loss can be generalized to thresholded transfer functions.…”
Section: Introductionmentioning
confidence: 78%
“…For example, a conjunctive query such as ∃x∃z (Mother(Ann, x) ∧ Gene(x, z, a)) can be transformed into a linear number of non-overlapping decomposable conjunctions; model counting can hence be evaluated in polynomial time. A second direction is to explore other learning strategies such as, for example, quasi-additive algorithms (Grove et al 2001;Gentile 2003) which might open the door to new learnability results. A third direction is to examine other types of learning interfaces.…”
Section: Discussionmentioning
confidence: 99%
“…(This choice of p was also suggested by Gentile (2001) in the context of p-norm perceptron algorithms.) A similar bound can be obtained, under the same assumption on the r t 's, by setting η = √ 2 ln N /t in the exponential potential.…”
Section: Corollarymentioning
confidence: 97%