Proceedings of the Eleventh Annual Conference on Computational Learning Theory 1998
DOI: 10.1145/279943.279985
|View full text |Cite
|
Sign up to set email alerts
|

Large margin classification using the perceptron algorithm

Abstract: We introduce and analyze a new algorithm for linear classification whichcombines Rosenblatt'sperceptron algorithm with Helmbold and Warmuth's leave-one-out method. Like Vapnik's maximalmargin classifier, our algorithm takes advantage of data that are linearly separable with large margins. Compared to Vapnik's algorithm, however, ours is much simpler to implement, and much more efficient in terms of computation time. We also show that our algorithm can be efficiently used in very high dimensional spaces using k… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

5
589
0
14

Year Published

2002
2002
2020
2020

Publication Types

Select...
4
3
3

Relationship

0
10

Authors

Journals

citations
Cited by 603 publications
(608 citation statements)
references
References 11 publications
5
589
0
14
Order By: Relevance
“…However such ad-hoc procedures of forcing convergence lead to bias in the final parameters. In the oscillatory case, one can choose any of the parameter selection heuristics commonly used in perceptron learning where convergence is also not guaranteed, e.g., the voted perceptron [14] [13]. In this work we simply used majority vote parameter setting, i.e., the parameters for which the training error was minimum.…”
Section: Experimental Observations: Parameter Learningmentioning
confidence: 99%
“…However such ad-hoc procedures of forcing convergence lead to bias in the final parameters. In the oscillatory case, one can choose any of the parameter selection heuristics commonly used in perceptron learning where convergence is also not guaranteed, e.g., the voted perceptron [14] [13]. In this work we simply used majority vote parameter setting, i.e., the parameters for which the training error was minimum.…”
Section: Experimental Observations: Parameter Learningmentioning
confidence: 99%
“…We compared the performances of our two-stage MB classifier with those of four widely used classifiers: a naïve Bayes classifier based on the multivariate Bernoulli distribution with Laplace prior for unseen words, discussed in Nigam et al [21], a support vector machine (SVM) classifier, discussed by Joachims [22], an implementation of the voted Perceptron, discussed in Freund and Schapire [23], and a maximum entropy conditional random field learner, introduced by Lafferty et al [24].…”
Section: Results and Analysismentioning
confidence: 99%
“…the minimal distance from any instance to the separating hyperplane. Freund and Schapire (1998) generalized this result to the inseparable case. The maximum-margin algorithm uses quadratic programming to find the weight vector that classifies all the training data correctly and maximizes the margin.…”
Section: Introductionmentioning
confidence: 87%