Proceedings of the 23rd International Conference on Machine Learning - ICML '06 2006
DOI: 10.1145/1143844.1143865
|View full text |Cite
|
Sign up to set email alerts
|

An empirical comparison of supervised learning algorithms

Abstract: A number of supervised learning methods have been introduced in the last decade. Unfortunately, the last comprehensive empirical evaluation of supervised learning was the Statlog Project in the early 90's. We present a large-scale empirical comparison between ten supervised learning methods: SVMs, neural nets, logistic regression, naive bayes, memory-based learning, random forests, decision trees, bagged trees, boosted trees, and boosted stumps. We also examine the effect that calibrating the models via Platt … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

36
1,122
1
29

Year Published

2009
2009
2022
2022

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 1,912 publications
(1,193 citation statements)
references
References 20 publications
36
1,122
1
29
Order By: Relevance
“…The modular nature of the inference machine architecture allows us to insert any supervised learning classifier as our choice of multi-class predictor g. As the data distribution is highly multi-modal, a high-capacity nonlinear predictor is required. In this work, we use a boosted classifier [28] with random forests for the weak learners, because random forests have been empirically shown to consistently outperform other methods on several datasets [29]. We learn our boosted classifier by optimizing the non-smooth hinge loss [30].…”
Section: Methodsmentioning
confidence: 99%
“…The modular nature of the inference machine architecture allows us to insert any supervised learning classifier as our choice of multi-class predictor g. As the data distribution is highly multi-modal, a high-capacity nonlinear predictor is required. In this work, we use a boosted classifier [28] with random forests for the weak learners, because random forests have been empirically shown to consistently outperform other methods on several datasets [29]. We learn our boosted classifier by optimizing the non-smooth hinge loss [30].…”
Section: Methodsmentioning
confidence: 99%
“…Most elementary undergraduate-level statistics courses teach frequentist inference rather than Bayesian inference [12]. Bayesian method has three parts [2] : Naive Bayes classifier [13][14][15] Case-based reasoning (CBR), broadly construed, is the process of solving new problems based on the solutions of similar past problems. An auto mechanic who fixes an engine by recalling another car that exhibited similar symptoms is using case-based reasoning.…”
Section: Taxonomy Of Supervised Learning Algorithmsmentioning
confidence: 99%
“…Most elementary undergraduate-level statistics courses teach frequentist inference rather than Bayesian inference [12]. Bayesian method has three parts [2] : Naive Bayes classifier [13][14][15], Bayesian network [16], [17] and Bayesian knowledge base [17], [18].…”
Section: Taxonomy Of Supervised Learning Algorithmsmentioning
confidence: 99%
“…Additionally, there are a large number of other metrics, which are more or less appropriate, for different classes of problems [3]. In fact, real-world applications often have to meet several criteria, which would require the use of customized multi-criteria metrics [18].…”
Section: Evaluation Metricsmentioning
confidence: 99%
“…However, the most common criterion for machine learning and data mining applications is arguably classification performance. This can be illustrated by the fact that one of the largest empirical studies of machine learning metrics focuses entirely on classification performance metrics [3]. The accuracy metric is often used to measure the performance of a classifier in terms of correctness.…”
Section: Introductionmentioning
confidence: 99%