Publication | Closed Access
An empirical comparison of supervised learning algorithms
2.7K
Citations
19
References
2006
Year
Unknown Venue
EngineeringMachine LearningClassification MethodData ScienceData MiningPattern RecognitionPlatt ScalingDecision Tree LearningStatisticsSupervised LearningPredictive AnalyticsSupervised Learning AlgorithmsKnowledge DiscoveryIntelligent ClassificationComputer ScienceStatistical Learning TheoryData ClassificationClassifier SystemDecision Trees
Supervised learning methods have proliferated over the last decade, yet the most recent comprehensive empirical evaluation was the Statlog Project of the early 1990s. The study aims to conduct a large-scale empirical comparison of ten supervised learning methods, including SVMs, neural nets, logistic regression, naive Bayes, memory-based learning, random forests, decision trees, bagged trees, boosted trees, and boosted stumps. The authors evaluate these methods across a variety of performance criteria, assess the effect of calibrating models with Platt Scaling and Isotonic Regression, and examine how calibration influences performance.
A number of supervised learning methods have been introduced in the last decade. Unfortunately, the last comprehensive empirical evaluation of supervised learning was the Statlog Project in the early 90's. We present a large-scale empirical comparison between ten supervised learning methods: SVMs, neural nets, logistic regression, naive bayes, memory-based learning, random forests, decision trees, bagged trees, boosted trees, and boosted stumps. We also examine the effect that calibrating the models via Platt Scaling and Isotonic Regression has on their performance. An important aspect of our study is the use of a variety of performance criteria to evaluate the learning methods.
| Year | Citations | |
|---|---|---|
2001 | 119.3K | |
1999 | 26.9K | |
1996 | 16.6K | |
1998 | 10.5K | |
1999 | 4.9K | |
2006 | 4.3K | |
1999 | 2.6K | |
1990 | 1.6K | |
2005 | 1.5K | |
2000 | 1.1K |
Page 1
Page 1