Sciweavers

ICML
2001
IEEE

Obtaining calibrated probability estimates from decision trees and naive Bayesian classifiers

14 years 5 months ago
Obtaining calibrated probability estimates from decision trees and naive Bayesian classifiers
Accurate, well-calibrated estimates of class membership probabilities are needed in many supervised learning applications, in particular when a cost-sensitive decision must be made about examples with example-dependent costs. This paper presents simple but successful methods for obtaining calibrated probability estimates from decision tree and naive Bayesian classifiers. Using the large and challenging KDD'98 contest dataset as a testbed, we report the results of a detailed experimental comparison of ten methods, according to four evaluation measures. We conclude that binning succeeds in significantly improving naive Bayesian probability estimates, while for improving decision tree probability estimates, we recommend smoothing by -estimation and a new variant of pruning that we call curtailment.
Bianca Zadrozny, Charles Elkan
Added 17 Nov 2009
Updated 17 Nov 2009
Type Conference
Year 2001
Where ICML
Authors Bianca Zadrozny, Charles Elkan
Comments (0)