skip to main content
10.1145/775047.775151acmconferencesArticle/Chapter ViewAbstractPublication PageskddConference Proceedingsconference-collections
Article

Transforming classifier scores into accurate multiclass probability estimates

Authors Info & Claims
Published:23 July 2002Publication History

ABSTRACT

Class membership probability estimates are important for many applications of data mining in which classification outputs are combined with other sources of information for decision-making, such as example-dependent misclassification costs, the outputs of other classifiers, or domain knowledge. Previous calibration methods apply only to two-class problems. Here, we show how to obtain accurate probability estimates for multiclass problems by combining calibrated binary probability estimates. We also propose a new method for obtaining calibrated two-class probability estimates that can be applied to any classifier that produces a ranking of examples. Using naive Bayes and support vector machine classifiers, we give experimental results from a variety of two-class and multiclass domains, including direct marketing, text categorization and digit recognition.

References

  1. E.L. Allwein, R. E. Schapire, and Y. Singer. Reducing multiclass to binary: A unifying approach for margin classifiers. Journal of Machine Learning Research, 1:113--141, 2000. Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. M. Ayer, H. Brunk, G. Ewing, W. Reid, and E. Silverman. An empirical distribution function for sampling with incomplete information. Annals of Mathematical Statistics, 26(4):641--647, 1955.Google ScholarGoogle ScholarCross RefCross Ref
  3. S. D. Bay. UCI KDD archive. Department of Information and Computer Sciences, University of California, lrvine, 2000. http://kdd.ics.uci.edu/.Google ScholarGoogle Scholar
  4. P. N. Bennett. Assessing the calibration of naive Bayes' posterior estimates. Technical Report CMU-CS-00-155, Carnegie Mellon University, 2000.Google ScholarGoogle Scholar
  5. C.L. Blake and C. J. Merz. UCI repository of machine learning databases. Department of Information and Computer Sciences, University of California, Irvine, 1998. hetp://www.ics.uci.edu/~mlearn/MLRepository.html.Google ScholarGoogle Scholar
  6. G. W. Brier. Verification of forecasts expressed in terms of probability. Monthly Weather Review, 78:1--3, 1950.Google ScholarGoogle ScholarCross RefCross Ref
  7. M. H. DeGroot and S. E. Fienberg. The comparison and evaluation of forecasters. Statistician, 32(1):12--22, 1982.Google ScholarGoogle Scholar
  8. T. G. Dietterich and G. Bakiri. Solving multiclass learning problems via error-correcting output codes. Journal of Artificial Intelligence Research, 2:263--286, 1995. Google ScholarGoogle ScholarCross RefCross Ref
  9. P. Domingos and M. Pazzani. Beyond independence: Conditions for the optimality of the simple Bayesian classifier. In Proceedings of the Thirteenth International Conference on Machine Learning, pages 105--112. Morgan Kaufmann Publishers, Inc., 1996.Google ScholarGoogle Scholar
  10. L. Dümbgen. Statistical software (MATLAB), 2000. Available at http://www.math.mu-luebeck.de/workers/duembgen/software/software.html.Google ScholarGoogle Scholar
  11. C. Elkan. Boosting and naive bayesian learning. Technical Report CS97-557, University of California, San Diego, 1997.Google ScholarGoogle Scholar
  12. C. Elkan. Magical thinking in data mining: Lessons from coil challenge 2000. In Proceedings of the Seventh International Conference on Knowledge Discovery and Data Mining, pages 426--431. ACM Press, 2001. Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. J. Georges and A. H. Milley. KDD'99 competition: Knowledge discovery contest report. Available at http://www-cse.ucsd.edu/users/elkan/kdresults.html, 1999.Google ScholarGoogle Scholar
  14. T. Hastie and R. Tibshirani. Classification by pairwise coupling. In Advances in Neural Information Processing Systems, volume 10. MIT Press, 1998. Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. E. G. Kong and T. G. Dietterich. Probability estimation using error-correcting output coding. In Int. Conf : Artificial Intelligence and Soft Computing, 1997.Google ScholarGoogle Scholar
  16. K. Lang. Newsweeder: Learning to filter netnews. In Proceedings of the Twelfth International Conference on Machine Learning, pages 331--339, 1995.Google ScholarGoogle ScholarCross RefCross Ref
  17. A. Murphy and R. Winkler. Reliability of subjective probability forecasts of precipitation and temperature. Applied Statistics, 26(1):41--47, 1977.Google ScholarGoogle ScholarCross RefCross Ref
  18. J. Platt. Probabilistic outputs for support vector rnachines and comparison to regularized likelihood methods. In Advances in Large Margin Classifiers. MIT Press, 1999.Google ScholarGoogle Scholar
  19. F. Provost and P. Domingos. Well-trained PETs: Improving probability estimation trees. CDER Working Paper #00-04-IS, Stern School of Business, New York University, NY, NY 10012, 2000.Google ScholarGoogle Scholar
  20. J. Rennie and R. Rifkin. Improving multiclass text classification with the support vector machine. Technical Report AIM-2001-026.2001, MIT, 2001.Google ScholarGoogle Scholar
  21. R. Rifkin. SvmFu 3, 2001. Available at http://five-percent-nation.mit.edu/SvmFu.Google ScholarGoogle Scholar
  22. T. Robertson, P. Wright, and R. Dykstra. Order Restricted Statistical Inference, chapter 1. John Wiley & Sons, 1988.Google ScholarGoogle Scholar
  23. B. Zadrozny. Reducing multiclass to binary by coupling probability estimates. In Advances in Neural Information Processing Systems (NIPS*2001), 2002. To appear.Google ScholarGoogle Scholar
  24. B. Zadrozny and C. Elkan. Learning and making decisions when costs and probabilities are both unknown. In Proceedings of the Seventh International Conference on Knowledge Discovery and Data Mining, pages 204--213. ACM Press, 2001. Google ScholarGoogle ScholarDigital LibraryDigital Library
  25. B. Zadrozny and C. Elkan. Obtaining calibrated probability estimates from decision trees and naive bayesian classifiers. In Proceedings of the Eighteenth International Conference on Machine Learning, pages 609--616. Morgan Kaufmann Publishers, Inc., 2001. Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. Transforming classifier scores into accurate multiclass probability estimates

                Recommendations

                Comments

                Login options

                Check if you have access through your login credentials or your institution to get full access on this article.

                Sign in
                • Published in

                  cover image ACM Conferences
                  KDD '02: Proceedings of the eighth ACM SIGKDD international conference on Knowledge discovery and data mining
                  July 2002
                  719 pages
                  ISBN:158113567X
                  DOI:10.1145/775047

                  Copyright © 2002 ACM

                  Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

                  Publisher

                  Association for Computing Machinery

                  New York, NY, United States

                  Publication History

                  • Published: 23 July 2002

                  Permissions

                  Request permissions about this article.

                  Request Permissions

                  Check for updates

                  Qualifiers

                  • Article

                  Acceptance Rates

                  KDD '02 Paper Acceptance Rate44of307submissions,14%Overall Acceptance Rate1,133of8,635submissions,13%

                  Upcoming Conference

                  KDD '24

                PDF Format

                View or Download as a PDF file.

                PDF

                eReader

                View online with eReader.

                eReader