Skip to main content

Calibrated Associative Classification

  • Chapter
  • First Online:
Demand-Driven Associative Classification

Part of the book series: SpringerBriefs in Computer Science ((BRIEFSCOMPUTER))

  • 427 Accesses

Abstract

Given an input \(x_i\) and an arbitrary output \(c_j,\) a classification algorithm usually works by estimating the probability of \(x_i\) being related to \(c_j\) (i.e., class membership probability). Well calibrated classification algorithms are those able to produce functions that provide accurate estimates of class membership probabilities, that is, the estimated probability \(\hat{p}(c_j|x_i)\) is close to \(p(c_j|\hat{p}(c_j|x_i)),\) which is the true, (unknown) empirical probability of \(x_i\) being related to output \(c_j\) given that the probability estimated by the classification algorithm is \(\hat{p}(c_j|x_i).\) Calibration is not a necessary property for producing an accurate approximation of the target function, and, thus, most of the research has focused on direct accuracy maximization strategies rather than on calibration. However, non-calibrated functions are problematic in applications where the reliability associated with a prediction must be taken into account (i.e., cost-sensitive classification , cautious classifications etc.). In these applications, a sensible use of the classification algorithm must be based on the reliability of its predictions (Veloso et al. Calibrated lazy associative classification, Inform Sci, 2009), and thus, the algorithm must produce well calibrated functions. In this chapter we introduce calibrated associative classification algorithms .

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    \(v\) can take the values 0 (the prediction is wrong) or 1 (otherwise), as shown in step 6 of Algorithm 14.

  2. 2.

    For each experiment, predictions were sorted from the most reliable to the least reliable.

  3. 3.

    The basic idea is to solicit a person \(x_i\) for whom the expected return \(\hat{p}(\hbox{donate}|x_i)y(x_i)\) is greater than the cost of mailing the solicitation.

References

  1. Boser, B., Guyon, I., Vapnik, V.: A training algorithm for optimal margin classifiers. In: Proceedings of the Annual Conference on Computational Learning Theory (COLT), pp. 144–152. Springer (1992)

    Google Scholar 

  2. Cestnik, B.: Estimating probabilities: a crucial task in machine learning. In: Proceedings of the European Conference on Artificial Intelligence (ECAI), pp. 147–149 (1990)

    Google Scholar 

  3. Chang, C.-C., Lin, C.-J.: LIBSVM: a library for support vector machines. Available at http://www.csie.ntu.edu.tw/∼cjlin/papers/libsvm.pdf (2001)

  4. Cohen, I., Goldszmidt, M.: Properties and benefits of calibrated classifiers. In: Proceedings of the European Conference on Principles of Data Mining and Knowledge Discovery (PKDD), pp. 125–136. Springer (2004)

    Google Scholar 

  5. Cussens, J.: Bayes and pseudo-bayes estimates of conditional probabilities and their reliability. In: Proceedings of the European Conference on Machine Learning (ECML), pp. 136–152. Springer (1993)

    Google Scholar 

  6. DeGroot, M., Fienberg, S.: The comparison and evalution of forecasters. Statistician 32, 12–22 (1982)

    Article  Google Scholar 

  7. Fayyad, U., Irani, K.: Multi interval discretization of continuous-valued attributes for classification learning. In: Proceedings of the International Joint Conference on Artificial Intelligence (IJCAI), pp. 1022–1027 (1993)

    Google Scholar 

  8. Platt, J.: Probabilistic outputs for support vector machines and comparison to regularized likelihood methods. Adv. Large Margin Classif. 61–74 (1999)

    Google Scholar 

  9. Quinlan, J.: C4.5: Programs for Machine Learning. Morgan Kaufmann, San Francisco (1993)

    Google Scholar 

  10. Rissanen, J.: Modeling by shortest data description. Automatica 14, 465–471 (1978)

    Article  MATH  Google Scholar 

  11. Zadrozny, B., Elkan, C.: Obtaining calibrated probability estimates from decision trees and naive bayesian classifiers. In: Proceedings of the International Conference on Machine Learning (ICML), pp. 609–616. IEEE Computer Society (2001)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Adriano Veloso .

Rights and permissions

Reprints and permissions

Copyright information

© 2011 Adriano Veloso

About this chapter

Cite this chapter

Veloso, A., Meira, W. (2011). Calibrated Associative Classification. In: Demand-Driven Associative Classification. SpringerBriefs in Computer Science. Springer, London. https://doi.org/10.1007/978-0-85729-525-5_7

Download citation

  • DOI: https://doi.org/10.1007/978-0-85729-525-5_7

  • Published:

  • Publisher Name: Springer, London

  • Print ISBN: 978-0-85729-524-8

  • Online ISBN: 978-0-85729-525-5

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics