Computational Intelligence Techniques in Medical Decision Making: the Data Mining Perspective

  • V. Maojo
  • J. Sanandres
  • H. Billhardt
  • J. Crespo
Part of the Studies in Fuzziness and Soft Computing book series (STUDFUZZ, volume 96)


In this chapter, we give an overview of computational approaches to medical decision making, with particular emphasis on data mining methods. Medicine has been one of the most challenging application areas for Artificial Intelligence since the 1970s. The first generation of expert systems was an academic success. However, these systems failed to have a clinical impact. Researchers realized that medicine, and particularly patient care, is a complex domain, where requirements are different from other areas. A host of approaches were later adopted, including extracting objective information and knowledge from institutional and clinical databases using data mining. We give an overview of data mining methods, including neural networks, fuzzy sets and other machine learning approaches. Several examples of practical medical applications are presented. Finally, we discuss several limitations that must be overcome to effectively apply data mining methods and results to patient care.


Data Mining Medical Decision Radial Basis Function Network Electronic Patient Record Medical Informatics 
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.


Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.


  1. [1]
    Shortliffe, E.H., Buchanan, B., and Feigenbaum, E. (1979), “Knowledge engineering for medical decision making• a review of computer-based clinical decision aids,” Proceedings of the IEEE 67, pp. 1207–1224.CrossRefGoogle Scholar
  2. [2]
    Szolovits, P. (Ed.) (1982), Artificial Intelligence in Medicine, Westview Press, Boulder, CO.Google Scholar
  3. [3]
    Kassirer, J. and Kopelman, R. (1991), Learning Clinical Reasoning, Williams and Wilkins, Baltimore, USA.Google Scholar
  4. [4]
    Shortliffe, E.H. and Perreault, L. (2001), Medical Informatics: Computer Applications in Health Care and Biomedicine, 2°a edition, Springer Verlag, New York, USA.Google Scholar
  5. [5]
    Buchanan, B. and Shortliffe, E.H. (1984), Rule-Based Expert Systems: the Mycin Experiments of the Stanford Heuristic Programming Project, Addison-Wesley, New York, USA.Google Scholar
  6. [6]
    Miller, R.A., Pople, H.E., and Myers, J.D. (1982), “Internist-1, an experimental computer-based diagnostic consultant for general internal medicine,” New England Journal of Medicine, vol. 307, pp. 468–476.PubMedCrossRefGoogle Scholar
  7. [7]
    Szolovits, P. and Pauker, S.G. (1978), “Categorical and probabilistic reasoning in medical diagnosis,” Artificial Intelligence, vol. 11, pp. 115–144.Google Scholar
  8. [8]
    Weiss, S., Kulikowski, C., and Safir, A. (1978), “Glaucoma consultation by computer,” Computers in Biology and Medicine, vol. 8, no. 1, p. 25.PubMedCrossRefGoogle Scholar
  9. [9]
    Patil, R. and Senyk, O. (1988), “Compiling causal knowledge for diagnostic reasoning,” in: Miller, P. (Ed.) Selected Topics in Medical Artificial IntelligenceSpringer Verlag, New York, USA.Google Scholar
  10. [10]
    Rennels, G., Shortliffe, E., Stockdale, F., and Miller, P. (1986), “Computational model of reasoning from the clinical literature,” Proceedings of the 10 th Annual Symposium on Computer Applications in Medical Care, Washington, DC, pp. 373–380.Google Scholar
  11. [11]
    Bratko, I., Mozetic, I., and Lavrac, N. (1989), KARDIO: a Study in Deep and Qualitative Knowledge for Expert Systems, The MIT Press, Boston, MA.Google Scholar
  12. [12]
    Pazos, A., Maojo, V., and Ezquerra, N. (1992), “Neural networks in nuclear medicine: a new approach to prognosis in myocardial infarction,” Proceedings of the Seventh World Congress in Medical Informatics (Medinfo 92), North Holland, Geneva, Switzerland.Google Scholar
  13. [13]
    Gordon, C. and Christensen, J.P. (Eds.). (1995), Health Telematics for Clinical Guidelines and Protocols, IOS Press, Amsterdam, Netherlands.Google Scholar
  14. [14]
    Herrero, C., Maojo, V., Crespo, J., Sanandrés, J., and Lazaro, P. (1996), “A specification language for clinical practice guidelines,” Proceedings of the IEEE Conference in Medicine and Biology, Amsterdam.Google Scholar
  15. [15]
    Fayyad, U., Piatetsky-Shapiro, G., and Smyth P. (1996), “From data mining to knowledge discovery: an overview,” in: Fayyad, U., Piatetsky-Shapiro, G., Smyth, P., and Uthurusamy, R. (Eds.), Advances in Knowledge Discovery and Data Mining, AAAI Press/ The MIT Press, Menlo Park, California, pp. 1–34.Google Scholar
  16. [16]
    Han, J. and Kamber, M. (2000), Data Mining: Concepts and Techniques,Morgan Kaufmann Publishers.Google Scholar
  17. [17]
    Goebel, M. and Gruenwald, L. (1999), “A survey of data mining and knowledge discovery software tools,” SIGKDD Explorations, ACM SIGKDD.Google Scholar
  18. [18]
    Langley, P. and Simon, H.A. (1995), “Applications of machine learning and rule induction,” Communications of the ACM, vol. 38, no. 11, pp. 55–64.CrossRefGoogle Scholar
  19. [19]
    Weiss, S. and Indurkhya, N. (1998) Predictive Data Mining. A Practical Guide, Morgan Kaufmann, San Francisco, CA.Google Scholar
  20. [20]
    Long, W.J., Griffith, J.L., Selker, H.P., and D’Agostino R.B. (1993), “A comparison of logistic regression to decision-tree induction in a medical domain,” Computers and Biomedical Research, vol. 26, pp. 74–97.PubMedCrossRefGoogle Scholar
  21. [21]
    Lavrac, N., Keravnou, E., and Zupan, B. (1997), “Intelligent data analysis in medicine and pharmacology: an overview,” Intelligent Data Analysis in Medicine and Pharmacology, Kluwer, pp. 1–13.Google Scholar
  22. [22]
    Bratko, I. and Muggleton, S. (1995), “Applications of inductive logic programming,” Communications of the ACM, vol. 38, no. 11, pp. 65–70.CrossRefGoogle Scholar
  23. [23]
    Lim, T.S., Loh, W.Y., and Shih Y.S. (1997), “An empirical comparison of decision trees and other classification methods,” Technical Report 979, Department of Statistics, University of Wisconsin-Madison, Madison, WI.Google Scholar
  24. [24]
    Dreiseitl, S., Ohno-Machado, L., Kittler, H., Vinterbo, S., Billhardt, H., and Binder, M. (2001), “A comparison of machine learning methods for the diagnosis of pigmented skin lesions,” Journal of Biomedical Informatics, vol. 34, no. 1, pp. 28–36.PubMedCrossRefGoogle Scholar
  25. [25]
    Brodley, C.E. (1995), “Recursive automatic bias selection for classifier construction,” Machine Learning, vol. 20, pp. 63–94.Google Scholar
  26. [26]
    Michalski, R.S. (1983), “A theory and methodology of inductive learning,” Artificial Intelligence, vol. 20, pp. 111–161.CrossRefGoogle Scholar
  27. [27]
    Pearl, J. (1988), Probabilistic Reasoning in Intelligent Systems: Networks of Plausible Inference, Morgan Kaufmann, San Mateo, CA.Google Scholar
  28. [28]
    Hojsgaard, S. and Thiesson, B. (1995), “BIFROST — block recursive models induced from relevant knowledge, observations and statistical techniques,” Computational Statistics and Data Analysis, vol. 19, pp. 155–175.CrossRefGoogle Scholar
  29. [29]
    Ramoni, M., and Sebastiani, P. (1997), “Learning Bayesian networks from incomplete databases,” Proceedings of the Thirteen Conference on Uncertainty in Artificial Intelligence, Morgan Kaufman, San Mateo, California, USA, pp. 401–408.Google Scholar
  30. [30]
    Breiman, L. and Friedman, J. (1985), “Estimating optimal transformations for multiple regression and correlation (with discussion),” Journal of the American Statistical Association, vol. 80, pp. 580–619.CrossRefGoogle Scholar
  31. [31]
    Friedman, J.H. (1991), “Multivariable adaptive regression splines (with discussion),” The Annals of Statistics, vol. 19, pp. 1–141.CrossRefGoogle Scholar
  32. [32]
    Quinlan, J.R. (1992), C4.5: Programs for Machine Learning, Morgan Kaufmann, San Mateo, CA.Google Scholar
  33. [33]
    Breiman, L., Friedman, J., Olshen, R., and Stone, C. (1984), Classification and Regression Trees,Wadsworth International Group.Google Scholar
  34. [34]
    Michalski, R.S., Mozetic, I., Hong, J., and Lavrac, N. (1986), “The AQ 15 inductive learning system: an overview and experiments,” Proceedings of IMAL 1986, Orsay, Université de Paris-Sud.Google Scholar
  35. [35]
    Michalski, R.S., Mozetic, I., Hong, J., and Lavrac., N. (1986), “The multi-purpose incremental learning system AQ 15 and its testing application to three medical domains,” Proceedings of the.5 th national conference on Artificial Intelligence (AAAI-86), Philadelphia, pp. 1041–1045.Google Scholar
  36. [36]
    Clark, P. and Boswell. R. (1991), “Rule induction with CN2: some recent improvements,” Kodratoff, Y. (Ed.), Proceedings of the European Working Session on Learning (EWSL-91), Porto, Portugal, Springer Verlag, pp. 151–163.Google Scholar
  37. [37]
    Clark, P. and Niblett, T. (1988), “The CN2 induction algorithm,” Machine Learning, vol. 3, pp. 261.Google Scholar
  38. [38]
    Muggleton, S. and Feng, C. (1990), “Efficient induction of logic programs,” Proceedings of the f t Conference on Algorithmic Learning Theory, Ohmsma, Tokyo, Japan, pp. 368–381.Google Scholar
  39. [39]
    Quinlan, J.R. (1990), “Learning logical definitions from relations,” Machine Learning, vol. 5, pp. 239–266.Google Scholar
  40. [40]
    Koton, P.A. (1988), Using Experience in Learning and Problem Solving,Ph.D. Thesis, Massachusetts Institute of Technology, Laboratory of Computer Science, MIT/LCS-TR441.Google Scholar
  41. [41]
    Evans, D. and Patel, V. (Eds.) (1989), Cognitive Science in Medicine. Biomedical Modeling, MIT Press, Boston, MA.Google Scholar
  42. [42]
    Bareiss, R. (1989), Exemplar-Based Knowledge Acquisition: a Unified Approach to Concept Representation, Classification and Learning,Academic Press.Google Scholar
  43. [43]
    Aha, D.W., Kibler, D., and Albert, M.K. (1991), “Instance-based learning algorithms,” Machine Learning, vol. 6, pp. 37–66.Google Scholar
  44. [44]
    Holland, J. (1975), Adaptation in Natural and Artificial Systems, University of Michigan Press, Ann Arbor, USA.Google Scholar
  45. [45]
    Holland, J.H., Holyoak K.J., Nisbett, R.E., and Thagard, P.R. (1987), Induction: Processes of Inference, Learning, and Discovery, MIT Press, Cambridge, MA.Google Scholar
  46. [46]
    Bourgoin, M. and Smith, S. (1995), “Big data — better returns, leveraging your hidden data assets to improve ROI,” in: Freedman et al. (Eds.), Artificial Intelligence in the Capital Markets, Probus Publishing company.Google Scholar
  47. [47]
    Ohm, A., Komorowski, J., Skowron, A., and Synak, P. (1998), “The design and implementation of a knowledge discovery toolkit based on rough sets: the ROSETTA system,” in: Polkowski, L. and Skowron, A. (Eds.), Rough Sets in Knowledge Discovery. Studies in Fuzziness and Soft Computing, Physica-Verlag, vol. 18, pp. 376–399.Google Scholar
  48. [48]
    Zadeh, L.A. (1978), “Fuzzy sets as a basis for a theory of possibility,” Fuzzy Sets and Systems, vol. 1, pp. 3–28.CrossRefGoogle Scholar
  49. [49]
    Pawlak, Z. (1991), Rough Sets: Theoretical Aspects of Reasoning about Data,Kluwer Academic Publishers.Google Scholar
  50. [50]
    Van Bemmel, J. and Musen, M. (1997), Handbook of Medical Informatics, Springer-Verlag, New York, USA.Google Scholar
  51. [51]
    Rosenblatt, F. (1962), Principles of Neurodynamics, Spartan Books, New York.Google Scholar
  52. [52]
    Minsky, M.L. and Papert, S. (1969), Perceptrons: An Introduction to Computational Geometry, MIT Press, Cambridge, MA.Google Scholar
  53. [53]
    Rumelhart, D.E., Hinton, G.E., and Williams, R.J. (1986), “Learning internal representations by error propagation,” in: Rumelhart, D.E. and McClelland, J.L. (Eds.), Parallel Distributed Processing: Explorations in the Microstructure of Cognition, vol. 1, MIT Press. Boston, USA.Google Scholar
  54. [54]
    Broomhead, D.S. and Lowe, D. (1988), “Multivariable functional interpolation and adaptive networks,” Complex Systems, vol. 2, pp. 321–355.Google Scholar
  55. [55]
    Fahlman, S.E. and Lebiere, C. (1990), “The cascade-correlation learning architecture,” in: Touretzky, D.S. (Ed.), Advances in Neural Information Processing Systems 2, pp. 524–532, Morgan Kaufman Publishers Inc., San Mateo, California.Google Scholar
  56. [56]
    Bledsoe, W. and Browning, I. (1959), “Pattern recognition and reading by machine,” Proceedings of the Eastern Joint Computer Conference.Google Scholar
  57. [57]
    Aleksander, I. and Stonham, T. (1979), “Guide to pattern recognition using random-access memories,” IEE Proceedings on Computers and Digital Techniques, vol. 2, pp. 29–40.CrossRefGoogle Scholar
  58. [58]
    Krishnaiah, P. and Kanal, L. (Eds.) (1982), Classification, Pattern Recognition, and Reduction of Dimensionality, vol. 2 of Handbook of Statistics, North Holland, Amsterdam.Google Scholar
  59. [59]
    Kohonen, T. (1984), Self-Organization and Associative Memory, Springer-Verlag, Berlin.Google Scholar
  60. [60]
    Witten, I. (1999), Data Mining: Practical Machine Learning Tools and Techniques with Java Implementations,Morgan Kaufmann.Google Scholar
  61. [61]
    Billhardt, H., Maojo, V., Martin, F., Crespo, J., Pazos, A., Alamo, S., Rodriguez, J., and Sanandrés, J. (1998), “El proyecto armeda - acceso unificado a bases de datos medicas a través de la World Wide Web,” Actas del Congreso INFORSALUDNET 98, Madrid.Google Scholar
  62. [62]
    Wasson, J.H., Sox, H.C., Goldman. L., and Neff, R.K. (1985), “Clinical prediction rules; applications and methodologic standards,” New England Journal of Medicine, vol. 313, pp. 793799.Google Scholar
  63. [63]
    Goldman, L., Cook, E.F., Brand, D.A., Lee, T.H., Rouan, G.W., Weisberg, M.C., Acampora, D., Stasiulewicz, C., Walshon, J., Terranova, G., et al. (1988), “A computer protocol to predict myocardial infarction in emergency department patients with chest pain,” The New England Journal of Medicine, vol. 318, no. 13, pp. 797–803.PubMedCrossRefGoogle Scholar
  64. [64]
    Lorenzi, N. (1999), Proceedings of the AMIA 99 Annual Symposium, Hanley and Belfus, Inc, Philadelphia, USA.Google Scholar
  65. [65]
    Cooper, G.F., Aliferis, C.F., Ambrosino, R., Aronis, J., Buchanan, B.G., Caruana, R., Fine, M.J., Glymour, C., Gordon, G., Hanusa, B.H., Janosky, J.E., Meek, C., Mitchell, T., Richardson, T., Spirtes, P. (1997), “An evaluation of machine-learning methods for predicting pneumonia mortality,” Artificial Intelligence in Medicine, vol. 9, pp. 107–138.PubMedCrossRefGoogle Scholar
  66. [66]
    Tu, J.V., Weinstein, M.C., McNeil, B.J., Naylor, C.D., et al. (1998), “Predicting mortality after coronary artery bypass surgery,” Medical Decision Making, vol. 18, pp. 229–235.PubMedGoogle Scholar
  67. [67]
    Ciruelo, E., Crespo, J., Gomez, A., Maojo, V., Montes, C., and Sanandrés, J. (1997), “Predreuma: modelo de inducciôn constructiva en prognosis y clasificaciôn en artritis reumatoide,” Actas del Congreso Inforsalud 97, Madrid.Google Scholar
  68. [68]
    Sanandrés, J., Maojo, V., Crespo, J., and Gomez, A. (2001), “A clustering-based constructive induction method and its application to rheumatoid arthritis,”. in: Quaglini, S., Barahona, P., and Andreassen, S. (Eds.), Proc. of Artificial Intelligence in Medicine.Google Scholar
  69. [69]
    Teach, R.L. and Shortliffe, E.H. (1981), “An analysis of physician attitudes regarding computer-based clinical computation systems,” Computers and Biomedical Research, vol. 14, no. 6, pp. 542–558.PubMedCrossRefGoogle Scholar
  70. [70]
    Maojo, V. (2000), “A survey of data mining in medicine. methods and limitations,” keynote speech at ISMDA 2000, Frankfurt.Google Scholar
  71. [71]
    Cios, K. and Kacprzyk, J. (Eds.) (2001), Medical Data Mining and Knowledge Discovery, Springer Verlag, New York, USA.Google Scholar

Copyright information

© Springer-Verlag Berlin Heidelberg 2002

Authors and Affiliations

  • V. Maojo
  • J. Sanandres
  • H. Billhardt
  • J. Crespo

There are no affiliations available

Personalised recommendations