Skip to main content

Preprocessing by a Cost-Sensitive Literal Reduction Algorithm: Reduce

  • Conference paper
Learning, Networks and Statistics

Part of the book series: International Centre for Mechanical Sciences ((CISM,volume 382))

Abstract

This study is concerned with whether it is possible to detect what information contained in the training data and background knowledge is relevant for solving the learning problem, and whether irrelevant information can be eliminated in preprocessing before starting the learning process. A case study of data preprocessing for a hybrid genetic algorithm shows that the elimination of irrelevant features can substantially improve the efficiency of learning. In addition, cost-sensitive feature elimination can be effective for reducing costs of induced hypotheses.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Caruana, R. and D. Freitag: Greedy Attribute Selection, in: Proceedings of the 11th International Conference on Machine Learning, Morgan Kaufmann, 1994, 28–36.

    Google Scholar 

  2. Fayyad, U.M. and K.B. Irani: On the handling of continuous-valued attributes in decision tree generation, Machine Learning, 8 (1992), 87–102.

    MATH  Google Scholar 

  3. Gamberger, D.: A Minimization Approach to Propositional Inductive Learning, in: Proceedings of the 8th European Conference on Machine Learning, Springer, 1995, 151–160.

    Google Scholar 

  4. Grefenstette, J.J.: Optimization of control parameters for genetic algorithms, IEEE Transactions on Systems, Man, and Cybernetics, 16 (1986), 122–128.

    Article  Google Scholar 

  5. John, G.H., R. Kohavi and K. Pfleger: Irrelevant Features and the Subset Selection Problem, in: Proceedings of the 11th International Conference on Machine Learning, Morgan Kaufmann, 1994, 190–198.

    Google Scholar 

  6. Lavrac, N., S. Dzeroski and M. Grobelnik:. Learning Nonrecursive Definitions of Relations with LINUS, in: Proceedings of the 5th European Working Session on Learning, Springer, 1991, 265–281.

    Google Scholar 

  7. Lavrac, N. and S. Dzeroski: Inductive Logic Programming: Techniques and Applications, Ellis Horwood, 1994.

    MATH  Google Scholar 

  8. Lavrac, N., D. Gamberger and S. Dzeroski: An Approach to Dimensionality Reduction in Learning from Deductive Databases, in: Proceedings of the 5th International Workshop on Inductive Logic Programming, Scientific Report, Katholieke Universiteit Leuven, 1995, 337–354.

    Google Scholar 

  9. Lavrac, N., D. Gamberger and P. Turney: Cost-Sensitive Feature Reduction Applied to a Hybrid Genetic Algorithm, in: Proceedings of the 7th International Workshop on Algorithmic Learning Theory, Springer, 1996, 127–134.

    Google Scholar 

  10. Michalski, R.S. and J.B. Larson: Inductive Inference of VL Decision Rules, ACM SIGART Newsletter, 63 (1977), 38–44.

    Google Scholar 

  11. Michalski, R.S.: A Theory and Methodology of Inductive Learning, in: Machine Learning: An Artificial Intelligence Approach (Eds. R. Michalski, J. Carbonell and T. Mitchell ), Tioga, 1983, 83–134.

    Google Scholar 

  12. Michie, D., S. Muggleton, D. Page and A. Srinivasan: To the International Computing Community: A new East-West Challenge. Oxford University Computing Laboratory, Oxford, 1994. [Available at URL http://ftp.comlab.ox.ac.uk/pub/Packages/ILP/trains.tar.Z.]

  13. Quinlan, J.R.: C4. 5: Programs for Machine Learning, Morgan Kaufmann, 1993.

    Google Scholar 

  14. Skalak, D: Prototype and Feature Selection by Sampling and Random Mutation Hill Climbing Algorithms, in: Proceedings of the 11th International Conference on Machine Learning, Morgan Kaufmann, 1994, 293–301.

    Google Scholar 

  15. Turney, P.: Cost-Sensitive Classification: Empirical Evaluation of a Hybrid Genetic Decision Tree Induction Algorithm, Journal of Artificial Intelligence Research, 2 (1995), 369–409. [Available at URL http://www.cs.washington.edu/research/ jair/home.html.]

  16. Turney, P.: Low Size-Complexity Inductive Logic Programming: The East-West Challenge as a Problem in Cost-Sensitive Classification, in: Advances in Inductive Logic Programming (Ed. L. De Raedt ), IOS Press, 1996, 308–321.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 1997 Springer-Verlag Wien

About this paper

Cite this paper

Lavrac, N., Gamberger, D., Turney, P. (1997). Preprocessing by a Cost-Sensitive Literal Reduction Algorithm: Reduce. In: Della Riccia, G., Lenz, HJ., Kruse, R. (eds) Learning, Networks and Statistics. International Centre for Mechanical Sciences, vol 382. Springer, Vienna. https://doi.org/10.1007/978-3-7091-2668-4_11

Download citation

  • DOI: https://doi.org/10.1007/978-3-7091-2668-4_11

  • Publisher Name: Springer, Vienna

  • Print ISBN: 978-3-211-82910-3

  • Online ISBN: 978-3-7091-2668-4

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics