Skip to main content

Classification Using Rough Random Forest

  • Conference paper
  • First Online:

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 9468))

Abstract

The Rough random forest is a classification model based on rough set theory. The Rough random forest uses the concept of random forest and rough set theory in a single model. It combines a collection of decision trees for classification instead of depending on a single decision tree. It uses the concept of bagging and random subspace method to improve the performance of the classification model. In the rough random forest the reducts of each decision tree are chosen on the basis of boundary region condition. Each decision tree uses a different subset of patterns and features. The class label of patterns is obtained by combining the decisions of all the decision trees by majority voting. Results are reported on a number of benchmark datasets and compared with other techniques. Rough random forest is found to give better performance.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

References

  1. Bonissone, P., Cadenas, J., Garrido, M., Dıaz-Valladares, R.: A fuzzy random forest: Fundamental for design and construction. In: Proceedings of the 12th International Conference on Information Processing and Management of Uncertainty in Knowledge-Based Systems (IPMU 2008), pp. 1231–1238 (2008)

    Google Scholar 

  2. Breiman, L.: Random forests. Mach. Learn. 45(1), 5–32 (2001)

    Article  MATH  Google Scholar 

  3. Breiman, L.: Bagging predictors. Mach. Learn. 24, 123–140 (1996)

    MathSciNet  MATH  Google Scholar 

  4. Dietterich, T.G.: An experimental comparison of three methods for constructing ensembles of decision trees: Bagging, boosting, and randomization. Mach. Learn. 40(2), 139–157 (2000)

    Article  Google Scholar 

  5. Ding, B., Zheng, Y., Zang, S.: A new decision tree algorithm based on rough set theory. In: 2009 Asia-Pacific Conference on Information Processing, pp. 326–329, July 2009

    Google Scholar 

  6. Hamza, M., Larocque, D.: An empirical comparison of ensemble methods based on classification trees. J. Stat. Comput. Simul. 75, 629–643 (2005)

    Article  MATH  Google Scholar 

  7. Huang, L., Huang, M., Guo, B., Zhuang, Z.: A new method for constructing decision tree based on rough set theory. In: 2007 IEEE International Conference on Granular Computing (GRC 2007), pp. 241–241, November 2007

    Google Scholar 

  8. Komorowski, J., Pawlak, Z., Polkowski, L., Skowron, A.: Rough sets: A tutorial. Rough fuzzy hybridization: A new trend in decision-making, pp. 3–98 (1999)

    Google Scholar 

  9. Lichman, M.: UCI machine learning repository (2013)

    Google Scholar 

  10. Patel, B.R., Rana, K.K.: A survey on decision tree algorithm for classification. Int. J. Eng. Dev. Res. IJEDR 2, 1–5 (2014)

    Google Scholar 

  11. Pawlak, Z.: Rough sets. Int. J. Comput. Inform. Sci. 11(5), 341–356 (1982)

    Article  MathSciNet  MATH  Google Scholar 

  12. Quinlan, R.J.: C4.5: Programs for Machine Learning. Morgan Kaufmann Publishers Inc., San Francisco (1993)

    Google Scholar 

  13. Quinlan, J.R.: Induction of decision trees. Mach. Learn. 1(1), 81–106 (1986)

    Google Scholar 

  14. Rokach, L., Maimon, O.: Data Mining with Decision Trees: Theroy and Applications. World Scientific Publishing Co., Inc., River Edge (2008)

    Google Scholar 

  15. Skurichina, M., Duin, R.P.W.: Bagging, boosting and the random subspace method for linear classifiers. Pattern Anal. Appl. 5, 121–135 (2002)

    Article  MathSciNet  MATH  Google Scholar 

  16. Suraj, Z.: An introduction to rough set theory and its applications. In: ICENCO, Cairo, Egypt (2004)

    Google Scholar 

  17. Wei, J., Huang, D., Wang, S., Ma, Z.: Rough set based decision tree. In: 2002 Proceedings of the 4th World Congress on Intelligent Control and Automation, vol. 1, pp. 426–431 (2002)

    Google Scholar 

  18. Yeh, C.C., Lin, F., Hsu, C.Y.: A hybrid KMV model, random forests and rough set theory approach for credit rating. Knowl. Based Syst. 33, 166–172 (2012)

    Article  Google Scholar 

  19. Zhang, X., Zhou, J., He, Y., Wang, Y., Liu, B.: Vibration fault diagnosis of hydro-turbine generating unit based on rough 1-v-1 multiclass support vector machine. In: 2010 Sixth International Conference on Natural Computation (ICNC), vol. 2, pp. 755–759, August 2010

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Rajhans Gondane .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2015 Springer International Publishing Switzerland

About this paper

Cite this paper

Gondane, R., Devi, V.S. (2015). Classification Using Rough Random Forest. In: Prasath, R., Vuppala, A., Kathirvalavakumar, T. (eds) Mining Intelligence and Knowledge Exploration. MIKE 2015. Lecture Notes in Computer Science(), vol 9468. Springer, Cham. https://doi.org/10.1007/978-3-319-26832-3_8

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-26832-3_8

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-26831-6

  • Online ISBN: 978-3-319-26832-3

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics