Skip to main content

Generational Feature Elimination to Find All Relevant Feature Subset

  • Conference paper
  • First Online:
Intelligent Decision Technologies 2017 (IDT 2017)

Part of the book series: Smart Innovation, Systems and Technologies ((SIST,volume 72))

Included in the following conference series:

Abstract

The recent increase of dimensionality of data is a target for many existing feature selection methods with respect to efficiency and effectiveness. In this paper, the all relevant feature selection method based on information gathered using generational feature elimination was introduced. The successive generations of feature subset were defined using DTLevelImp algorithm and in each step the subset of most important features were eliminated from the primary investigated dataset. This process was executed until the most important feature reach importance value on the level similar to importance of the random shadow features. The proposed method was also initially tested on well-know artificial and real-world datasets and the results confirm its efficiency. Thus, it can be concluded that selected attributes are relevant.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 129.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 169.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 169.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Kuhn, M., Johnson, K.: Applied Predictive Modeling, pp. 487–500. Springer, New York (2013)

    Book  MATH  Google Scholar 

  2. Nilsson, R., Peña, J.M., Björkegren, J., Tegnér, J.: Detecting multivariate differentially expressed genes. BMC Bioinform. 8, 150 (2007)

    Article  Google Scholar 

  3. Phuong, T.M., Lin, Z., Altman, R.B.: Choosing SNPs using feature selection. In: Proceedings of the IEEE Computational Systems Bioinformatics Conference, CSB 2005, pp. 301–309 (2005)

    Google Scholar 

  4. Paja, W., Wrzesień, M., Niemiec, R., Rudnicki, W.R.: Application of all-relevant feature selection for the failure analysis of parameter-induced simulation crashes in climate models. Geoscientific Model Dev. 9, 1065–1072 (2016)

    Article  Google Scholar 

  5. Pancerz, K., Paja, W., Gomuła, J.: Random forest feature selection for data coming from evaluation sheets of subjects with ASDs. In: Ganzha, M., Maciaszek, L., Paprzycki, M. (eds.) Proceedings of the 2016 Federated Conference on Computer Science and Information Systems (FedCSIS), pp. 299–302 (2016)

    Google Scholar 

  6. Bermingham, M.L., Pong-Wong, R., Spiliopoulou, A., Hayward, C., Rudan, I., Campbell, H., Wright, A.F., Wilson, J.F., Agakov, F., Navarro, P., Haley, C.S.: Application of high-dimensional feature selection: evaluation for genomic prediction in man. Sci. Rep. 5, 10312 (2015)

    Article  Google Scholar 

  7. Kohavi, R., John, G.H.: Wrappers for feature subset selection. Artif. Intell. 97, 273–324 (1997)

    Article  MATH  Google Scholar 

  8. Rudnicki, W.R., Wrzesień, M., Paja, W.: All relevant feature selection methods and applications. In: Stańczyk, U., Jain, L. (eds.) Feature Selection for Data and Pattern Recognition. Studies in Computational Intelligence, vol. 584, pp. 11–28. Springer-Verlag, Germany (2015)

    Google Scholar 

  9. Zhu, Z., Ong, Y.S., Dash, M.: Wrapper-filter feature selection algorithm using a memetic framework. IEEE Trans. Syst. Man Cybern. Part B Cybern. 37, 70–76 (2007)

    Google Scholar 

  10. Guyon, I., Weston, J., Barnhill, S., Vapnik, V.: Mach. Learn. 46, 389–422 (2002)

    Article  Google Scholar 

  11. Johannes, M., Brase, J.C., Frohlich, H., Gade, S., Gehrmann, M., Falth, M., Sultmann, H., Beiflbarth, T.: Integration of pathway knowledge into a reweighted recursive feature elimination approach for risk stratification of cancer patients. Bioinformatics 26(17), 2136–2144 (2010)

    Article  Google Scholar 

  12. Stoppiglia, H., Dreyfus, G., Dubois, R., Oussar, Y.: Ranking a random feature for variable and feature selection. J. Mach. Learn. Res. 3, 1399–1414 (2003)

    MATH  Google Scholar 

  13. Tuv, E., Borisov, A., Torkkola, K.: Feature selection using ensemble based ranking against artificial contrasts. In: International Symposium on Neural Networks, pp. 2181–2186 (2006)

    Google Scholar 

  14. Paja, W.: Feature selection methods based on decision rule and tree models. In: Czarnowski, I., Caballero, A.M., Howlett, R.J., Jain, L.C. (eds.) Intelligent Decision Technologies 2016: Proceedings of the 8th KES International Conference on Intelligent Decision Technologies (KES-IDT 2016) - Part II, pp. 63–70. Springer, Cham (2016)

    Google Scholar 

  15. Guyon, I., Gunn, S., Ben-Hur, A., Dror, G.: Result analysis of the NIPS 2003 feature selection challenge. Adv. Neural Inf. Process. Syst. 17, 545–552 (2013)

    Google Scholar 

  16. Lucas, D.D., Klein, R., Tannahill, J., Ivanova, D., Brandon, S., Domyancic, D., Zhang, Y.: Failure analysis of parameter-induced simulation crashes in climate models. Geosci. Model Dev. Discuss. 6, 585–623 (2013)

    Article  Google Scholar 

Download references

Acknowledgment

This work was supported by the Center for Innovation and Transfer of Natural Sciences and Engineering Knowledge at the University of Rzeszów.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to W. Paja .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2018 Springer International Publishing AG

About this paper

Cite this paper

Paja, W. (2018). Generational Feature Elimination to Find All Relevant Feature Subset. In: Czarnowski, I., Howlett, R., Jain, L. (eds) Intelligent Decision Technologies 2017. IDT 2017. Smart Innovation, Systems and Technologies, vol 72. Springer, Cham. https://doi.org/10.1007/978-3-319-59421-7_13

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-59421-7_13

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-59420-0

  • Online ISBN: 978-3-319-59421-7

  • eBook Packages: EngineeringEngineering (R0)

Publish with us

Policies and ethics