Abstract
The recent increase of dimensionality of data is a target for many existing feature selection methods with respect to efficiency and effectiveness. In this paper, the all relevant feature selection method based on information gathered using generational feature elimination was introduced. The successive generations of feature subset were defined using DTLevelImp algorithm and in each step the subset of most important features were eliminated from the primary investigated dataset. This process was executed until the most important feature reach importance value on the level similar to importance of the random shadow features. The proposed method was also initially tested on well-know artificial and real-world datasets and the results confirm its efficiency. Thus, it can be concluded that selected attributes are relevant.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Kuhn, M., Johnson, K.: Applied Predictive Modeling, pp. 487–500. Springer, New York (2013)
Nilsson, R., Peña, J.M., Björkegren, J., Tegnér, J.: Detecting multivariate differentially expressed genes. BMC Bioinform. 8, 150 (2007)
Phuong, T.M., Lin, Z., Altman, R.B.: Choosing SNPs using feature selection. In: Proceedings of the IEEE Computational Systems Bioinformatics Conference, CSB 2005, pp. 301–309 (2005)
Paja, W., Wrzesień, M., Niemiec, R., Rudnicki, W.R.: Application of all-relevant feature selection for the failure analysis of parameter-induced simulation crashes in climate models. Geoscientific Model Dev. 9, 1065–1072 (2016)
Pancerz, K., Paja, W., Gomuła, J.: Random forest feature selection for data coming from evaluation sheets of subjects with ASDs. In: Ganzha, M., Maciaszek, L., Paprzycki, M. (eds.) Proceedings of the 2016 Federated Conference on Computer Science and Information Systems (FedCSIS), pp. 299–302 (2016)
Bermingham, M.L., Pong-Wong, R., Spiliopoulou, A., Hayward, C., Rudan, I., Campbell, H., Wright, A.F., Wilson, J.F., Agakov, F., Navarro, P., Haley, C.S.: Application of high-dimensional feature selection: evaluation for genomic prediction in man. Sci. Rep. 5, 10312 (2015)
Kohavi, R., John, G.H.: Wrappers for feature subset selection. Artif. Intell. 97, 273–324 (1997)
Rudnicki, W.R., Wrzesień, M., Paja, W.: All relevant feature selection methods and applications. In: Stańczyk, U., Jain, L. (eds.) Feature Selection for Data and Pattern Recognition. Studies in Computational Intelligence, vol. 584, pp. 11–28. Springer-Verlag, Germany (2015)
Zhu, Z., Ong, Y.S., Dash, M.: Wrapper-filter feature selection algorithm using a memetic framework. IEEE Trans. Syst. Man Cybern. Part B Cybern. 37, 70–76 (2007)
Guyon, I., Weston, J., Barnhill, S., Vapnik, V.: Mach. Learn. 46, 389–422 (2002)
Johannes, M., Brase, J.C., Frohlich, H., Gade, S., Gehrmann, M., Falth, M., Sultmann, H., Beiflbarth, T.: Integration of pathway knowledge into a reweighted recursive feature elimination approach for risk stratification of cancer patients. Bioinformatics 26(17), 2136–2144 (2010)
Stoppiglia, H., Dreyfus, G., Dubois, R., Oussar, Y.: Ranking a random feature for variable and feature selection. J. Mach. Learn. Res. 3, 1399–1414 (2003)
Tuv, E., Borisov, A., Torkkola, K.: Feature selection using ensemble based ranking against artificial contrasts. In: International Symposium on Neural Networks, pp. 2181–2186 (2006)
Paja, W.: Feature selection methods based on decision rule and tree models. In: Czarnowski, I., Caballero, A.M., Howlett, R.J., Jain, L.C. (eds.) Intelligent Decision Technologies 2016: Proceedings of the 8th KES International Conference on Intelligent Decision Technologies (KES-IDT 2016) - Part II, pp. 63–70. Springer, Cham (2016)
Guyon, I., Gunn, S., Ben-Hur, A., Dror, G.: Result analysis of the NIPS 2003 feature selection challenge. Adv. Neural Inf. Process. Syst. 17, 545–552 (2013)
Lucas, D.D., Klein, R., Tannahill, J., Ivanova, D., Brandon, S., Domyancic, D., Zhang, Y.: Failure analysis of parameter-induced simulation crashes in climate models. Geosci. Model Dev. Discuss. 6, 585–623 (2013)
Acknowledgment
This work was supported by the Center for Innovation and Transfer of Natural Sciences and Engineering Knowledge at the University of Rzeszów.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2018 Springer International Publishing AG
About this paper
Cite this paper
Paja, W. (2018). Generational Feature Elimination to Find All Relevant Feature Subset. In: Czarnowski, I., Howlett, R., Jain, L. (eds) Intelligent Decision Technologies 2017. IDT 2017. Smart Innovation, Systems and Technologies, vol 72. Springer, Cham. https://doi.org/10.1007/978-3-319-59421-7_13
Download citation
DOI: https://doi.org/10.1007/978-3-319-59421-7_13
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-59420-0
Online ISBN: 978-3-319-59421-7
eBook Packages: EngineeringEngineering (R0)