Abstract
The exponential growth of DNA dataset in the scientific repository has been encouraging interdisciplinary research on ecology, computer science, and bioinformatics. For better classification of cancer (DNA gene expression), many technologies are useful as demonstrated by a prior experimental study. The major challenging task of gene selection method is extracting informative genes contribution in the classification from the DNA microarray datasets at low computational cost. In this paper, amalgamation of Spearman’s correlation (SC) and filter-based feature selection (FS) methods is proposed. We demonstrate the extensive comparison of the effect of Spearman’s correlation with FS methods, i.e., Relief-F, Joint Mutual Information (JMI), and max-relevance and min-redundancy (MRMR). To measure the classification performance, four diverse supervised classifiers, i.e., K-nearest neighbor (K-NN), support vector machines (SVM), naïve Bayes (NB), and decision tree (DT), have been used on DLBCL dataset. The result demonstrates that Spearman’s correlation in conglomeration with MRMR performs better than other combinations.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
H. Liu, H. Motoda, R. Setiono, and Z. Zhao, “Feature Selection : An Ever Evolving Frontier in Data Mining,” J. Mach. Learn. Res. Work. Conf. Proc. 10 Fourth Work. Featur. Sel. Data Min., pp. 4–13, 2010.
H. Peng, F. Long, and C. Ding, “Feature selection based on mutual information: Criteria of Max-Dependency, Max-Relevance, and Min-Redundancy,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 27, no. 8, pp. 1226–1238, 2005.
H. H. Yang and J. Moody, “Feature selection based on joint mutual information,” Proc. Int. ICSC Symp. Adv. Intell. Data Anal., pp. 22–25, 1999.
M. R.- Sikonja, “Theoretical and Empirical Analysis of ReliefF and RReliefF,” Mach. Learn. J., vol. 1, no. 53, pp. 23–69, 2003.
H. Lai, Y. Tang, H. Luo, and Y. Pan, “Greedy feature selection for ranking,” Proc. 2011 15th Int. Conf. Comput. Support. Coop. Work Des. CSCWD 2011, pp. 42–46, 2011.
I. Guyon and A. Elisseeff, “An Introduction to Variable and Feature Selection,” J. Mach. Learn. Res., vol. 3, no. 3, pp. 1157–1182, 2003.
X. Liu, A. Krishnan, and A. Mondry, “An entropy-based gene selection method for cancer classification using microarray data.,” BMC Bioinformatics, vol. 6, no. 1, p. 76, 2005.
H. Lu, J. Chen, K. Yan, Q. Jin, Y. Xue, and Z. Gao, “A Hybrid Feature Selection Algorithm for Gene Expression Data Classification,” Neurocomputing, no. 2017, 2016.
A. S. Ghareb, A. A. Bakar, and A. R. Hamdan, “Hybrid feature selection based on enhanced genetic algorithm for text categorization,” Expert Syst. Appl., vol. 49, pp. 31–47, 2016.
P. Moradi and M. Gholampour, “A hybrid particle swarm optimization for feature subset selection by integrating a novel local search strategy,” Appl. Soft Comput. J., vol. 43, pp. 117–130, 2016.
S. A. Medjahed, T. A. Saadi, A. Benyettou, and M. Ouali, “Kernel-based learning and feature selection analysis for cancer diagnosis,” Appl. Soft Comput. J., vol. 51, pp. 39–48, 2017.
Y. Sun, “Iterative RELIEF for feature weighting: Algorithms, theories, and applications,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 29, no. 6, pp. 1035–1051, 2007.
A. Arauzo-Azofra, J. Benitez, and J. Castro, “A feature set measure based on relief,” Proc. fifth Int. Conf. Recent Adv. Soft Comput., pp. 104–109, 2004.
R. Kohavi and H. John, “Wrappers for feature subset selection,” Artif. Intell., vol. 97, no. 97, pp. 273–324, 1997.
C. Spearman, “The Proof and Measurement of Association between Two Things,” Am. J. Psychol., vol. 15, no. 1, pp. 72–101, 2017.
K. Q. Weinberger and L. K. Saul, “Distance Metric Learning for Large Margin Nearest Neighbor Classification,” J. Mach. Learn. Res., vol. 10, pp. 207–244, 2009.
A. Ben-hur and J. Weston, “A user’s guide to support vector machines,” Data Min. Tech. life Sci., pp. 223–39, 2010.
N. Friedman, D. Geiger, and M. Goldszmidt, “Bayesian Network Classifiers,” Mach. Learn., vol. 29, pp. 131–163, 1997.
W. Loh, “Classification and regression trees,” Data Min. Knowl. Discov., vol. 1, no. February, pp. 14–23, 2011.
“Diffuse Large B-cell Lymphoma Dataset.” [Online]. Available: https://llmpp.nih.gov/lymphoma/data/clones.txt.
M. Sokolova, N. Japkowicz, and S. Szpakowicz, “Beyond accuracy, F-Score and ROC: A family of discriminant measures for performance evaluation,” Adv. Artif. Intell., vol. 4304, pp. 1015–1021, 2006.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2019 Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Shukla, A.K., Singh, P., Vardhan, M. (2019). DNA Gene Expression Analysis on Diffuse Large B-Cell Lymphoma (DLBCL) Based on Filter Selection Method with Supervised Classification Method. In: Behera, H., Nayak, J., Naik, B., Abraham, A. (eds) Computational Intelligence in Data Mining. Advances in Intelligent Systems and Computing, vol 711. Springer, Singapore. https://doi.org/10.1007/978-981-10-8055-5_69
Download citation
DOI: https://doi.org/10.1007/978-981-10-8055-5_69
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-10-8054-8
Online ISBN: 978-981-10-8055-5
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)