Abstract
Exploiting label correlation is important for multi-label learning, where each instance is associated with a set of labels. However, most of existing multi-label feature selection methods ignore the label correlation. Therefore, we propose a Label Correlation Based Weighting Feature Selection Approach for Multi-Label Data, called MLLCWFS. It is a framework developed from traditional filtering feature selection methods for single-label data. To exploit the label correlation, we compute the importance of each label in mutual information, and adopt three weighting strategies to evaluate the correlation between features and labels. Extensive experiments conducted on four benchmark data sets using two base classifiers demonstrate that our approach is superior to the state-of-the-art feature selection algorithms for multi-label data.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Gu, Q.-Q., Li, Z.-H., Han, J.-W.: Correlated multi-label feature selection. In: Proceedings of the 20th ACM International Conference on Information And Knowledge management, pp. 1087–1096 (2011)
Jungjit, S., Freitas, A.A., Michaelis, M., Cinatl, J.: Two extensions to multi-label correlation-based feature selection: a case study in bioinformatics. In: Proceedings of the IEEE International Conference on Systems, Man, and Cybernetics, pp. 1519–1524 (2013)
Zhang, M.-L., Zhou, Z.-H.: A review on multi-label learning algorithms. IEEE Trans. Knowl. Date Eng. 26(8), 1819–1837 (2014)
Boutell, M.R., Luo, J.-B., Shen, X.-P., Brown, C.M.: Learning multi-label scene classification. Pattern Recogn. 37(9), 1757–1771 (2004)
McCallum, A.K.: Multi-label text classification with a mixture model trained by EM. In: Proceedings of the AAAI 1999 Workshop Text Learning (1999)
Ueda, N., Saito, K.: Parametric mixture models for multi-labeled text. Adv. neural Inf. Process. Syst. 15, 721–728 (2003)
Elisseeff, A., Weston, J.: A kernel method for multi-labelled classification. Adv. Neural Inf. Process. Syst. 14, 681–687 (2002)
Zhang, Y., Zhou, Z.-H.: Multilabel dimensionality reduction via dependence maximization. ACM Trans. Knowl. Discov. Data 4(3), 1–21 (2010). Article 14
Lee, J., Kim, D.-W.: Feature selection for multi-label classification using multivariate mutual information. Pattern Recogn. Lett. 34(3), 349–357 (2013)
Huang, S.-J., Zhou, Z.-H.: “Multi-label Learning by Exploiting Label Correlations Locally. In: Proceedings of the 26th AAAI Conference on Artificial Intelligence, pp. 949–955 (2012)
Spolaôr, N., Cherman, E.A., Lee, H.D.: ReliefF for multi-label feature selection. In: Brazilian Conference on Intelligent Systems, pp. 6–11 (2013)
Zhang, Z.-H., Li, S.-N., Li, Z.-G., Chen, H.: Multi-label feature selection algorithm based on information entropy. J. Comput. Res. Dev. 50(6), 1177–1184 (2013). (in Chinese)
Liu, H., Motoda, H.: Computational Methods of Feature Selection (Chapman & Hall/Crc Data Mining and Knowledge Discovery Series). Chapman & Hall/CRC, New York (2008)
Spolaôr, N., Monard, M.C., Lee, H.D., Tsoumakas, G.: Label construction for multi-label feature selection. In: Brazilian Conference on Intelligent Systems, pp. 247–252 (2014)
Peng, H.-C., Long, F.-H., Ding, C.: Feature selection based on mutual information criteria of max-dependency, max-relevance, and min-redundancy. IEEE Trans. Pattern Anal. Mach. Intell. 27(8), 1226–1238 (2005)
Tsoumakas, G., Katakis, I., Vlahavas, I.: Mining multi-label data. In: Maimon, O., Rokach, L. (eds.) Data Mining and Knowledge Discovery Handbook, pp. 667–685. Springer, Heidelberg (2010)
Doquire, G., Verleysen, M.: Feature selection for multi-label classification problems. In: Advances in Computational Intelligence, pp. 9–16 (2011)
Li, S.-N., Zhang, Z.-H., Duan, J.-Q.: An ensemble multi-label feature selection algorithm based on information entropy. Int. Arab J. Inf. Technol. 11(4), 379–386 (2014)
Zhang, M.-L., Zhou, Z.-H.: ML-kNN: A lazy learning approach to multi-label learning. Pattern Recogn. 40(7), 2038–2048 (2007)
Hall, M.A.: Correlation-based feature selection for discrete and numeric class machine learning. In: Proceedings of the 17th International Conference on Machine Learning, ICML-2000, pp. 359–366 (2000)
Hoque, N., Bhattacharyya, D.K., Kalita, J.K.: MIFS-ND: A mutual information-based feature selection method. Expert Syst. Appl. 41(14), 6371–6385 (2014)
Robnik-Šikonja, M., Kononenko, I.: Theoretical and empirical analysis of relief and RReliefF. Mach. Learn. 53(1), 23–69 (2003)
Quinlan, J.R.: C4.5: Programs for Machine Learning. Morgan Kaufmann Publishers Inc., San Francisco (1993)
Tsoumakas, G., Spyromitros-Xioufis, E., Vilcek, J., Vlahavas, I.: Mulan: A java library for multi-label learning. J. Mach. Learn. Res. 12, 2411–2414 (2011)
Acknowledgments
This work is supported in part by the National 973 Program of China under grant 2013CB329604, the Program for Changjiang Scholars and Innovative Research Team in University (PCSIRT) of the Ministry of Education, China, under grant IRT13059, the Specialized Research Fund for the Doctoral Program of Higher Education under grant 20130111110011, the Natural Science Foundation of China under grants (61503112, 61273292, 61273297, 61229301, 61305063), and the Specified Research Fund for the Doctoral Program of HFUT under grant JZ2015HGBZ0461.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2016 Springer International Publishing Switzerland
About this paper
Cite this paper
Liu, L., Zhang, J., Li, P., Zhang, Y., Hu, X. (2016). A Label Correlation Based Weighting Feature Selection Approach for Multi-label Data. In: Cui, B., Zhang, N., Xu, J., Lian, X., Liu, D. (eds) Web-Age Information Management. WAIM 2016. Lecture Notes in Computer Science(), vol 9659. Springer, Cham. https://doi.org/10.1007/978-3-319-39958-4_29
Download citation
DOI: https://doi.org/10.1007/978-3-319-39958-4_29
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-39957-7
Online ISBN: 978-3-319-39958-4
eBook Packages: Computer ScienceComputer Science (R0)