Skip to main content
Log in

The monarch butterfly optimization algorithm for solving feature selection problems

  • S.I. : Healthcare Analytics
  • Published:
Neural Computing and Applications Aims and scope Submit manuscript

Abstract

Feature selection (FS) is considered to be a hard optimization problem in data mining and some artificial intelligence fields. It is a process where rather than studying all of the features of a whole dataset, some associated features of a problem are selected, the aim of which is to increase classification accuracy and reduce computational time. In this paper, a recent optimization algorithm, the monarch butterfly optimization (MBO) algorithm, is implemented with a wrapper FS method that uses the k-nearest neighbor (KNN) classifier. Experiments were implemented on 18 benchmark datasets. The results showed that, in comparison with four metaheuristic algorithms (WOASAT, ALO, GA and PSO), MBO was superior, giving a high rate of classification accuracy of, on average, 93% for all datasets as well as reducing the selection size significantly. Therefore, the use of the MBO to solve the FS problems has been proven through the results obtained to be effective and highly efficient in this field, and the results have also proven the strength of the balance between global and local search of MBO.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Institutional subscriptions

Fig. 1
Fig. 2
Fig. 3
Fig. 4

Similar content being viewed by others

References

  1. Arora S, Anand P (2019) Binary butterfly optimization approaches for feature selection. Expert Syst Appl 116:147–160

    Google Scholar 

  2. Bennasar M, Hicks Y, Setchi R (2015) Feature selection using joint mutual information maximisation. Expert Syst Appl 42:8520–8532

    Google Scholar 

  3. Teisseyre P, Zufferey D, Słomka M (2019) Cost-sensitive classifier chains: selecting low-cost features in multi-label classification. Pattern Recogn 86:290–319

    Google Scholar 

  4. Alweshah M, Abdullah S (2015) Hybridizing firefly algorithms with a probabilistic neural network for solving classification problems. Appl Soft Comp 35:513–524

    Google Scholar 

  5. Alweshah M (2018) Construction biogeography-based optimization algorithm for solving classification problems. Neural Comp Appl 31(10):1–10

    Google Scholar 

  6. Singh HR, Biswas SK, Bordoloi M (2019) Recent neuro-fuzzy approaches for feature selection and classification. In: Sarfraz M (ed) Exploring critical approaches of evolutionary computation, ed: IGI Global, pp 1–19

  7. Liu H, Yu L (2005) Toward integrating feature selection algorithms for classification and clustering. IEEE Trans knowledge Data Eng 17:491–502

    Google Scholar 

  8. Liu H, Motoda H (2012) Feature selection for knowledge discovery and data mining. Springer, Berlin

    MATH  Google Scholar 

  9. Yang Y, Pedersen JO (1997) A comparative study on feature selection in text categorization. In: DH Fisher (ed) Icml, pp 412–420

  10. Peng H, Long F, Ding C (2005) Feature selection based on mutual information criteria of max-dependency, max-relevance, and min-redundancy. IEEE Trans Pattern Anal Mach Intell 27:1226–1238

    Google Scholar 

  11. Han J, Pei J, Kamber M (2011) Data mining: concepts and techniques. Elsevier, Amsterdam

    MATH  Google Scholar 

  12. Yan C, Ma J, Luo H, Patel A (2019) Hybrid binary coral reefs optimization algorithm with simulated annealing for feature selection in high-dimensional biomedical datasets. Chemometr Intell Laborat Syst 184:102–111

    Google Scholar 

  13. Yuan M, Yang Z, Ji G (2019) Partial maximum correlation information: a new feature selection method for microarray data classification. Neurocomputing 323:231–243

    Google Scholar 

  14. Dash M, Liu H (1997) Feature selection for classification. Intell Data Anal 1:131–156

    Google Scholar 

  15. Yusta SC (2009) Different metaheuristic strategies to solve the feature selection problem. Pattern Recogn Lett 30:525–534

    Google Scholar 

  16. Tahir MA, Smith J (2010) Creating diverse nearest-neighbour ensembles using simultaneous metaheuristic feature selection. Pattern Recogn Lett 31:1470–1480

    Google Scholar 

  17. Kumar L, Bharti KK (2019) An improved BPSO algorithm for feature selection. In: Khare A, Tiwary US, Sethi IK, Singh N (eds) Recent trends in communication, computing, and electronics, ed: Springer, pp 505–513

  18. Yang XS (2010) Nature-inspired metaheuristic algorithms: Luniver press

  19. Osman IH, Kelly JP (1996) Meta-heuristics: an overview. In: Osman IH, Kelly JP (eds) Meta-heuristics, ed: Springer, pp 1–21

  20. Stützle T, López-Ibáñez M (2019) Automated design of metaheuristic algorithms. In: Gendreau M, Potvin JY (eds) Handbook of Metaheuristics, ed: Springer, pp 541–579

  21. Ahmad SR, Bakar AA, Yaakub MR (2015) Metaheuristic algorithms for feature selection in sentiment analysis. Sci Inf Conf (SAI) 2015:222–226

    Google Scholar 

  22. Kannan S, Slochanal SMR, Padhy NP (2005) Application and comparison of metaheuristic techniques to generation expansion planning problem. IEEE Trans Power Syst 20:466–475

    Google Scholar 

  23. Emary E, Zawbaa HM, Grosan C, Hassenian AE (2015) Feature subset selection approach by gray-wolf optimization. In: Afro-European Conference for Industrial Advancement, pp 1–13

  24. Alweshah M, Hammouri AI, Tedmori S (2017) Biogeography-based optimisation for data classification problems. Int J Data Mining Modell Manag 9:142–162

    Google Scholar 

  25. Zhang Y, Gong D, Hu Y, Zhang W (2015) Feature selection algorithm based on bare bones particle swarm optimization. Neurocomputing 148:150–157

    Google Scholar 

  26. Alweshah M, Ramadan E, Ryalat MH, Almi’ani M, Hammouri AI (2020) Water evaporation algorithm with probabilistic neural network for solving classification problems. Jordanian J Comput Inf Technol (JJCIT) 6(14):2020

    Google Scholar 

  27. Zorarpacı E, Özel SA (2016) A hybrid approach of differential evolution and artificial bee colony for feature selection. Expert Syst Appl 62:91–103

    Google Scholar 

  28. Wang Y, Liu Y, Feng L, Zhu X (2015) Novel feature selection method based on harmony search for email classification. Knowledge-Based Syst 73:311–323

    Google Scholar 

  29. Lin K-C, Zhang K-Y, Huang Y-H, Hung JC, Yen N (2016) Feature selection based on an improved cat swarm optimization algorithm for big data classification. J Supercomput 72:3210–3221

    Google Scholar 

  30. Ghareb AS, Bakar AA, Hamdan AR (2016) Hybrid feature selection based on enhanced genetic algorithm for text categorization. Expert Syst Appl 49:31–47

    Google Scholar 

  31. Lin S-W, Lee Z-J, Chen S-C, Tseng T-Y (2008) Parameter determination of support vector machine and feature selection using simulated annealing approach. Appl Soft Comput 8:1505–1512

    Google Scholar 

  32. Mohammed Al-Weshah SAK, Almomani A, Al-Refai M, Qashi R (2019) Metaheuristic algorithms based feature selection approach for intrusion detection. In: Brij QZS, Gupta B (eds) Machine learning for computer and cyber security: principle, algorithms, and practices. Taylor & Francis, USA

    Google Scholar 

  33. Al Nsour H, Alweshah M, Hammouri AI, Al Ofeishat H, Mirjalili S (2019) A hybrid grey wolf optimiser algorithm for solving time series classification problems. J Intell Syst 29(1):846–857

    Google Scholar 

  34. Alshareef AM, Bakar AA, Hamdan AR, Abdullah SMS, Alweshah M (2015) A case-based reasoning approach for pattern detection in Malaysia rainfall data. Int J Big Data Intell 2:285–302

    Google Scholar 

  35. Alweshah M (2018) Construction biogeography-based optimization algorithm for solving classification problems. Neural Comput Appl 29:1–10

    Google Scholar 

  36. Alweshah M, Alzubi OA, Alzubi JA, Alaqeel S (2016) Solving attribute reduction problem using wrapper genetic programming,”. Int J Comput Sci Netw Secur (IJCSNS) 16:77

    Google Scholar 

  37. Alweshah M, Hammouri AI, Rashaideh H, Ababneh M, Tayyeb H (2017) Solving time series classification problems using combined of support vector machine and neural network. Int J Data Anal Tech Strat 9:2017

    Google Scholar 

  38. Wang GG, Zhao X, Deb S (2015) A novel monarch butterfly optimization with greedy strategy and self-adaptive. In: Soft computing and machine intelligence (ISCMI), 2015 Second international conference on, pp 45–50

  39. Feng Y, Wang G-G, Li W, Li N (2018) Multi-strategy monarch butterfly optimization algorithm for discounted 0–1 knapsack problem. Neural Comput Appl 30:3019–3036

    Google Scholar 

  40. Altman NS (1992) An introduction to kernel and nearest-neighbor nonparametric regression. Am Stat 46:175–185

    MathSciNet  Google Scholar 

  41. Afifi AA, Azen SP (1979) Statistical analysis: a computer oriented approach. Academic Press Inc, Orlando

    MATH  Google Scholar 

  42. Selvakumar B, Muneeswaran K (2019) Firefly algorithm based feature selection for network intrusion detection. Comput Secur 81:148–155

    Google Scholar 

  43. Ghosh M, Malakar S, Bhowmik S, Sarkar R, Nasipuri M (2019) feature selection for handwritten word recognition using memetic algorithm. In: Mandal JK, Dutta P, Mukhopadhyay S (eds) Advances in intelligent computing, ed: Springer, pp 103–124

  44. Goswami S, Chakraborty S, Guha P, Tarafdar A, Kedia A (2019) Filter-Based Feature Selection Methods Using Hill Climbing Approach. In: Li X, Wong, KC (eds) Natural computing for unsupervised learning, ed: Springer, pp 213–234

  45. Zawbaa HM, Emary E, Parv B (2015) Feature selection based on antlion optimization algorithm. In: Complex systems (WCCS), 2015 Third World Conference on, 2015, pp 1–7

  46. Sabeena S, Sarojini B (2015) Optimal feature subset selection using ant colony optimization. Indian J Sci Technol 8:1–5

    Google Scholar 

  47. Wan Y, Wang M, Ye Z, Lai X (2016) A feature selection method based on modified binary coded ant colony optimization algorithm. Appl Soft Comp 49:248–258

    Google Scholar 

  48. Aghdam MH, Kabiri P (2016) Feature selection for intrusion detection system using ant colony optimization. IJ Netw Secur 18:420–432

    Google Scholar 

  49. Wu S (2015) Comparative analysis of particle swarm optimization algorithms for text feature selection. In: Master’s Projects. 386. https://doi.org/10.31979/etd.k4cc-tvzq. https://scholarworks.sjsu.edu/etd_projects/386

  50. Samsani S, Suma GJ (2016) A binary approach of artificial bee colony optimization technique for feature subset selection

  51. Ghanem WAH, Jantan A (2016) Novel multi-objective artificial bee Colony optimization for wrapper based feature selection in intrusion detection. Int J Adv Soft Comp Appl 8:1–12

    Google Scholar 

  52. Zawbaa HM, Emary E, Grosan C (2016) Feature selection via chaotic antlion optimization. PLoS ONE 11:e0150652

    Google Scholar 

  53. Wang J, Xue B, Gao X, Zhang M (2016) A differential evolution approach to feature selection and instance selection. In: Pacific Rim International Conference on Artificial Intelligence, pp 588–602

  54. Shahbeig S, Sadjad K, Sadeghi M (2016) Feature selection from iron direct reduction data based on binary differential evolution optimization. Bull de la Société Royale des Sciences de Liège 85:114–122

    Google Scholar 

  55. Mafarja M, Aljarah I, Heidari AA, Hammouri AI, Faris H, Ala’M M, Mirjalili S (2018) Evolutionary population dynamics and grasshopper optimization approaches for feature selection problems. Knowledge-Based Syst 145:25–45

    Google Scholar 

  56. Barbu A, She Y, Ding L, Gramajo G (2017) Feature selection with annealing for computer vision and big data learning. IEEE Trans Pattern Anal Mach Intell 39:272–286

    Google Scholar 

  57. Cerrada M, Sánchez RV, Cabrera D, Zurita G, Li C (2015) Multi-stage feature selection by using genetic algorithms for fault diagnosis in gearboxes based on vibration signal. Sensors 15:23903–23926

    Google Scholar 

  58. Aalaei S, Shahraki H, Rowhanimanesh A, Eslami S (2016) Feature selection using genetic algorithm for breast cancer diagnosis: experiment on three different datasets. Iran J Basic Med Sci 19:476

    Google Scholar 

  59. Malakar S, Ghosh M, Bhowmik S, Sarkar R, Nasipuri M (2019) A GA based hierarchical feature selection approach for handwritten word recognition. Neural Comp Appl 32(7):1–20

    Google Scholar 

  60. Saidi R, Bouaguel W, Essoussi N (2019) Hybrid feature selection method based on the genetic algorithm and pearson correlation coefficient. In: Hassanien AE (ed) Machine learning paradigms: theory and application, ed: Springer, pp 3–24

  61. Emary E, Zawbaa HM, Hassanien AE (2016) Binary ant lion approaches for feature selection. Neurocomputing 213:54–65

    Google Scholar 

  62. Basiri ME, Nemati S (2009) A novel hybrid ACO-GA algorithm for text feature selection. In: Tyrrell A, Sarfraz M (eds) Evolutionary computation, CEC’09. IEEE congress on, 2009, Kuwait University, Kuwait, pp 2561–2568

  63. Jona J, Nagaveni N (2014) Ant-cuckoo colony optimization for feature selection in digital mammogram. Pak J Biol Sci PJBS 17:266–271

    Google Scholar 

  64. Babatunde R, Olabiyisi S, Omidiora E (2014) Feature dimensionality reduction using a dual level metaheuristic algorithm. Optimization 7:49–52

    Google Scholar 

  65. Mafarja M, Abdullah S (2013) Investigating memetic algorithm in solving rough set attribute reduction. Int J Comput Appl Technol 48:195–202

    Google Scholar 

  66. Azmi R, Pishgoo B, Norozi N, Koohzadi M, Baesi F (2010) A hybrid GA and SA algorithms for feature selection in recognition of hand-printed Farsi characters. In: Intelligent Computing and Intelligent Systems (ICIS), IEEE International Conference on, 2010, pp. 384-387

  67. Olabiyisi SO, Fagbola TM, Omidiora EO, Oyeleye AC (2012) Hybrid metaheuristic feature extraction technique forsolving timetabling problem.Int. J Sci Engi Res 3(8):1–6

    Google Scholar 

  68. Chen Z, LinT Tang N, Xia X (2016) A parallel genetic algorithm based feature selection and parameter optimization for support vector machine. Sci Programm 2016:1–11

    Google Scholar 

  69. Alzaqebah M, Alrefai N, Ahmed EA, Jawarneh S, Alsmadi MK (2020) Neighborhood search methods with Moth Optimization algorithm as a wrapper method for feature selection problems. Int J Electr Comp Eng 10:3672

    Google Scholar 

  70. Too J, Rahim Abdullah A (2020) Binary atom search optimisation approaches for feature selection. Conn Sci. https://doi.org/10.1080/09540091.2020.1741515

    Article  Google Scholar 

  71. Faris H, Hassonah MA, Ala’M AZ, Mirjalili S, Aljarah I (2018) A multi-verse optimizer approach for feature selection and optimizing SVM parameters based on a robust system architecture. Neural Comput Appl 30:2355–2369

    Google Scholar 

  72. Jain K, Bhadauria SS (2016) Enhanced content based image retrieval using feature selection using teacher learning based optimization. Int J Comput Sci Inf Secur (IJCSIS) 14:1052–1057

    Google Scholar 

  73. Pashaei E, Aydin N (2017) Binary black hole algorithm for feature selection and classification on biological data. Appl Soft Comput 56:94–106

    Google Scholar 

  74. Sreeja N (2019) A weighted pattern matching approach for classification of imbalanced data with a fireworks-based algorithm for feature selection. Conn Sci 31:143–168

    Google Scholar 

  75. Tuba E,. Strumberger I, Bacanin N, Jovanovic R, Tuba M (2019) Bare bones fireworks algorithm for feature selection and SVM optimization. In: 2019 IEEE Congress on Evolutionary Computation (CEC), pp 2207–2214

  76. Sindhu R, Ngadiran R, Yacob YM, Hanin Zahri NA, Hariharan M, Polat K (2019) A hybrid SCA inspired BBO for feature selection problems. Math Prob Eng 2019:1–18

    Google Scholar 

  77. Emary E, Zawbaa HM, Ghany KKA, Hassanien AE, Parv B (2015) Firefly optimization algorithm for feature selection. In: Proceedings of the 7th balkan conference on informatics conference, pp 1–7

  78. Alweshah M, Al-Sendah M, Dorgham OM, Al-Momani A, Tedmori S (2020) Improved water cycle algorithm with probabilistic neural network to solve classification problems. Cluster Comp. https://doi.org/10.1007/s10586-019-03038-5

    Article  Google Scholar 

  79. Alweshah M, Qadoura MA, Hammouri AI, Azmi MS, AlKhalaileh S (2020) Flower pollination algorithm for solving classification problems. Int J Adv Soft Comp Appl 12(1):15–34

    Google Scholar 

  80. Alzubi OA, Alzubi JA, Alweshah M, Qiqieh I, Al-Shami S, Ramachandran M (2020) An optimal pruning algorithm of classifier ensembles: dynamic programming approach. Neural Comp Appl. https://doi.org/10.1007/s00521-020-04761-6

    Article  Google Scholar 

  81. Mafarja MM, Mirjalili S (2017) Hybrid Whale Optimization Algorithm with simulated annealing for feature selection. Neurocomputing 260:302–312

    Google Scholar 

  82. Chakrabarty S, Pal AK, Dey N, Das D, Acharjee S (2014) Foliage area computation using Monarch butterfly algorithm. In: Non conventional energy (ICONCE), 2014 1st International conference on, 2014, pp 249–253

  83. Feng Y, Yang J, Wu C, Lu M, Zhao X-J (2018) Solving 0–1 knapsack problems by chaotic monarch butterfly optimization algorithm with Gaussian mutation. Memetic Comp 10:135–150

    Google Scholar 

  84. Ghanem WA, Jantan A (2018) Hybridizing artificial bee colony with monarch butterfly optimization for numerical optimization problems. Neural Comp Appl 30:163–181

    Google Scholar 

  85. Sambariya D, Gupta T (2017) Optimal design of PID controller for an AVR system using monarch butterfly optimization. In: Information, communication, instrumentation and control (ICICIC), 2017 International Conference on, 2017, pp 1–6

  86. Devikanniga D, Raj RJS (2018) Classification of osteoporosis by artificial neural network based on monarch butterfly optimisation algorithm. Healthcare Technol Lett 5:70–75

    Google Scholar 

  87. Strumberger I, Sarac M, Markovic D, Bacanin N (2018) Hybridized monarch butterfly algorithm for global optimization problems. Int J Comp 3:63–68

    Google Scholar 

  88. Faris H, Aljarah I, Mirjalili S (2018) Improved monarch butterfly optimization for unconstrained global search and neural network training. Appl Intell 48:445–464

    Google Scholar 

  89. Yazdani S, Hadavandi E (2019) LMBO-DE: a linearized monarch butterfly optimization algorithm improved with differential evolution. Soft Comp 23:8029–8043. https://doi.org/10.1007/s00500-018-3439-8

    Article  Google Scholar 

  90. Stromberger I, Tuba E, Bacanin N, Beko M, Tuba M (2018) Monarch butterfly optimization algorithm for localization in wireless sensor networks. In: Radioelektronika (RADIOELEKTRONIKA), 2018 28th International Conference, pp 1-6

  91. Mafarja MM, Mirjalili S (2017) Hybrid Whale Optimization Algorithm with simulated annealing for feature selection. Neurocomputing 260:302–312

    Google Scholar 

  92. Blake CL, Merz CJ (1998) UCI Repository of machine learning databases [http://wwwics.uci.edu/~mlearn/MLRepository.html]. Irvine, CA: University of California, Department of Information and Computer Science, vol 55, pp 12–21. Accessed 2019

Download references

Acknowledgements

The research reported in this publication was supported by the Deanship of Scientific Research at Al-Balqa Applied University in Jordan.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Mohammed Alweshah.

Ethics declarations

Conflict of interest

The authors certify that they have no affiliations with or involvement in any organization or entity with any financial interest, or non-financial interest in the subject matter or materials discussed in this manuscript.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Alweshah, M., Khalaileh, S.A., Gupta, B.B. et al. The monarch butterfly optimization algorithm for solving feature selection problems. Neural Comput & Applic 34, 11267–11281 (2022). https://doi.org/10.1007/s00521-020-05210-0

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s00521-020-05210-0

Keywords

Navigation