Skip to main content
Log in

An intelligent feature selection approach based on moth flame optimization for medical diagnosis

  • Original Article
  • Published:
Neural Computing and Applications Aims and scope Submit manuscript

Abstract

In this work, an enhanced moth flame optimization (MFO) algorithm is proposed as a search strategy within a wrapper feature selection (FS) framework. It aims mainly to improve the classification tasks in medical applications. FS is an NP-hard problem because the run time of its procedure grows exponentially. The persistent necessity for a powerful FS system and the promising results of swarming behavior in various optimization scenarios motivated our attempts to develop a new FS approach. This paper uses a recently developed methodology inspired by the moth movement style in searching for a near-optimal feature subset for reliable disease diagnosis. The proposed modification strategy is based on two stages of enhancement. In the first stage, eight binary variants are produced using eight transfer functions. In the second stage, the Levy flight operator is integrated into the structure of MFO in combination with transfer functions. The main target is to increase the diversity of the algorithm and support the exploration of the search space. Twenty-three medical data sets downloaded from UCI, Keel, Kaggle data repositories are used to validate the proposed approaches. It has been demonstrated that the proposed approach significantly outperforms other well-known wrapper approaches across 83% of data sets. Furthermore, the proposed approach outperforms other methods in the literature across 75% of the data sets. The comparisons with filter-based approach reveal superior performance across 70% of the data sets. This work also conducts an extensive investigation of the parameters and studies the effects of different settings on the performance of the FS process. The empirical results and various comparisons reveal a remarkable effect of the Levy flight operator and transfer functions on the performance of MFO. This can be achieved through improving the diversity, achieving a greater exploration/exploitation trade-off, and enhancing the convergence trends of the optimizer.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9

Similar content being viewed by others

References

  1. Yu L, Liu H (2003) Feature selection for high-dimensional data: A fast correlation-based filter solution. In: Proceedings of the 20th international conference on machine learning (ICML-03), pp 856–863

  2. Liu H, Motoda H (1998) Feature extraction, construction and selection: a data mining perspective, vol 453. Springer, Berlin

    Book  Google Scholar 

  3. Khurma RA, Aljarah I, Sharieh A, Mirjalili S (2020) Evolopy-fs: An open-source nature-inspired optimization framework in python for feature selection. In: Evolutionary machine learning techniques, Springer, pp 131–173

  4. Dash M, Liu H (1997) Feature selection for classification. Intell Data Anal 1:131–156

    Article  Google Scholar 

  5. Aljarah I, Mafarja M, Heidari AA, Faris H, Zhang Y, Mirjalili S (2018) Asynchronous accelerating multi-leader salp chains for feature selection. Appl Soft Comput 71:964–979

    Article  Google Scholar 

  6. Al-Madi N, Faris H, Abukhurma R (2018) Cost-sensitive genetic programming for churn prediction and identification of the influencing factors in telecommunication market. Int J Adv Sci Technol, pp 13–28

  7. Faris H, Abukhurma R, Almanaseer W, Saadeh M, Mora AM, Castillo PA, Aljarah I (2019) Improving financial bankruptcy prediction in a highly imbalanced class distribution using oversampling and ensemble learning: a case from the spanish market. Prog Artif Intell, pp 1–23

  8. Mirjalili S, Mirjalili SM, Yang X-S (2014) Binary bat algorithm. Neural Comput Appl 25:663–681

    Article  Google Scholar 

  9. Li S, Wu X, Tan M (2008) Gene selection using hybrid particle swarm optimization and genetic algorithm. Soft Comput 12:1039–1048

    Article  Google Scholar 

  10. El Aziz MA, Hassanien AE (2018) Modified cuckoo search algorithm with rough sets for feature selection. Neural Comput Appl 29:925–934

    Article  Google Scholar 

  11. Li Q, Chen H, Huang H, Zhao X, Cai Z, Tong C, Liu W, Tian X (2017) An enhanced grey wolf optimization based feature selection wrapped kernel extreme learning machine for medical diagnosis. Computational and mathematical methods in medicine 2017

  12. Faris H, Aljarah I, Alqatawna JF (2015) Optimizing feedforward neural networks using krill herd algorithm for e-mail spam detection. In: 2015 IEEE Jordan conference on applied electrical engineering and computing technologies (AEECT), IEEE, pp 1–5

  13. Khurma RA, Aljarah I, Sharieh A (2020) An efficient moth flame optimization algorithm using chaotic maps for feature selection in the medical applications. In: Proceedings of the 9th international conference on pattern recognition applications and methods - Volume 1: ICPRAM, INSTICC, SciTePress, pp 175–182

  14. Mirjalili S (2015) Moth-flame optimization algorithm: a novel nature-inspired heuristic paradigm. Knowl Based Syst 89:228–249

    Article  Google Scholar 

  15. Faris H, Aljarah I, Mirjalili S (2017) Evolving radial basis function networks using moth–flame optimizer. In: Handbook of neural computation. Academic Press, pp 537–550

  16. Khurma RA, Aljarah I, Sharieh A (2020) Rank based moth flame optimisation for feature selection in the medical application. In: 2020 IEEE congress on evolutionary computation (CEC). IEEE, pp 1–8

  17. Handels H, Roß T, Kreusch J, Wolff HH, Poeppl SJ (1999) Feature selection for optimized skin tumor recognition using genetic algorithms. Artif Intell Medi 16:283–297

    Article  Google Scholar 

  18. Babaoglu İ, Findik O, Ülker E (2010) A comparison of feature selection models utilizing binary particle swarm optimization and genetic algorithm in determining coronary artery disease using support vector machine. Expert Syst Appl 37:3177–3183

    Article  Google Scholar 

  19. Martinez E, Alvarez MM, Trevino V (2010) Compact cancer biomarkers discovery using a swarm intelligence feature selection algorithm. Comput Biol Chem 34:244–250

    Article  Google Scholar 

  20. Vieira SM, Mendonça LF, Farinha GJ, Sousa JM (2013) Modified binary pso for feature selection using svm applied to mortality prediction of septic patients. Appl Soft Comput 13:3494–3504

    Article  Google Scholar 

  21. Inbarani HH, Azar AT, Jothi G (2014) Supervised hybrid feature selection based on pso and rough sets for medical diagnosis. Comput Methods Programs Biomed 113:175–185

    Article  Google Scholar 

  22. Banka H, Dara S (2015) A hamming distance based binary particle swarm optimization (hdbpso) algorithm for high dimensional feature selection, classification and validation. Pattern Recognit Lett 52:94–100

    Article  Google Scholar 

  23. Ahmad F, Isa NAM, Hussain Z, Osman MK, Sulaiman SN (2015) A ga-based feature selection and parameter optimization of an ann in diagnosing breast cancer. Pattern Anal Appl 18:861–870

    Article  MathSciNet  Google Scholar 

  24. Sheikhpour R, Sarram MA, Sheikhpour R (2016) Particle swarm optimization for bandwidth determination and feature selection of kernel density estimation based classifiers in diagnosis of breast cancer. ApplSoft Comput 40:113–131

    MATH  Google Scholar 

  25. Gunasundari S, Janakiraman S, Meenambal S (2016) Velocity bounded boolean particle swarm optimization for improved feature selection in liver and kidney disease diagnosis. Expert Syst Appl 56:28–47

    Article  Google Scholar 

  26. Sayed GI, Hassanien AE, Nassef TM, Pan J-S (2016) Alzheimer’s disease diagnosis based on moth flame optimization. In: International conference on genetic and evolutionary computing, Springer, pp 298–305

  27. Wang M, Chen H, Yang B, Zhao X, Hu L, Cai Z, Huang H, Tong C (2017) Toward an optimal kernel extreme learning machine using a chaotic moth-flame optimization strategy with applications in medical diagnoses. Neurocomputing 267:69–84

    Article  Google Scholar 

  28. Sayed GI, Hassanien AE (2017) Moth-flame swarm optimization with neutrosophic sets for automatic mitosis detection in breast cancer histology images. Appl Intell 47:397–408

    Article  Google Scholar 

  29. Ibrahim HT, Mazher WJ, Ucan ON, Bayat O (2018) A grasshopper optimizer approach for feature selection and optimizing svm parameters utilizing real biomedical data sets. Neural Comput Appl, pp 1–10

  30. González-López J, Ventura S, Cano A (2019) Distributed selection of continuous features in multilabel classification using mutual information. IEEE Trans Neural Netw Learn Syst

  31. Gonzalez-Lopez J, Ventura S, Cano A (2020) Distributed multi-label feature selection using individual mutual information measures. Knowl Based Syst 188:105052

    Article  Google Scholar 

  32. Al-Betar MA, Alomari OA, Abu-Romman SM (2020) A triz-inspired bat algorithm for gene selection in cancer classification. Genomics 112:114–126

    Article  Google Scholar 

  33. Wolpert DH, Macready WG et al (1997) No free lunch theorems for optimization. IEEE Trans Evolut Comput 1:67–82

    Article  Google Scholar 

  34. Mirjalili S, Lewis A (2013) S-shaped versus v-shaped transfer functions for binary particle swarm optimization. Swarm Evolut Comput 9:1–14

    Article  Google Scholar 

  35. Kennedy J, Eberhart RC (1997) A discrete binary version of the particle swarm algorithm. In: IEEE International conference on systems, man, and cybernetics. Computational cybernetics and simulation, vol 5, IEEE, pp 4104–4108

  36. Rashedi E, Nezamabadi-Pour H, Saryazdi S (2010) Bgsa: binary gravitational search algorithm. Nat Comput 9:727–745

    Article  MathSciNet  Google Scholar 

  37. Sims DW, Humphries NE, Bradford RW, Bruce BD (2012) Lévy flight and brownian search patterns of a free-ranging predator reflect different prey field characteristics. J Animal Ecol 81:432–442

    Article  Google Scholar 

  38. Li Z, Zhou Y, Zhang S, Song J (2016) Lévy-flight moth-flame algorithm for function optimization and engineering design problems. Mathe Probl Eng 2016:1–22

    Google Scholar 

  39. Faris H, Aljarah I, Mirjalili S, Castillo PA, Guervós JJM (2016) EvoloPy: an open-source nature-inspired optimization framework in Python. In: IJCCI (ECTA), pp 171–177

  40. Derrac J, García S, Molina D, Herrera F (2011) A practical tutorial on the use of nonparametric statistical tests as a methodology for comparing evolutionary and swarm intelligence algorithms. Swarm Evolut Comput 1:3–18

    Article  Google Scholar 

  41. Xue B, Zhang M, Browne WN (2013) Novel initialisation and updating mechanisms in pso for feature selection in classification. In: European conference on the applications of evolutionary computation, Springer, pp 428–438

  42. Kashef S, Nezamabadi-pour H (2015) An advanced aco algorithm for feature subset selection. Neurocomputing 147:271–279

    Article  Google Scholar 

  43. Mafarja M, Sabar NR (2018) Rank based binary particle swarm optimisation for feature selection in classification. In: Proceedings of the 2nd international conference on future networks and distributed systems, ACM, p 19

  44. Mafarja M, Jarrar R, Ahmad S, Abusnaina AA (2018) Feature selection using binary particle swarm optimization with time varying inertia weight strategies. In: Proceedings of the 2nd international conference on future networks and distributed systems, ACM, p 18

  45. Emary E, Zawbaa HM, Hassanien AE (2016) Binary grey wolf optimization approaches for feature selection. Neurocomputing 172:371–381

    Article  Google Scholar 

  46. Mafarja M, Aljarah I, Heidari AA, Hammouri AI, Faris H, Ala’M A-Z, Mirjalili S (2018) Evolutionary population dynamics and grasshopper optimization approaches for feature selection problems. Knowl Based Syst 145:25–45

    Article  Google Scholar 

  47. Chen L-F, Su C-T, Chen K-H, Wang P-C (2012) Particle swarm optimization for feature selection with application in obstructive sleep apnea diagnosis. Neural Comput Appl 21:2087–2096

    Article  Google Scholar 

  48. Moradi P, Gholampour M (2016) A hybrid particle swarm optimization for feature subset selection by integrating a novel local search strategy. Appl Soft Comput 43:117–130

    Article  Google Scholar 

  49. Liu Y, Wang G, Chen H, Dong H, Zhu X, Wang S (2011) An improved particle swarm optimization for feature selection. J Bionic Eng 8:191–200

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Ibrahim Aljarah.

Ethics declarations

Conflict of Interest

The authors declare that they have no conflict of interest.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Appendix A. Data sets description

Appendix A. Data sets description

See Table 23

Table 23 Description of the used disease diagnosis data sets

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Abu Khurmaa, R., Aljarah, I. & Sharieh, A. An intelligent feature selection approach based on moth flame optimization for medical diagnosis. Neural Comput & Applic 33, 7165–7204 (2021). https://doi.org/10.1007/s00521-020-05483-5

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s00521-020-05483-5

Keywords

Navigation