Abstract
Feature selection is an important data preprocessing method before classification. Multi-objective optimization algorithms have been proved an effective way to solve feature selection problems. However, there are few studies on multi-objective optimization feature selection methods for multi-label data. In this paper, a multi-objective multi-label filter feature selection algorithm based on two particle swarms (MOMFS) is proposed. We use mutual information to measure the relevance between features and label sets, and the redundancy between features, which are taken as two objectives. In order to avoid Particle Swarm Optimization (PSO) from falling into the local optimum and obtaining a false Pareto front, we employ two swarms to optimize the two objectives separately and propose an improved hybrid topology based on particle’s fitness value. Furthermore, an archive maintenance strategy is introduced to maintain the distribution of archive. In order to study the effectiveness of the proposed algorithm, we select five multi-label evaluation criteria and perform experiments on seven multi-label data sets. MOMFS is compared with classic single-objective multi-label feature selection algorithms, multi-objective filter and wrapper feature selection algorithms. The experimental results show that MOMFS can effectively reduce the multi-label data dimension and perform better than other approaches on five evaluation criteria.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.References
Zhang M, Zhou Z (2007) ML-KNN: a lazy learning approach to multi-label learning. Pattern Recogn 40(7):2038–2048
Tahir MA, Kittler J, Bouridane A (2012) Multilabel classification using heterogeneous ensemble of multi-label classifiers. Pattern Recogn Lett 33(5):513–523
Stojanova D, Ceci M, Malerba D, Dzeroski S (2013) Using PPI network autocorrelation in hierarchical multi-label classification trees for gene function prediction. BMC Bioinformatics 14(1):285
Oramas S, Nieto O, Barbieri F et al (2017) Multi-label music genre classification from audio, text, and images using deep features. In: proceedings of the 18th International Society for Music Information Retrieval Conference, pp 23-30
Wang J, Yang Y, Mao J et al (2016) Cnn-rnn: a unified framework for multi-label image classification. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 2285–2294
Zhao Z, Morstatter F, Sharma S et al (2010) Advancing feature selection research. ASU feature selection repository 1–28
Rejer I, Twardochleb M (2018) Gamers’ involvement detection from EEG data with cGAAM-A method for feature selection for clustering. Expert Syst Appl 101:196–204
Lin S, Zhe Z, Xing L (2016) Particle swarm optimization-based feature selection in sentiment classification. Soft Comput 20(10):1–14
Mistry K, Zhang L, Neoh SC, Lim CP, Fielding B (2017) A micro-GA embedded PSO feature selection approach to intelligent facial emotion recognition. IEEE Transs Cybernetics 47(6):1496–1509
Fernandez-Lozano C, Seoane JA, Gestal M, Gaunt TR, Dorado J, Campbell C (2015) Texture classification using feature selection and kernel-based techniques. Soft Comput 19(9):2469–2480
Unler A, Murat A, Chinnam RB (2011) mr2 PSO: a maximum relevance minimum redundancy feature selection method based on swarm intelligence for support vector machine classification. Inf Sci 181(20):4625–4641
Xue B, Zhang M, Browne WN, Yao X (2016) A survey on evolutionary computation approaches to feature selection. IEEE Trans Evol Comput 20(4):606–626
Xue B, Zhang M, Browne WN (2013) Particle swarm optimization for feature selection in classification: a multi-objective approach. IEEE Trans Cybernetics 43(6):1656–1671
Tran, B, Xue B, Zhang M (2014) Overview of particle swarm optimisation for feature selection in classification. In: Asia-Pacific conference on simulated evolution and learning, pp 605–617
Zhang Y, Wang S, Phillips P, Ji G (2014) Binary PSO with mutation operator for feature selection using decision tree applied to spam detection. Knowl-Based Syst 64(1):22–31
Dong H, Sun J, Li T, Li L et al (2018) An improved niching binary particle swarm optimization for feature selection. In: 2018 IEEE international conference on systems, man, and cybernetics, pp 3571-3577
Yadav S, Ekbal A, Saha S (2017) Feature selection for entity extraction from multiple biomedical corpora: a PSO-based approach. Soft Comput 22(20):6881–6904
Teng X, Dong H, Sun J (2017) Co-evolutionary algorithm for feature selection. CAAI Trans Intelligent Syst 01:24–31
Feng T, Fu X, Zhang Y et al (2008) A genetic algorithm-based method for feature subset selection. Soft Comput 12(2):111–120
Dong H, Li T, Ding R, Sun J (2018) A novel hybrid genetic algorithm with granular information for feature selection and optimization. Appl Soft Comput 65:33–46
Khushaba RN, Al-Ani A, Al-Jumaily A (2011) Feature subset selection using differential evolution and a statistical repair mechanism. Expert Syst Appl 38(9):11515–11526
Xue B, Fu W, Zhang M (2014) Differential evolution (DE) for multi-objective feature selection in classification. In: Proceedings of the Companion Publication of the 2014 Annual Conference on Genetic and Evolutionary Computation, pp 83–84
Zorarpacı E, Özel SA (2016) A hybrid approach of differential evolution and artificial bee colony for feature selection. Expert Syst Appl 62:91–103
Li J, Cheng K, Wang S et al (2016) Feature selection: a data perspective. ACM Comput Surv 50(6):1–45
Xiong N (2002) A hybrid approach to input selection for complex processes. IEEE Transactions on Systems, Man, and Cybernetics-Part A: Systems and Humans 32(4):532–536
Mao KZ (2004) Feature subset selection for support vector machines through discriminative function pruning analysis. IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics) 34(1):60–67
Hsu CN, Huang H, Dietrich S (2002) The ANNIGMA-wrapper approach to fast feature selection for neural nets. IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics) 32(2):207–212
Zhang D, Chen S, Zhou Z (2008) Constraint score: a new filter method for feature selection with pairwise constraints. Pattern Recogn 41(5):1440–1451
Lei Y, Liu H (2003) Feature selection for high-dimensional data: a fast correlation-based filter solution. In: proceedings of the 20th international conference on machine learning, pp 856-863
Dash M, Choi K, Scheuermann P et al (2002) Feature selection for clustering-a filter solution. In: IEEE International Conference on Data Mining, pp 115–122
Hancer E, Xue B, Zhang M et al (2017) Pareto front feature selection based on artificial bee Colony optimization. Inf Sci 422:462–479
Spolaôr N, Lorena AC, Lee HD (2018) Feature selection via Pareto multi-objective genetic algorithms. Appl Artif Intell 31(32):1–28
Sikdar UK, Ekbal A, Saha S (2015) MODE: multiobjective differential evolution for feature selection and classifier ensemble. Soft Comput 19(12):3529–3549
Xue B, Zhang M, Browne WN (2013) Particle swarm optimization for feature selection in classification: a multi-objective approach. IEEE Trans Cybernetics 43(6):1656–1671
Jiménez F, Pérez-Sánchez H, Palma J, Sánchez G, Martínez C (2019) A methodology for evaluating multi-objective evolutionary feature selection for classification in the context of virtual screening. Soft Comput 23(18):8775–8800
Zhang Y, Gong D, Sun X, Guo YN (2017) A PSO-based multi-objective multi-label feature selection method in classification. Sci Rep 7(1):376
Khan M A, Ekbal A, Eneldo Loza Mencía et al (2017) Multi-objective optimisation-based feature selection for multi-label classification. In: International Conference on Applications of Natural Language to Information Systems, pp 38–41
Zhang Y, Gong D, Rong M (2015) Multi-objective differential evolution algorithm for multi-label feature selection in classification. In: International Conference in Swarm Intelligence, pp 339–345
Yin J, Tao T, Xu J (2015) A multi-label feature selection algorithm based on multi-objective optimization. In: International Joint Conference on Neural Networks, pp 1–7
Zhan Z, Li J, Cao J et al (2013) Multiple populations for multiple objectives: a Coevolutionary technique for solving multiobjective optimization problems. IEEE Trans Syst Man Cybernetics Part B 43(2):445–463
Kennedy J (2002) Particle swarm optimization. In: proceedings of the ICNN95-international conference on neural networks, pp 1942-1948
Eberhart RC, Shi Y (2007) Chapter four - evolutionary computation implementations. Computational intelligence 95-143
Kalita DJ, Singh S (2020) SVM hyper-parameters optimization using quantized multi-PSO in dynamic environment. Soft Comput 24(2):1225–1241
Xuan T, Siarry P, Oulhadj H (2018) Integrating fuzzy entropy clustering with an improved PSO for MRI brain image segmentation. Appl Soft Comput 65:230–242
Zhao L, Wei J (2019) A nested particle swarm algorithm based on sphere mutation to solve bi-level optimization. Soft Comput 23(21):11331–11341
Guo Z, Quan L (2013) Financial time series forecasting using LPP and SVM optimized by PSO. Soft Comput 17(5):805–818
Kennedy J, Eberhart RC (1997) A discrete binary version of the particle swarm algorithm. In: International conference on systems, man, and cybernetics, pp 4104–4108
Kennedy J (1999) Small worlds and mega-minds: effects of neighborhood topology on particle swarm performance. In: proceedings of the 1999 congress on evolutionary computation, pp 1931-1938
Liang J, Qin A, Suganthan PN et al (2006) Comprehensive learning particle swarm optimizer for global optimization of multimodal functions. IEEE Trans Evol Comput 10(3):281–295
Shannon CE (1948) A mathematical theory of communication. Bell Labs Tech J 27(3):379–423
Battiti R (1994) Using mutual information for selecting features in supervised neural net learning. IEEE Trans Neural Netw 5(4):537–550
Liu H, Sun J, Lei L et al (2009) Feature selection with dynamic mutual information. Pattern Recogn 42(7):1330–1339
Bostani H, Sheikhan M (2015) Hybrid of binary gravitational search algorithm and mutual information for feature selection in intrusion detection systems. Soft Comput 21(9):1–18
Sluga D, Lotrič U (2017) Quadratic mutual information feature selection. Entropy 19(4):157–173
Bidgoli, A. A., Rahnamayan, S., Ebrahimpour-Komleh, H. (2019). Opposition-based multi-objective binary differential evolution for multi-label feature selection. In: international conference on evolutionary multi-criterion optimization, pp 553–564
Raquel CR, Naval Jr PC (2005) An effective use of crowding distance in multiobjective particle swarm optimization. In:Proceedings of the 7th annual conference on Genetic and evolutionary computation, pp 257–264
Tsoumakas G, Spyromitros-Xioufis E, Vilcek J et al (2011) Mulan: a java library for multi-label learning. J Mach Learn Res 12:2411–2414
Zitzler E, Thiele L. Multiobjective Optimization Using Evolutionary Algorithms-A Comparative Study. In: International conference on parallel problem solving from nature, pp 292–301
Spolaôr N, Cherman EA, Monard MC, Lee HD (2013) A comparison of multi-label feature selection methods using the problem transformation approach. Electronic Notes Theoretical Comput Sci 292:135–151
Peng H, Long F, Ding C (2005) Feature selection based on mutual information: criteria of max-dependency, max-relevance, and min-redundancy. IEEE Trans Pattern Anal Mach Intell 27(8):1226–1238
Lin Y, Hu Q, Liu J, Duan J (2015) Multi-label feature selection based on max-dependency and min-redundancy. Neurocomputing 168:92–103
Huang R, Jiang W, Sun G (2018) Manifold-based constraint Laplacian score for multi-label feature selection. Pattern Recogn Lett 112:346–352
Zhang J, Luo Z, Li C, Zhou C, Li S (2019) Manifold regularized discriminative feature selection for multi-label learning. Pattern Recogn 95:136–150
Deb K, Pratap A, Agarwal S, Meyarivan T (2002) A fast and elitist multiobjective genetic algorithm: NSGA-II. IEEE Trans Evol Comput 6(2):182–197
Coello CC, Lechuga MS (2002) MOPSO: a proposal for multiple objective particle swarm optimization. In: proceedings of the 2002 congress on evolutionary computation pp 1051-1056
Acknowledgments
This work was supported by the National Science Foundation of China (No. 61472095), Preparatory Research Foundation of Education Department of Heilongjiang (1354MSYYB003) and Research Foundation of Mudanjiang Normal University (YB2020010).
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
Hongbin Dong declares that he has no conflict of interest. Jing Sun declares that she has no conflict of interest. Tao Li declares that he has no conflict of interest. Rui Ding declares that she has no conflict of interest. Xiaohang Sun declares that he has no conflict of interest.
Ethical approval
This article does not contain any studies with human participants or animals performed by any of the authors.
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Dong, H., Sun, J., Li, T. et al. A multi-objective algorithm for multi-label filter feature selection problem. Appl Intell 50, 3748–3774 (2020). https://doi.org/10.1007/s10489-020-01785-2
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10489-020-01785-2