Abstract
Feature selection, which is a commonly used data prepossessing technique, focuses on improving model performance and efficiency by removing redundant or irrelevant features. However, an implicit assumption made by traditional feature selection approaches is that data are independent and identically distributed (IID). To further obtain more complex and significant information, an effective feature selection construction should consider the couplings (non-IIDness) contained within feature values and relevance between features. Hence, referring to rough set theory, this paper first introduces a new coupled similarity measure to discover the value-to-feature-to-class coupling information, which can be used to calculate object neighbor and update feature weights. Second, using mutual information, a new coupled relevance measure is defined to capture the feature-to-feature coupling relationships. On this basis, an effective feature-selection algorithm based on coupling learning is developed for categorical data. To demonstrate the proposed algorithm, four common classifiers and 12 UCI data sets are employed in the experiments. The experimental results confirm the feasibility of the new algorithm and its effectiveness.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.References
Cao LB, Ou Y, Yu PS (2012) Coupled behavior analysis with applications. IEEE Trans Knowl Data Eng 24(8):1378–1392
Cao LB (2014) Non-IIDness learning in behavioral and social data. Comput J 57(9):1358–1370
Zhu CZ, Cao LB, Yin JP (2020) Unsupervised heterogeneous coupling learning for categorical representation. IEEE Trans Pattern Anal Mach Intell. https://doi.org/10.1109/TPAMI.2020.3010953
Dash M, Liu H (2003) Consistency-based search in feature selection. Artif Intell 151:155–176
Guyon I, Elisseeff A (2003) An introduction to variable feature selection. Mach Learn Res 3:1157–1182
Hu QH, Zhang LJ, Zhou YC, Pedrycz W (2018) Large-scale multimodality attribute reduction with multi-kernel fuzzy rough sets. IEEE Trans Fuzzy Syst 26(1):226–238
Zhao H, Hu QH, Zhu PF, Wang Y et al (2019) A recursive regularization based feature selection framework for hierarchical classification. IEEE Trans Knowl Data Eng. https://doi.org/10.1109/TKDE.2019.2960251
Jensen R, Shen Q (2004) Semantics-preserving dimensionality reduction: rough and fuzzy-rough-based approaches. IEEE Trans Knowl Data Eng 16(12):1457–1471
Zhan JM, Jiang HB, Yao YY (2020) Three-way multi-attribute decision-making based on outranking relations. IEEE Trans Fuzzy Syst. https://doi.org/10.1109/TFUZZ.2020.3007423
Geng X, Liu TY, Qin T, Li H (2007) Feature selection for ranking. In: Proceedings of the 30th annual international ACM SIGIR conference on research and development in information retrieval (SIGIR), pp 407–414
Rodriguez-Lujan I, Huerta R, Elkan C, Cruz CS (2010) Quadratic programming feature selection. J Mach Learn Res 11:1491–1516
Kohavi R, John GH (1997) Wrappers for feature subset selection. Artif Intell 97(1C2):273–324
Roffo G, Melzi S, Castellani U, Vinciarelli A, Cristani M (2021) Infinite feature selection: a graph-based feature filtering approach. IEEE Trans Pattern Anal Mach Intell 43(12):4396–4410
Pawlak Z (1998) Rough set theory and its applications in data analysis. Cybern Syst 29:661–688
Liang JY, Chin KS, Dang CY, Yam Richid CM (2002) A new method for measuring uncertainty and fuzziness in rough set theory. Int J Gen Syst 31(4):331–342
Dai JH, Hu QH, Zhang JH, Hu H et al (2017) Attribute selection for partially labeled categorical data by rough set approach. IEEE Trans Cybern 47(9):2460–2471
Zhang XY, Yao H, Lv ZY, Miao DQ (2021) Class-specific information measures and attribute reducts for hierarchy and systematicness. Inf Sci 563:196–225
Kira K, Rendell LA (1992) The feature selection problem: traditional methods and a new algorithm. Proc AAAI 92:129–134
Kryszkiewicz M, Lasek P (2008) FUN: fast discovery of minimal sets of attributes functionally determining a decision attribute. Trans Rough Sets 9:76–95
Kwak N, Choi CH (2002) Input feature selection by mutual information based on Parzen window. IEEE Trans Pattern Anal Mach Intell 24(12):1667–1671
Liang JY, Bai L, Cao FY (2010) K-modes clustering algorithm based on a new distance measure. J Comput Res Dev 47(10):1749–1755
Wang C, Dong XJ, Zhou F, Cao LB, Chi CH (2015) Coupled attribute similarity learning on categorical data. IEEE Trans Neural Net Learn Syst 26(4):781–797
Jian SL, Cao LB, Lu K, Gao H (2018) Unsupervised coupled metric similarity for non-IID categorical data. IEEE Trans Knowl Data Eng 1
Pawlak Z, Skowron A (2007) Rough sets and Boolean reasoning. Inf Sci 177(1):41–73
Xia SY, Wang C, Wang GY, Gao XB et al (2022) An unified model of Pawlak rough set and neighborhood rough set. arXiv e-prints
Xia SY, Zhang H, Li WH, Wang GY et al (2022) GBNRS: a novel rough set algorithm for fast adaptive attribute reduction in classification. IEEE Trans Knowl Data Eng 34(3):1231–1242
Li TR, Ruan D, Geert W, Song J, Xu Y (2007) A rough sets based characteristic relation approach for dynamic attribute generalization in data mining. Knowl Based Syst 20(5):485–494
Diday E, Bock HH (2000) Analysis of symbolic data: exploratory methods for extracting statistical information from complex data. Springer, Berlin
Kononenko I, Robnik-Sikonja M, Pompe U (1996) Relieff for estimation and discretization of attributes in classification, regression, and ILP problems. In: Artificial intelligence: methodology system application: proceedings of AIMSA 96(31–40)
Mitra P, Murthy CA, Pal SK (2002) Unsupervised feature selection using feature similarity. IEEE Trans Pattern Anal Mach Intell 24(3):301–312
Benabdeslem K, Hindawi M (2004) Efficient semi-supervised feature selection: constraint, relevance and redundancy. IEEE Trans Knowl Data Eng 26:1131–1143
Acknowledgements
This work was supported by National Natural Science Fund of China (No. 62276158, 72171137) and Research Project Supported by Shanxi Scholarship Council of China (No. 2021-007).
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Wang, F., Liang, J. & Song, P. Coupling learning for feature selection in categorical data. Int. J. Mach. Learn. & Cyber. 14, 2455–2465 (2023). https://doi.org/10.1007/s13042-023-01775-z
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s13042-023-01775-z