Coupling learning for feature selection in categorical data | International Journal of Machine Learning and Cybernetics Skip to main content
Log in

Coupling learning for feature selection in categorical data

  • Original Article
  • Published:
International Journal of Machine Learning and Cybernetics Aims and scope Submit manuscript

Abstract

Feature selection, which is a commonly used data prepossessing technique, focuses on improving model performance and efficiency by removing redundant or irrelevant features. However, an implicit assumption made by traditional feature selection approaches is that data are independent and identically distributed (IID). To further obtain more complex and significant information, an effective feature selection construction should consider the couplings (non-IIDness) contained within feature values and relevance between features. Hence, referring to rough set theory, this paper first introduces a new coupled similarity measure to discover the value-to-feature-to-class coupling information, which can be used to calculate object neighbor and update feature weights. Second, using mutual information, a new coupled relevance measure is defined to capture the feature-to-feature coupling relationships. On this basis, an effective feature-selection algorithm based on coupling learning is developed for categorical data. To demonstrate the proposed algorithm, four common classifiers and 12 UCI data sets are employed in the experiments. The experimental results confirm the feasibility of the new algorithm and its effectiveness.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
¥17,985 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price includes VAT (Japan)

Instant access to the full article PDF.

Fig. 1

Similar content being viewed by others

Explore related subjects

Discover the latest articles, news and stories from top researchers in related subjects.

References

  1. Cao LB, Ou Y, Yu PS (2012) Coupled behavior analysis with applications. IEEE Trans Knowl Data Eng 24(8):1378–1392

    Article  Google Scholar 

  2. Cao LB (2014) Non-IIDness learning in behavioral and social data. Comput J 57(9):1358–1370

    Article  Google Scholar 

  3. Zhu CZ, Cao LB, Yin JP (2020) Unsupervised heterogeneous coupling learning for categorical representation. IEEE Trans Pattern Anal Mach Intell. https://doi.org/10.1109/TPAMI.2020.3010953

    Article  Google Scholar 

  4. Dash M, Liu H (2003) Consistency-based search in feature selection. Artif Intell 151:155–176

    Article  MathSciNet  MATH  Google Scholar 

  5. Guyon I, Elisseeff A (2003) An introduction to variable feature selection. Mach Learn Res 3:1157–1182

    MATH  Google Scholar 

  6. Hu QH, Zhang LJ, Zhou YC, Pedrycz W (2018) Large-scale multimodality attribute reduction with multi-kernel fuzzy rough sets. IEEE Trans Fuzzy Syst 26(1):226–238

    Article  Google Scholar 

  7. Zhao H, Hu QH, Zhu PF, Wang Y et al (2019) A recursive regularization based feature selection framework for hierarchical classification. IEEE Trans Knowl Data Eng. https://doi.org/10.1109/TKDE.2019.2960251

    Article  Google Scholar 

  8. Jensen R, Shen Q (2004) Semantics-preserving dimensionality reduction: rough and fuzzy-rough-based approaches. IEEE Trans Knowl Data Eng 16(12):1457–1471

    Article  Google Scholar 

  9. Zhan JM, Jiang HB, Yao YY (2020) Three-way multi-attribute decision-making based on outranking relations. IEEE Trans Fuzzy Syst. https://doi.org/10.1109/TFUZZ.2020.3007423

    Article  Google Scholar 

  10. Geng X, Liu TY, Qin T, Li H (2007) Feature selection for ranking. In: Proceedings of the 30th annual international ACM SIGIR conference on research and development in information retrieval (SIGIR), pp 407–414

  11. Rodriguez-Lujan I, Huerta R, Elkan C, Cruz CS (2010) Quadratic programming feature selection. J Mach Learn Res 11:1491–1516

    MathSciNet  MATH  Google Scholar 

  12. Kohavi R, John GH (1997) Wrappers for feature subset selection. Artif Intell 97(1C2):273–324

    Article  MATH  Google Scholar 

  13. Roffo G, Melzi S, Castellani U, Vinciarelli A, Cristani M (2021) Infinite feature selection: a graph-based feature filtering approach. IEEE Trans Pattern Anal Mach Intell 43(12):4396–4410

    Article  Google Scholar 

  14. Pawlak Z (1998) Rough set theory and its applications in data analysis. Cybern Syst 29:661–688

    Article  MATH  Google Scholar 

  15. Liang JY, Chin KS, Dang CY, Yam Richid CM (2002) A new method for measuring uncertainty and fuzziness in rough set theory. Int J Gen Syst 31(4):331–342

    Article  MathSciNet  MATH  Google Scholar 

  16. Dai JH, Hu QH, Zhang JH, Hu H et al (2017) Attribute selection for partially labeled categorical data by rough set approach. IEEE Trans Cybern 47(9):2460–2471

    Article  Google Scholar 

  17. Zhang XY, Yao H, Lv ZY, Miao DQ (2021) Class-specific information measures and attribute reducts for hierarchy and systematicness. Inf Sci 563:196–225

    Article  MathSciNet  Google Scholar 

  18. Kira K, Rendell LA (1992) The feature selection problem: traditional methods and a new algorithm. Proc AAAI 92:129–134

    Google Scholar 

  19. Kryszkiewicz M, Lasek P (2008) FUN: fast discovery of minimal sets of attributes functionally determining a decision attribute. Trans Rough Sets 9:76–95

    Google Scholar 

  20. Kwak N, Choi CH (2002) Input feature selection by mutual information based on Parzen window. IEEE Trans Pattern Anal Mach Intell 24(12):1667–1671

    Article  Google Scholar 

  21. Liang JY, Bai L, Cao FY (2010) K-modes clustering algorithm based on a new distance measure. J Comput Res Dev 47(10):1749–1755

    Google Scholar 

  22. Wang C, Dong XJ, Zhou F, Cao LB, Chi CH (2015) Coupled attribute similarity learning on categorical data. IEEE Trans Neural Net Learn Syst 26(4):781–797

    Article  MathSciNet  Google Scholar 

  23. Jian SL, Cao LB, Lu K, Gao H (2018) Unsupervised coupled metric similarity for non-IID categorical data. IEEE Trans Knowl Data Eng 1

  24. Pawlak Z, Skowron A (2007) Rough sets and Boolean reasoning. Inf Sci 177(1):41–73

    Article  MathSciNet  MATH  Google Scholar 

  25. Xia SY, Wang C, Wang GY, Gao XB et al (2022) An unified model of Pawlak rough set and neighborhood rough set. arXiv e-prints

  26. Xia SY, Zhang H, Li WH, Wang GY et al (2022) GBNRS: a novel rough set algorithm for fast adaptive attribute reduction in classification. IEEE Trans Knowl Data Eng 34(3):1231–1242

    Article  Google Scholar 

  27. Li TR, Ruan D, Geert W, Song J, Xu Y (2007) A rough sets based characteristic relation approach for dynamic attribute generalization in data mining. Knowl Based Syst 20(5):485–494

    Article  Google Scholar 

  28. Diday E, Bock HH (2000) Analysis of symbolic data: exploratory methods for extracting statistical information from complex data. Springer, Berlin

    MATH  Google Scholar 

  29. Kononenko I, Robnik-Sikonja M, Pompe U (1996) Relieff for estimation and discretization of attributes in classification, regression, and ILP problems. In: Artificial intelligence: methodology system application: proceedings of AIMSA 96(31–40)

  30. Mitra P, Murthy CA, Pal SK (2002) Unsupervised feature selection using feature similarity. IEEE Trans Pattern Anal Mach Intell 24(3):301–312

    Article  Google Scholar 

  31. Benabdeslem K, Hindawi M (2004) Efficient semi-supervised feature selection: constraint, relevance and redundancy. IEEE Trans Knowl Data Eng 26:1131–1143

    Article  Google Scholar 

Download references

Acknowledgements

This work was supported by National Natural Science Fund of China (No. 62276158, 72171137) and Research Project Supported by Shanxi Scholarship Council of China (No. 2021-007).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Jiye Liang.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Wang, F., Liang, J. & Song, P. Coupling learning for feature selection in categorical data. Int. J. Mach. Learn. & Cyber. 14, 2455–2465 (2023). https://doi.org/10.1007/s13042-023-01775-z

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s13042-023-01775-z

Keywords

Navigation