The Data Dimensionality Reduction and Features Weighting in the Classification Process Using Forest Optimization Algorithm | SpringerLink
Skip to main content

The Data Dimensionality Reduction and Features Weighting in the Classification Process Using Forest Optimization Algorithm

  • Chapter
  • First Online:
Intelligent Information and Database Systems: Recent Developments (ACIIDS 2019)

Abstract

The paper presents the data dimensionality reduction in the classification process, with a special presentation of using the ability of features weighting by determining the level of importance of a given attribute in the data vector. This reduction was implemented using the Forest Optimization Algorithm (FOA) and the use of a classifier allowing to enter the importance of each attribute for a data vector. The paper presents both, a description of the capability of using the FOA algorithm as well as the possibility of introducing modifications which allows to regulate the objective function between the obtained classification result and the number of reduced features. The conducted tests and obtained results were also presented. At the end of paper, a summary and the final conclusions are provided.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
¥17,985 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
JPY 3498
Price includes VAT (Japan)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
JPY 11439
Price includes VAT (Japan)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Hardcover Book
JPY 14299
Price includes VAT (Japan)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

References

  1. Agrawal, R., Imielinski, T., Swami, A.: Database mining: a performance perspective. IEEE Trans. Knowl. Data Eng. 5(6), 914–925 (1993)

    Google Scholar 

  2. Aha, D.W., Kibler, D., Albert, M.K.: Instance-based learning algorithms. Mach. Learn. 6(1), 37–66 (1991)

    Google Scholar 

  3. Aha, D.W.: Feature weighting for lazy learning algorithms. In: Feature Extraction, Construction and Selection, pp. 13–32. Springer (1998)

    Google Scholar 

  4. Arie, B.D.: Comparison of classification accuracy using cohen’s weighted kappa. Expert. Syst. Appl. 34(2), 825–832 (2008)

    Google Scholar 

  5. Bach, M., Werner, A.: Cost-sensitive feature selection for class imbalance problem. In: International Conference on Information Systems Architecture and Technology, pp. 182–194. Springer (2017)

    Google Scholar 

  6. Costa, E., Lorena, A., Carvalho, A., Freitas, A.: A review of performance evaluation measures for hierarchical classifiers. In: Evaluation Methods for Machine Learning II: Papers from the AAAI-2007 Workshop, pp. 1–6 (2007)

    Google Scholar 

  7. Doak, J.: An evaluation of feature selection methods and their application to computer security. Technical Report, Department of Computer Science, University of California, Davis CA (1992)

    Google Scholar 

  8. Gasca, E., Sánchez, J.S., Alonso, R.: Eliminating redundancy and irrelevance using a new mlp-based feature selection method. Pattern Recognit. 39(2), 313–315 (2006)

    MATH  Google Scholar 

  9. Ghaemi, M., Feizi-Derakhshi, M.R.: Forest optimization algorithm. Expert. Syst. Appl. 41(15), 6676–6687 (2014)

    Google Scholar 

  10. Ghaemi, M., Feizi-Derakhshi, M.R.: Feature selection using forest optimization algorithm. Pattern Recognit. 60, 121–129 (2016)

    Google Scholar 

  11. Guyon, I., Elisseeff, A.: An introduction to variable and feature selection. J. Mach. Learn. Res. 3(Mar), 1157–1182 (2003)

    Google Scholar 

  12. Hall, M., Frank, E., Holmes, G., Pfahringer, B., Reutemann, P., Witten, I.H.: The weka data mining software: an update. ACM SIGKDD Explor. Newsl. 11(1), 10–18 (2009)

    Google Scholar 

  13. Kostrzewa, D., Brzeski, R.: The data dimensionality reduction in the classification process through greedy backward feature elimination. In: International Conference on Man–Machine Interactions, pp. 397–407. Springer (2017)

    Google Scholar 

  14. Kostrzewa, D., Brzeski, R., Kubanski, M.: The classification of music by the genre using the knn classifier. In: International Conference: Beyond Databases, Architectures and Structures, pp. 233–242. Springer (2018)

    Google Scholar 

  15. Liu, H., Motoda, H.: Feature Extraction, Construction and Selection: A Data Mining Perspective, vol. 453. Springer (1998)

    Google Scholar 

  16. Liu, H., Motoda, H.: Feature Selection for Knowledge Discovery and Data Mining, vol. 454. Springer (2012)

    Google Scholar 

  17. Özşen, S., Güneş, S.: Attribute weighting via genetic algorithms for attribute weighted artificial immune system (awais) and its application to heart disease and liver disorders problems. Expert. Syst. Appl. 36(1), 386–392 (2009)

    Google Scholar 

  18. Papakostas, G., Koulouriotis, D., Polydoros, A., Tourassis, V.: Evolutionary feature subset selection for pattern recognition applications. In: Evolutionary Algorithms. InTech (2011)

    Google Scholar 

  19. Powers, D.M.: Evaluation: from precision, recall and f-measure to roc, informedness, markedness and correlation (2011)

    Google Scholar 

  20. Rosner, A., Weninger, F., Schuller, B., Michalak, M., Kostek, B.: Influence of low-level features extracted from rhythmic and harmonic sections on music genre classification. In: Man-Machine Interactions, vol. 3, pp. 467–473. Springer (2014)

    Google Scholar 

  21. Tosun, A., Turhan, B., Bener, A.B.: Feature weighting heuristics for analogy-based effort estimation models. Expert. Syst. Appl. 36(7), 10325–10333 (2009)

    Google Scholar 

  22. Tzanetakis, G., Cook, P.: Musical genre classification of audio signals. IEEE Trans. Speech Audio Process. 10(5), 293–302 (2002)

    Google Scholar 

  23. UCI Machine Learning Repository: Dermatology data set. Irvine, CA: University of California, School of Information and Computer Science. http://archive.ics.uci.edu/ml/datasets/dermatology

  24. Wu, X., Kumar, V., Quinlan, J.R., Ghosh, J., Yang, Q., Motoda, H., McLachlan, G.J., Ng, A., Liu, B., Philip, S.Y., et al.: Top 10 algorithms in data mining. Knowl. Inf. Syst. 14(1), 1–37 (2008)

    Google Scholar 

  25. Yang, J., Honavar, V.: Feature subset selection using a genetic algorithm. In: Feature Extraction, Construction and Selection, pp. 117–136. Springer (1998)

    Google Scholar 

  26. Yang, Z.M., He, J.Y., Shao, Y.H.: Feature selection based on linear twin support vector machines. Procedia Comput. Sci. 17, 1039–1046 (2013)

    Google Scholar 

Download references

Acknowledgements

This work was supported by BKM-509/RAU2/2017 (DK) and BK-213/RAu2/2018 (RB) grants from the Institute of Informatics, Silesian University of Technology, Gliwice, Poland.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Daniel Kostrzewa .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2020 Springer Nature Switzerland AG

About this chapter

Check for updates. Verify currency and authenticity via CrossMark

Cite this chapter

Kostrzewa, D., Brzeski, R. (2020). The Data Dimensionality Reduction and Features Weighting in the Classification Process Using Forest Optimization Algorithm. In: Huk, M., Maleszka, M., Szczerbicki, E. (eds) Intelligent Information and Database Systems: Recent Developments. ACIIDS 2019. Studies in Computational Intelligence, vol 830. Springer, Cham. https://doi.org/10.1007/978-3-030-14132-5_8

Download citation

Publish with us

Policies and ethics