Abstract
“The curse of dimensionality” is pertinent to many learning algorithms, and it denotes the drastic raise of computational complexity and the classification error in high dimensions. In this paper, different feature extraction techniques as means of (1) dimensionality reduction, and (2) constructive induction are analyzed with respect to the performance of a classifier. Three commonly used classifiers are taken for the analysis: kNN, Naïve Bayes and C4.5 decision tree. One of the main goals of this paper is to show the importance of the use of class information in feature extraction for classification and (in)appropriateness of random projection or conventional PCA to feature extraction for classification for some data sets. Two eigenvector-based approaches that take into account the class information are analyzed. The first approach is parametric and optimizes the ratio of between-class variance to the within-class variance of the transformed data. The second approach is a nonparametric modification of the first one based on the local calculation of the between-class covariance matrix. In experiments on benchmark data sets these two approaches are compared with each other, with conventional PCA, with random projection and with plain classification without feature extraction for each classifier.
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
References
Achlioptas, D.: Database-friendly random projections. In: Proceedings of the twentieth ACM SIGMOD-SIGACT-SIGART symposium on Principles of database systems, Santa Barbara, California. ACM Press, New York (2001)
Aivazyan, S.: Applied statistics: classification and dimension reduction. In: Finance and Statistics, Moscow (1989)
Aladjem, M.: Multiclass discriminant mappings. Signal Processing 35, 1–18 (1994)
Aladjem, M.: Parametric and nonparametric linear mappings of multidimensional data. Pattern Recognition 24(6), 543–553 (1991)
Bellman, R.: Adaptive Control Processes: A Guided Tour. Princeton University Press, Princeton (1961)
Bingham, E., Mannila, H.: Random projection in dimensionality reduction: applications to image and text data. In: Proceedings of the seventh ACM SIGKDD international conference on Knowledge discovery and data mining. ACM Press, San Francisco (2001)
Blake, C.L., Merz, C.J.: UCI Repository of Machine Learning Databases. Dept. of Information and Computer Science. University of California, Irvine CA (1998)
Fayyad, U.M.: Data Mining and Knowledge Discovery: Making Sense Out of Data. IEEE Expert 11(5), 20–25 (1996)
Fradkin, D., Madigan, D.: Experiments with random projections for machine learning. In: Proceedings of the ninth ACM SIGKDD international conference on Knowledge discovery and data mining. ACM Press, Washington (2003)
Fukunaga, K.: Introduction to statistical pattern recognition. Academic Press, London (1990)
Jimenez, L., Landgrebe, D.: High Dimensional Feature Reduction Via Projection Pursuit. PhD Thesis and School of Electrical & Computer Engineering Technical Report TR-ECE 96-5 (1995)
Jolliffe, I.T.: Principal Component Analysis. Springer, New York (1986)
Kiang, M.: A comparative assessment of classification methods. Decision Support Systems 35, 441–454 (2003)
Kohavi, R., Sommerfield, D., Dougherty, J.: Data mining using MLC++: a machine learning library in C++. Tools with Artificial Intelligence, pp. 234–245. IEEE CS Press, Los Alamitos (1996)
Liu, H.: Feature Extraction, Construction and Selection: A Data Mining Perspective. Kluwer Academic Publishers, Dordrecht (1998)
Michalski, R.S.: Seeking Knowledge in the Deluge of Facts. Fundamenta Informaticae 30, 283–297 (1997)
Oza, N.C., Tumer, K.: Dimensionality Reduction Through Classifier Ensembles. Technical Report NASA-ARC-IC-1999-124, Computational Sciences Division, NASA Ames Research Center, Moffett Field, CA (1999)
Quinlan, J.R.: C4.5 Programs for Machine Learning. Morgan Kaufmann, San Mateo (1993)
Tsymbal, A., Pechenizkiy, M., Puuronen, S., Patterson, D.W.: Dynamic integration of classifiers in the space of principal components. In: Kalinichenko, L.A., Manthey, R., Thalheim, B., Wloka, U. (eds.) ADBIS 2003. LNCS, vol. 2798, pp. 278–292. Springer, Heidelberg (2003)
Tsymbal, A., Puuronen, S., Pechenizkiy, M., Baumgarten, M., Patterson, D.: Eigenvector-based feature extraction for classification. In: Proc. 15th Int. FLAIRS Conference on Artificial Intelligence, Pensacola, FL, USA, pp. 354–358. AAAI Press, Menlo Park (2002)
Witten, I., Frank, E.: Data Mining: Practical machine learning tools with Java implementations. Morgan Kaufmann, San Francisco (2000)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2005 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Pechenizkiy, M. (2005). The Impact of Feature Extraction on the Performance of a Classifier: kNN, Naïve Bayes and C4.5. In: Kégl, B., Lapalme, G. (eds) Advances in Artificial Intelligence. Canadian AI 2005. Lecture Notes in Computer Science(), vol 3501. Springer, Berlin, Heidelberg. https://doi.org/10.1007/11424918_28
Download citation
DOI: https://doi.org/10.1007/11424918_28
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-25864-3
Online ISBN: 978-3-540-31952-8
eBook Packages: Computer ScienceComputer Science (R0)