Abstract
Multiple kernel learning (MKL) is a widely used kernel learning method, but how to select kernel is lack of theoretical guidance. The performance of MKL is depend on the users’ experience, which is difficult to choose the proper kernels in practical applications. In this paper, we propose a MKL method based on minimal redundant maximal relevance criterion and kernel alignment. The main feature of this method compared to others in the literature is that the selection of kernels is considered as a feature selection issue in the Hilbert space, and can obtain a set of base kernels with the highest relevance to the target task and the minimal redundancies among themselves. Experimental results on several benchmark classification data sets show that our proposed method can enhance the performance of MKL.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
References
Sonnenburg, S., Rätsch, G., Schäfer, C., Schölkopf, B.: Large scale multiple kernel learning. The Journal of Machine Learning Research 7, 1531–1565 (2006)
Duan, L., Tsang, I.W., Xu, D., Maybank, S.J.: Domain transfer svm for video concept detection. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 1375–1381. IEEE (2009)
Qi, M., Tsang, I.W.: Efficient Multi-template Learning for Structured Prediction. IEEE Transactions on Neural Networks and Learning Systems 24(2), 248–261 (2013)
Vapnik, V.: The nature of statistical learning theory. Springer (1999)
Lanckriet, G.R., Cristianini, N., Bartlett, P., et al.: Learning the kernel matrix with semidefinite programming. The Journal of Machine Learning Research 5, 27–72 (2004)
Bach, F.R., Lanckriet, G.R., Jordan, M.I.: Multiple kernel learning, conic duality, and the smo algorithm. In: Proceedings of the Twenty-First International Conference on Machine Learning, pp. 41–48. ACM (2004)
Rakotomamonjy, A., Bach, F., Canu, S., et al.: Simplemkl. Journal of Machine Learning Research 9, 2491–2521 (2008)
Gönen, M., Alpaydin, E.: Loclized algorithms for multiple kernel learning. Pattern Recognition 46, 795–807 (2013)
Cortes, C., Mohri, M., Rostamizadeh, A.: Algorithms for learning kernels based on centered alignment. The Journal of Machine Learning Research 13, 795–828 (2012)
Guyon, I., Elisseeff, A.: An introduction to variable and feature selection. The Journal of Machine Learning Research 3, 1157–1182 (2003)
Peng, H., Long, F., Ding, C.: Feature selection based on mutual information criteria of max-dependency, max-relevance, and min-redundancy. IEEE Transactions on Pattern Analysis and Machine Intelligence 27(8), 1226–1238 (2005)
Afkanpour, A., Szepesvari, C., Bowling, M.: Alignment based kernel learning with a continuous set of base kernels. arXiv preprint arXiv:1112.4607 (2011)
Cristianini, N., Shawe-Taylor, J., Elisseeff, A., et al.: On kernel-target alignment. In: NIPS, pp. 367–373 (2001)
Bache, K., Lichman, M.: UCI machine learning repository (2013)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2013 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Wu, P., Duan, F., Guo, P. (2013). Multiple Kernel Learning Method Using MRMR Criterion and Kernel Alignment. In: Lee, M., Hirose, A., Hou, ZG., Kil, R.M. (eds) Neural Information Processing. ICONIP 2013. Lecture Notes in Computer Science, vol 8226. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-42054-2_15
Download citation
DOI: https://doi.org/10.1007/978-3-642-42054-2_15
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-42053-5
Online ISBN: 978-3-642-42054-2
eBook Packages: Computer ScienceComputer Science (R0)