Abstract
Regularization theory presents a sound framework to solving supervised learning problems. However, there is a gap between the theoretical results and practical suitability of regularization networks (RN). Radial basis function networks (RBF) can be seen as a special case of regularization networks with a selection of learning algorithms. We study a relationship between RN and RBF, and experimentally evaluate their approximation and generalization ability with respect to number of hidden units.
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
References
Haykin, S.: Neural Networks: A Comprehensive Foundation, 2nd edn. Tom Robins (1999)
Xu, L., Krzyżak, A., Yuille, A.: On Radial Basis Function Nets and Kernel Regression: Statistical Consistency, Convergence Rates, and Receptive Field Size. Neural Netw. 7(4), 609–628 (1994)
Corradi, V., White, H.: Regularized Neural Networks: Some Convergence Rate Results. Neural Computation 7, 1225–1244 (1995)
Kuková, V., Sanguineti, M.: Learning with Generalization Capability by Kernal Methods of Bounded Complexity. J. Complex 21(3), 350–367 (2005)
Tikhonov, A., Arsenin, V.: Solutions of Ill-posed Problems. W.H. Winston, Washington (1977)
Poggio, T., Girosi, F.: A Theory of Networks for Approximation and Learning. Technical report, Cambridge, MA, USA (1989)
Poggio, T., Smale, S.: The Mathematics of Learning: Dealing with Data. Notices of the AMS 50, 536–544 (2003)
Powel, M.: Radial Basis Functions for Multivariable Interpolation: A review. In: IMA Conference on Algorithms for the Approximation of Functions and Data, RMCS, Shrivenham, England, pp. 143–167 (1985)
Neruda, R., Kudová, P.: Learning Methods for Radial Basis Functions Networks. Future Generation Computer Systems 21, 1131–1142 (2005)
Prechelt, L.: PROBEN1 – A Set of Benchmarks and Benchmarking Rules for Neural Network Training Algorithms. Technical Report 21/94, Universitaet Karlsruhe (1994)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2008 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Vidnerová, P., Neruda, R. (2008). Testing Error Estimates for Regularization and Radial Function Networks. In: Sun, F., Zhang, J., Tan, Y., Cao, J., Yu, W. (eds) Advances in Neural Networks - ISNN 2008. ISNN 2008. Lecture Notes in Computer Science, vol 5263. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-87732-5_61
Download citation
DOI: https://doi.org/10.1007/978-3-540-87732-5_61
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-87731-8
Online ISBN: 978-3-540-87732-5
eBook Packages: Computer ScienceComputer Science (R0)