Combination of Boosted Classifiers Using Bounded Weights | SpringerLink
Skip to main content

Combination of Boosted Classifiers Using Bounded Weights

  • Conference paper
Pattern Recognition and Data Mining (ICAPR 2005)

Part of the book series: Lecture Notes in Computer Science ((LNIP,volume 3686))

Included in the following conference series:

  • 1885 Accesses

Abstract

A recently developed neural network model that is based on bounded weights is used for the estimation of an optimal set of weights for ensemble members provided by the AdaBoost algorithm. Bounded neural network model is firstly modified for this purpose where ensemble members are used to replace the kernel functions. The optimal set of classifier weights are then obtained by the minimization of a least squares error function. The proposed weight estimation approach is compared to the AdaBoost algorithm with original weights. It is observed that better accuracies can be obtained by using a subset of the ensemble members.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

Similar content being viewed by others

References

  1. Schapire, R.E.: A brief introduction to boosting. In: Proceedings of the Sixteenth International Joint Conference on Artificial Intelligence (1999)

    Google Scholar 

  2. Schapire, R.E.: The boosting approach to machine learning: An overview. In: MSRI Workshop on Nonlinear Estimation and Classification, Berkeley, CA (2002)

    Google Scholar 

  3. Liao, Y., Fang, S.C., Nuttle, H.L.W.: A neural network model with bounded-weights for pattern classification. Computers and Operations Research 31, 1411–1426 (2004)

    Article  MATH  Google Scholar 

  4. Burges, C.J.C.: A tutorial on support vector machines for pattern recognition. Data Min. Knowl. Discov. 2(2), 121–167 (1998)

    Article  Google Scholar 

  5. Scholköpf, B.: Statistical learning and kernel methods. Technical Report, MSR-TR-2000-23, Microsoft Research (2000)

    Google Scholar 

  6. Luenberger, D.G.: Introduction to linear and nonlinear programming. Addison-Wesley Pub. Co., Reading (1973)

    Google Scholar 

  7. Hashem, S.: Optimal linear combinations of neural networks. Neural Networks 10(4), 599–614 (1997)

    Article  Google Scholar 

  8. Duda, R.O., Hart, P.E., Stork, D.G.: Pattern Classification. John Wiley and Sons, Chichester (2000)

    Google Scholar 

  9. Duin, R.P.W.: PRTOOLS (version 4.0). A Matlab toolbox for pattern recognition. In: Pattern Recognition Group. Delft University, Netherlands (2004)

    Google Scholar 

  10. Skurichina, M., Duin, R.P.W.: Bagging, boosting and the random subspace method for linear classifiers. Pattern Analysis and Applications 5, 121–135 (2002)

    Article  MATH  MathSciNet  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2005 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Altınçay, H., Tüzel, A. (2005). Combination of Boosted Classifiers Using Bounded Weights. In: Singh, S., Singh, M., Apte, C., Perner, P. (eds) Pattern Recognition and Data Mining. ICAPR 2005. Lecture Notes in Computer Science, vol 3686. Springer, Berlin, Heidelberg. https://doi.org/10.1007/11551188_16

Download citation

  • DOI: https://doi.org/10.1007/11551188_16

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-28757-5

  • Online ISBN: 978-3-540-28758-2

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics