Learning of Boolean Functions Using Support Vector Machines | SpringerLink
Skip to main content

Learning of Boolean Functions Using Support Vector Machines

  • Conference paper
  • First Online:
Algorithmic Learning Theory (ALT 2001)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 2225))

Included in the following conference series:

  • 702 Accesses

Abstract

This paper concerns the design of a Support Vector Machine (SVM) appropriate for the learning of Boolean functions. This is motivated by the need of a more sophisticated algorithm for classification in discrete attribute spaces. Classification in discrete attribute spaces is reduced to the problem of learning Boolean functions from examples of its input/output behavior. Since any Boolean function can be written in Disjunctive Normal Form (DNF), it can be represented as a weighted linear sum of all possible conjunctions of Boolean literals. This paper presents a particular kernel function called the DNF kernel which enables SVMs to efficiently learn such linear functions in the high-dimensional space whose coordinates correspond to all possible conjunctions. For a limited form of DNF consisting of positive Boolean literals, the monotone DNF kernel is also presented. SVMs employing these kernel functions can perform the learning in a high-dimensional feature space whose features are derived from given basic attributes. In addition, it is expected that SVMs’ well-founded capacity control alleviates overfitting. In fact, an empirical study on learning of randomly generated Boolean functions shows that the resulting algorithm outperforms C4.5. Furthermore, in comparison with SVMs employing the Gaussian kernel, it is shown that DNF kernel produces accuracy comparable to best adjusted Gaussian kernels.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

Similar content being viewed by others

References

  1. D. Angluin. Queries and concept learning. Machine Learning, 2:319–342, 1988.

    MathSciNet  Google Scholar 

  2. M. Bellare and P. Rogaway. The complexity of approximating a nonlinear program. In Complexity of Numerical Optimization, pages 16–32. World Scientific, 1993.

    Google Scholar 

  3. V. Cherkassky and F. Mulier. Learning from Data. Wiley, 1998.

    Google Scholar 

  4. N. Cristianini and J. Shawe-Taylor. An Introduction to Support Vector Machines. Cambridge Press, 2000.

    Google Scholar 

  5. P. Domingos. On the optimality of the simple Bayesian classifier under zero-one loss. Machine Learning, 29:103-130, 1997.

    Article  Google Scholar 

  6. P. Domingos. The role of Occam’s razor in knowledge discovery. Data Mining and Knowledge Discovery, 3:409–425, 1999.

    Article  Google Scholar 

  7. M. Kearns, Ming Li, L. Pitt, and L.G. Valiant. On the learnability of Boolean formulae. ACM, pages 285–295, 1987.

    Google Scholar 

  8. R. Khardon, D. Roth, and R. Servedio. Efficiency versus convergence of Boolean kernels for on-line learning algorithms. Manuscript, 2001.

    Google Scholar 

  9. H. Liu and H. Motoda. Feature transformation and subset selection. IEEE Intelligent Systems, 13(2):26–28, 1998.

    Article  Google Scholar 

  10. L. Pitt and L.G. Valiant. Computational limitations on learning from examples. Journal of the Association for Computing Machinery, 35(4):965–984, 1988.

    Article  MathSciNet  Google Scholar 

  11. J.C. Platt. Fast training of support vector machines using sequential minimal optimization. In Advances in Kernel Methods-Support Vector Learning, pages 185–208. MIT Press, 1998.

    Google Scholar 

  12. J.R. Quinlan. An empirical comparison of genetic and decision-tree classifiers. In Proceedings of International Conference on Machine Learning, pages 135–141, 1988.

    Google Scholar 

  13. J.R. Quinlan. C4.5: Programs for Machine Learning. Morgan Kaufmann, 1993.

    Google Scholar 

  14. L.G. Valiant. A theory of the learnable. Communications of the ACM, 27(11):1134–1142, 1984.

    Article  Google Scholar 

  15. V. Vapnik. The Nature of Statistical Learning Theory. Springer-Verlag, 1995.

    Google Scholar 

  16. V. Vapnik. Statistical Learning Theory. Wiley, 1998.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2001 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Sadohara, K. (2001). Learning of Boolean Functions Using Support Vector Machines. In: Abe, N., Khardon, R., Zeugmann, T. (eds) Algorithmic Learning Theory. ALT 2001. Lecture Notes in Computer Science(), vol 2225. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-45583-3_10

Download citation

  • DOI: https://doi.org/10.1007/3-540-45583-3_10

  • Published:

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-42875-6

  • Online ISBN: 978-3-540-45583-7

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics