Solving Partially Observable Problems by Evolution and Learning of Finite State Machines | SpringerLink
Skip to main content

Solving Partially Observable Problems by Evolution and Learning of Finite State Machines

  • Conference paper
  • First Online:
Evolvable Systems: From Biology to Hardware (ICES 2001)

Part of the book series: Lecture Notes in Computer Science ((LNCS,volume 2210))

Included in the following conference series:

Abstract

Finite state machines (FSM) have been successfully used to implement the control of an agent to solve particular sequential tasks. Nevertheless, finite state machines must be hand-coded by the engineer, which might be very difficult for complex tasks. Researchers have used evolutionary techniques to evolve finite state machines and find automatic solutions to sequential tasks. Their approach consists on encoding the state-transition table defining a finite state machine in the genome. However, the search space of such approach tends to be innecesarily huge. In this article, we propose an alternative approach for the automatic design of finite state machines using artificial evolution and learning techniques: the SOS-algorithm. We have obtained very impresive results on experimental work solving partially observable problems.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
¥17,985 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
JPY 3498
Price includes VAT (Japan)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
JPY 5719
Price includes VAT (Japan)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
JPY 7149
Price includes VAT (Japan)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

Similar content being viewed by others

References

  1. P.J. Angeline, G.M. Saunders, and J. Pollack. A evolutionary algorithm that constructs recurrent neural networks. IEEE Transactions on Neural Networks, 5(1):54–65, January 1994.

    Article  Google Scholar 

  2. A.R. Cassandra. Exact and Approximate Algorithms for Partially Observable MarkovDe cision Processes. PhD thesis, Brown University, 1998.

    Google Scholar 

  3. Thomas G. Dietterich. The MAXQ method for hierarchical reinforcement learning. In Proc. 15th International Conf. on Machine Learning, pages 118–126. Morgan Kaufmann, San Francisco, CA, 1998.

    Google Scholar 

  4. L.J. Fogel and A.J. Owens. Artificial Intelligence through Simulated Evolution. John Wiley and sons, New York, 1966.

    MATH  Google Scholar 

  5. M. Hauskrecht. Planning and Control in Stochastic Domains with Imperfect Information. PhD thesis, MIT, Cambridge, MA, 1997.

    Google Scholar 

  6. C. Jacob. Illustrating Evolutionary Computing with Mathematica. Morgan Kaufmann, San Francisco, 2001.

    Google Scholar 

  7. J.R. Koza. Genetic Programming: On the Programming of Computers by Means of Natural Selection. The MIT Press, Cambridge, Massachusetts, 1992.

    MATH  Google Scholar 

  8. J.L. Lin and T.M. Mitchell. Reinforcement learning with hidden states. In JA. Meyer, H.L. Roitblat, and S.W. Wilson, editors, From Animals to Animats: Proceedings of the Second International Conference on Simulation of Adaptive Behavior (SAB92), pages 281–290, 1992.

    Google Scholar 

  9. N. Meuleau, L. Peshkin, K.E. Kim, and L.P. Kaebling. Learning Finite-State Controllers for Partially Observable Environments. In Proceedings of the Conference on Uncertainty and Artificial Intelligence UAI’99, Stockholm, Sweden, 1999.

    Google Scholar 

  10. Z. Michalewicz. Genetic Algorithms + Data Structures = Evolution Programs. Springer-Verlag, 1992.

    Google Scholar 

  11. A. Pérez-Uribe. Structure-Adaptable Digital Neural Networks, Ph.D Thesis 2052. PhD thesis, Swiss Federal Institute of Technology-Lausanne, 1999.

    Google Scholar 

  12. D. Precup, R.S. Sutton, and S. Singh. Theoretical Results on Reinforcement Learning with Temporally Abstract Behaviors. In Proceedings of the 10th European Conference on Machine Learning ECML’98, pages 382–393, Chemnitz, Germany, 1998.

    Google Scholar 

  13. M.B. Ring. Continual learning in reinforcement environments. PhD thesis, The University of Texas at Austin, August 1994.

    Google Scholar 

  14. E. Sanchez. An Introduction to Digital Systems. In D. Mange and M. Tomassini, editors, Bio-Inspired Computing Machines: Toward Novel Computational Machines, pages 13–48. Presses Polytechniques et Universitaires Romandes, Lausanne, Switzerland, 1998.

    Google Scholar 

  15. Satinder P. Singh, Tommi Jaakkola, and Michael I. Jordan. Reinforcement learning with soft state aggregation. In G. Tesauro, D. Touretzky, and T. Leen, editors, Advances in Neural Information Processing Systems, volume 7, pages 361–368. The MIT Press, 1995.

    Google Scholar 

  16. R. Sun and T. Peterson. Multi-agent reinforcement learning with bidding for segmenting action sequences. In J-A. Meyer, A. Bethoz, D. Floreano, D. Roitblat, and S. Wilson, editors, From Animals to Animats: Proceedings of the Sixth International Conference on Simulation of Adaptive Behavior (SAB2000), pages 325–332, 2000.

    Google Scholar 

  17. R.S. Sutton. Learning to predict by the methods of Temporal Differences. Machine Learning, 3:9–44, 1988.

    Google Scholar 

  18. R.S. Sutton and A.G. Barto. Reinforcement Learning: An Introduction. The MIT Press, 1998.

    Google Scholar 

  19. S.D. Whitehead and D.H. Ballard. Learning to Perceive and Act by Trial. Machine Learning, 7(1):45–83, 1991.

    Google Scholar 

  20. M. Wiering and J. Schmidhuber. HQ-Learning. Adaptive Behavior, 6(2), 1997.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2001 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Sanchez, E., Pérez-Uribe, A., Mesot, B. (2001). Solving Partially Observable Problems by Evolution and Learning of Finite State Machines. In: Liu, Y., Tanaka, K., Iwata, M., Higuchi, T., Yasunaga, M. (eds) Evolvable Systems: From Biology to Hardware. ICES 2001. Lecture Notes in Computer Science, vol 2210. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-45443-8_24

Download citation

  • DOI: https://doi.org/10.1007/3-540-45443-8_24

  • Published:

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-42671-4

  • Online ISBN: 978-3-540-45443-4

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics