A consideration of human immunity-based reinforcement learning with continuous states | Artificial Life and Robotics Skip to main content

Advertisement

Log in

A consideration of human immunity-based reinforcement learning with continuous states

  • Original Article
  • Published:
Artificial Life and Robotics Aims and scope Submit manuscript

Abstract

Many reinforcement learning methods have been studied on the assumption that a state is discretized and the environment size is predetermined. However, an operating environment may have a continuous state and its size may not be known in advance, e.g., in robot navigation and control. When applying these methods to the environment described above, we may need a large amount of time for learning or failing to learn. In this study, we improve our previous human immunity-based reinforcement learning method so that it will work in continuous state space environments. Since our method selects an action based on the distance between the present state and the memorized action, information about the environment (e.g., environment size) is not required in advance. The validity of our method is demonstrated through simulations for the swingup control of an inverted pendulum.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
¥17,985 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price includes VAT (Japan)

Instant access to the full article PDF.

Similar content being viewed by others

Explore related subjects

Discover the latest articles, news and stories from top researchers in related subjects.

References

  1. Ito J, Nakano K, Sakurama K, et al (2008) Adaptive immunity-based reinforcement learning. Artif Life Robotics 13(1):188–193

    Article  Google Scholar 

  2. Watkins CJCH, Dayan P (1992) Technical note: q-learning. Mach Learn 8(3–4):279–292

    MATH  Google Scholar 

  3. Grefenstette JJ (1988) Credit assignment in rule discovery systems based on genetic algorithms. In: Shavlik JW, Dietterich TG (eds) Readings in machine learning. Kaufmann, San Mateo, pp 524–534

  4. Matsui T, Inuzuka N, Seki H (2002) Profit sharing with linear function approximation (in Japanese). 16th Annual Conference of the Japanese Society for Artificial Intelligence, pp 2D3–03

  5. Kimura H, Kobayashi S (2000) An analysis of actor-critic algorithms using eligibility traces: reinforcement learning with imperfect value functions (in Japanese). J Jpn Soc Artif Intell 15(2):267–275

    Google Scholar 

  6. Kakiuchi S, Ikebuchi K, Ota K, et al (eds) (2006) Immunology handbook, vol. 1 (in Japanese), Ohm sha

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Shu Hosokawa.

Additional information

This work was presented in part at the 15th International Symposium on Artificial Life and Robotics, Oita, Japan, February 4–6, 2010

About this article

Cite this article

Hosokawa, S., Nakano, K. & Sakurama, K. A consideration of human immunity-based reinforcement learning with continuous states. Artif Life Robotics 15, 560–564 (2010). https://doi.org/10.1007/s10015-010-0867-7

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10015-010-0867-7

Key words

Navigation