Learning to Negotiate Optimally in Non-stationary Environments | SpringerLink
Skip to main content

Learning to Negotiate Optimally in Non-stationary Environments

  • Conference paper
Cooperative Information Agents X (CIA 2006)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 4149))

Included in the following conference series:

Abstract

We adopt the Markov chain framework to model bilateral negotiations among agents in dynamic environments and use Bayesian learning to enable them to learn an optimal strategy in incomplete information settings. Specifically, an agent learns the optimal strategy to play against an opponent whose strategy varies with time, assuming no prior information about its negotiation parameters. In so doing, we present a new framework for adaptive negotiation in such non-stationary environments and develop a novel learning algorithm, which is guaranteed to converge, that an agent can use to negotiate optimally over time. We have implemented our algorithm and shown that it converges quickly in a wide range of cases.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
¥17,985 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
JPY 3498
Price includes VAT (Japan)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
JPY 5719
Price includes VAT (Japan)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
JPY 7149
Price includes VAT (Japan)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

Similar content being viewed by others

References

  1. Filar, J., Vrieze, K.: Competitive Markov Decision Processes. Springer, Heidelberg (1996)

    Google Scholar 

  2. Fudenberg, D., Levine, D.K.: The Theory of Learning in Games. MIT Press, Cambridge (1998)

    MATH  Google Scholar 

  3. Fudenberg, D., Tirole, J.: Game Theory. MIT Press, Cambridge (1991)

    Google Scholar 

  4. Hu, J., Wellman, M.P.: Multiagent reinforcement learning: Theoretical framework and an algorithm. In: Proceedings of the 11th International Conference on Machine Learning, pp. 242–250 (1998)

    Google Scholar 

  5. Kalai, E., Lehrer, E.: Rational learning leads to nash equilibrium. Econometrica 61(5), 1019–1045 (1993)

    Article  MATH  MathSciNet  Google Scholar 

  6. Karlin, S., Taylor, H.: First Course in Stochastic Processes. Academic Press, London (1974)

    Google Scholar 

  7. Kraus, S., Subrahmanian, V.S.: Multiagent reasoning with probability, time, and beliefs. International Journal of Intelligent Systems 10(5), 459–499 (1995)

    Article  MATH  Google Scholar 

  8. Kulkarni, V.: Modelling and Analysis of Stochastic Systems. Chapman Hall/CRC (1996)

    Google Scholar 

  9. Littman, M.L.: Markov games as a framework for multi-agent reinforcement learning. In: Proceedings of the 11th International Conference on Machine Learning, pp. 157–163 (1994)

    Google Scholar 

  10. Narayanan, V., Jennings, N.R.: An adaptive bilateral negotiation model for e-commerce settings. In: Proc. 7th Int. IEEE Conf. on E-Commerce Technology, Munich, Germany, pp. 34–39 (2005)

    Google Scholar 

  11. Sutton, R.S., Barto, A.G.: Reinforcement Learning: An Introduction. MIT Press, Cambridge (1998)

    Google Scholar 

  12. Weinberg, M., Rosenschein, J.: Best-response multiagent learning in non-stationary environments. In: The Third International Joint Conference on Autonomous Agents and Mutli-Agent Systems, pp. 506–513 (2004)

    Google Scholar 

  13. Zeng, D., Sycara, K.: Bayesian learning in negotiation. Int. J. Human-Computer Studies 48, 125–141 (1998)

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2006 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Narayanan, V., Jennings, N.R. (2006). Learning to Negotiate Optimally in Non-stationary Environments. In: Klusch, M., Rovatsos, M., Payne, T.R. (eds) Cooperative Information Agents X. CIA 2006. Lecture Notes in Computer Science(), vol 4149. Springer, Berlin, Heidelberg. https://doi.org/10.1007/11839354_21

Download citation

  • DOI: https://doi.org/10.1007/11839354_21

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-38569-1

  • Online ISBN: 978-3-540-38570-7

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics