Skip to main content

Faster Convergence to Cooperative Policy by Autonomous Detection of Interference States in Multiagent Reinforcement Learning

  • Conference paper
  • First Online:
PRICAI 2016: Trends in Artificial Intelligence (PRICAI 2016)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 9810))

Included in the following conference series:

Abstract

In this paper, we propose a method for ameliorating the state-space explosion that can occur in the context of multiagent reinforcement learning. In our method, an agent considers other agents’ states only when they interfere with each other in attaining their goals. Our idea is that the initial state-space of each agent does not include information about other spaces. Agents then automatically expand their state-space if they detect interference states. We adopt the information theory measure of entropy to detect the interference states for which agents should consider the state information of other agents. We demonstrate the advantage of our method with respect to the efficiency of global convergence.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Busoniu, L., De Schutter, B., Babuška, R.: A comprehensive survey of multiagent reinforcement learning. IEEE Trans. Syst. Man Cybern. Part C Appl. Rev. 38(2), 156–172 (2008)

    Article  Google Scholar 

  2. Watkins, C., Dayan, P.: Q-learning. Mach. Learn. 8(3), 279–292 (1992)

    MATH  Google Scholar 

  3. Littman, M.: Markov games as a framework for multi-agent reinforcement learning. In: Proceedings of the Eleventh International Conference on Machine Learning, pp. 242–250 (1994)

    Google Scholar 

  4. Littman, M.: Value-function reinforcement learning in Markov games. Cogn. Syst. Res. 2(1), 55–66 (2001)

    Article  Google Scholar 

  5. Wang, X., Sandholm, T.: Reinforcement learning to play an optimal Nash equilibrium in team Markov games. In: Advances in Neural Information Processing Systems vol. 15, pp. 1571–1578 (2002)

    Google Scholar 

  6. Hu, J., Wellman, M.: Nash Q-learning for general-sum stochastic games. J. Mach. Learn. Res. 4, 1039–1069 (2003)

    MathSciNet  MATH  Google Scholar 

  7. Greenwald, A., Zinkevich, M., Kaelbling, P.: Correlated Q-learning. In: Proceedings of the Twentieth International Conference on Machine Learning, pp. 242–249 (2003)

    Google Scholar 

  8. Busoniu, L., De Schutter, B., Babuška, R.: Multiagent reinforcement learning with adaptive state focus. In: Proceedings of the Seventeenth Belgian-Dutch Conference on Artificial Intelligence, pp. 35–42 (2005)

    Google Scholar 

  9. Kok, J.R., Vlassis, N.: Sparse cooperative Q-learning. In: Proceedings of the Twenty-First International Conference on Machine Learning, pp. 61–68 (2004)

    Google Scholar 

  10. Kok, J.R., Hoen, P., Bakker, B., Vlassis, N.: Utile coordination: learning interdependencies among cooperative agents. In: Proceedings of the IEEE Symposium on Computational Intelligence and Games (CIG), pp. 29–36 (2005)

    Google Scholar 

  11. Spaan, M.T.J., Melo, F.S.: Interaction-driven Markov games for decentralized multiagent planning under uncertainty. In: Proceedings of the 7th International Conference on Autonomous Agents and Multiagent Systems, pp. 525–532 (2008)

    Google Scholar 

  12. Melo, F.S., Veloso, M.: Learning of coordination: exploiting sparse interactions in multiagent systems. In: Proceedings of the 8th International Conference on Autonomous Agents and Multiagent Systems, pp. 773–780 (2009)

    Google Scholar 

  13. De Hauwere, Y., Vrancx, P., Nowé, A.: Learning what to observe in multi-agent systems. In: Proceedings of the Twentieth Belgian-Dutch Conference on Artificial Intelligence, pp. 83–90 (2009)

    Google Scholar 

  14. De Hauwere, Y., Vrancx, P., Nowé, A.: Learning multi-agent state space representations. In: Proceedings of the 9th International Conference on Autonomous Agents and Multiagent Systems, pp. 715–722 (2010)

    Google Scholar 

  15. De Hauwere, Y., Vrancx, P., Nowé, A.: Adaptive state representations for multiagent reinforcement learning. In: Proceedings of the 3rd International Conference on Agents and Artificial Intelligence, pp. 181–189 (2011)

    Google Scholar 

  16. Arai, S., Ishigaki, Y.: Information theoretic approach for measuring interaction in multiagent domain. J. Adv. Comput. Intell. Intell. Inform. 13(6), 649–657 (2009)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Sachiyo Arai .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2016 Springer International Publishing Switzerland

About this paper

Cite this paper

Arai, S., Xu, H. (2016). Faster Convergence to Cooperative Policy by Autonomous Detection of Interference States in Multiagent Reinforcement Learning. In: Booth, R., Zhang, ML. (eds) PRICAI 2016: Trends in Artificial Intelligence. PRICAI 2016. Lecture Notes in Computer Science(), vol 9810. Springer, Cham. https://doi.org/10.1007/978-3-319-42911-3_2

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-42911-3_2

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-42910-6

  • Online ISBN: 978-3-319-42911-3

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics