Skip to main content

The Effects of Locality and Asymmetry in Large-Scale Multiagent MDPs

  • Chapter
Coordination of Large-Scale Multiagent Systems

Summary

As multiagent systems scale up, the complexity of interactions between agents (cooperative coordination in teams, or strategic reasoning in the case of self-interested agents) often increases exponentially. In particular, in multiagent MDPs, it is generally necessary to consider the joint state space of all agents, making the size of the problem and the solution exponential in the number of agents. However, often interactions between the agents are only local, which suggests a more compact problem representation. We consider a subclass of multiagent MDPs with local interactions where dependencies between agents are asymmetric, meaning that agents can affect others in a unidirectional manner. This asymmetry, which often occurs in large-scale domains with authority-driven relationships between agents, allows us to make better use of the locality of agents’ interactions. We discuss a graphical model that exploits this form of problem structure and use it to analyze the effects of locality and asymmetry on the complexity and structure of optimal policies. For problems where the solutions retain some of the compactness of problem representation, we present computationally-efficient algorithms for constructing optimal multiagent policies.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 129.00
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 169.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 169.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

Similar content being viewed by others

References

  1. R. Bellman. Adaptive Control Processes: A Guided Tour. Princeton University Press, 1961.

    Google Scholar 

  2. Richard Bellman. Dynamic Programming. Princeton University Press, 1957.

    Google Scholar 

  3. Craig Boutilier. Sequential optimality and coordination in multiagent systems. In Proceedings of the 1999 International Joint Conference on Artificial Intelligence, pages 478–485, 1999.

    Google Scholar 

  4. Craig Boutilier, Thomas Dean, and Steve Hanks. Decision-theoretic planning: Structural assumptions and computational leverage. Journal of Artificial Intelligence Research, 11:1–94, 1999.

    MathSciNet  Google Scholar 

  5. Craig Boutilier, Richard Dearden, and Moises Goldszmidt. Stochastic dynamic programming with factored representations. Artificial Intelligence, 121(l–2):49–107, 2000.

    Article  MathSciNet  Google Scholar 

  6. D. P. de Farias and B. Van Roy. The linear programming approach to approximate dynamic programming. Operations Research, 51(6), 2003.

    Google Scholar 

  7. Daniela de Farias and Benjamin Van Roy. On constraint sampling in the linear programming approach to approximate dynamic programming. Mathematics of Operations Research, 29(3):462–478, 2004.

    Article  MathSciNet  Google Scholar 

  8. Thomas Dean and Keiji Kanazawa. A model for reasoning about persistence and causation. Computational Intelligence, 5(3): 142–150, 1989.

    Google Scholar 

  9. Dmitri A. Dolgov and Edmund H. Durfee. Optimal resource allocation and policy formulation in loosely-coupled Markov decision processes. In Proceedings of the Fourteenth International Conference on Automated Planning and Scheduling (ICAPS 04), pages 315–324, June 2004.

    Google Scholar 

  10. C. Guestrin, D. Roller, R. Parr, and S Venkataraman. Efficient solution algorithms for factored MDPs. Journal of Artificial Intelligence Research, 19:399–468, 2003.

    MathSciNet  Google Scholar 

  11. M.I. Jordan. Graphical models. Statistical Science (Special Issue on Bayesian Statistics), 19:140–155, 2004.

    MATH  Google Scholar 

  12. Michael Kearns, Michael L. Littman, and Satinder Singh. Graphical models for game theory. In Proceedings of the 17th Conference in Uncertainty in Artificial Intelligence (UAI01), pages 253–260, 2001.

    Google Scholar 

  13. Daphne Roller and Brian Milch. Multi-agent influence diagrams for representing and solving games. In Proceedings of Seventeenth International Joint Conference on Artificial Intelligence, pages 1027–1036, 2001.

    Google Scholar 

  14. Daphne Roller and Ronald Parr. Computing factored value functions for policies in structured MDPs. In Proceedings of the Sixteenth International Conference on Artificial Intelligence IJCAI-99, pages 1332–1339, 1999.

    Google Scholar 

  15. Michael L. Littman. Markov games as a framework for multi-agent reinforcement learning. In Proceedings of the 11th International Conference on Machine Learning (ML-94), pages 157–163, New Brunswick, NJ, 1994. Morgan Kaufmann.

    Google Scholar 

  16. Guillermo Owen. Game Theory: Second Edition. Academic Press, Orlando, Florida, 1982.

    Google Scholar 

  17. M. L. Puterman. Markov Decision Processes. John Wiley & Sons, New York, 1994.

    Google Scholar 

  18. D. Pynadath and M. Tambe. Multiagent teamwork: Analyzing the optimality and complexity of key theories and models. In In Proceedings of the First Conference on autonomous agents and multiagent systems (AAMAS-2002), 2002.

    Google Scholar 

  19. Satinder Singh and David Cohn. How to dynamically merge Markov decision processes. In Michael I. Jordan, Michael J. Kearns, and Sara A. Solla, editors, Advances in Neural Information Processing Systems, volume 10. The MIT Press, 1998.

    Google Scholar 

  20. Robert St-Aubin, Jesse Hoey, and Craig Boutilier. Apricodd: Approximate policy construction using decision diagrams. In NIPS, pages 1089–1095, 2000.

    Google Scholar 

  21. J. van der Wai. Stochastic dynamic programming. Mathematical Centre Tracts, 139, 1981.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2006 Springer Science+Business Media, Inc.

About this chapter

Cite this chapter

Dolgov, D.A., Durfee, E.H. (2006). The Effects of Locality and Asymmetry in Large-Scale Multiagent MDPs. In: Scerri, P., Vincent, R., Mailler, R. (eds) Coordination of Large-Scale Multiagent Systems. Springer, Boston, MA. https://doi.org/10.1007/0-387-27972-5_1

Download citation

  • DOI: https://doi.org/10.1007/0-387-27972-5_1

  • Publisher Name: Springer, Boston, MA

  • Print ISBN: 978-0-387-26193-5

  • Online ISBN: 978-0-387-27972-5

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics