Summary
As multiagent systems scale up, the complexity of interactions between agents (cooperative coordination in teams, or strategic reasoning in the case of self-interested agents) often increases exponentially. In particular, in multiagent MDPs, it is generally necessary to consider the joint state space of all agents, making the size of the problem and the solution exponential in the number of agents. However, often interactions between the agents are only local, which suggests a more compact problem representation. We consider a subclass of multiagent MDPs with local interactions where dependencies between agents are asymmetric, meaning that agents can affect others in a unidirectional manner. This asymmetry, which often occurs in large-scale domains with authority-driven relationships between agents, allows us to make better use of the locality of agents’ interactions. We discuss a graphical model that exploits this form of problem structure and use it to analyze the effects of locality and asymmetry on the complexity and structure of optimal policies. For problems where the solutions retain some of the compactness of problem representation, we present computationally-efficient algorithms for constructing optimal multiagent policies.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
References
R. Bellman. Adaptive Control Processes: A Guided Tour. Princeton University Press, 1961.
Richard Bellman. Dynamic Programming. Princeton University Press, 1957.
Craig Boutilier. Sequential optimality and coordination in multiagent systems. In Proceedings of the 1999 International Joint Conference on Artificial Intelligence, pages 478–485, 1999.
Craig Boutilier, Thomas Dean, and Steve Hanks. Decision-theoretic planning: Structural assumptions and computational leverage. Journal of Artificial Intelligence Research, 11:1–94, 1999.
Craig Boutilier, Richard Dearden, and Moises Goldszmidt. Stochastic dynamic programming with factored representations. Artificial Intelligence, 121(l–2):49–107, 2000.
D. P. de Farias and B. Van Roy. The linear programming approach to approximate dynamic programming. Operations Research, 51(6), 2003.
Daniela de Farias and Benjamin Van Roy. On constraint sampling in the linear programming approach to approximate dynamic programming. Mathematics of Operations Research, 29(3):462–478, 2004.
Thomas Dean and Keiji Kanazawa. A model for reasoning about persistence and causation. Computational Intelligence, 5(3): 142–150, 1989.
Dmitri A. Dolgov and Edmund H. Durfee. Optimal resource allocation and policy formulation in loosely-coupled Markov decision processes. In Proceedings of the Fourteenth International Conference on Automated Planning and Scheduling (ICAPS 04), pages 315–324, June 2004.
C. Guestrin, D. Roller, R. Parr, and S Venkataraman. Efficient solution algorithms for factored MDPs. Journal of Artificial Intelligence Research, 19:399–468, 2003.
M.I. Jordan. Graphical models. Statistical Science (Special Issue on Bayesian Statistics), 19:140–155, 2004.
Michael Kearns, Michael L. Littman, and Satinder Singh. Graphical models for game theory. In Proceedings of the 17th Conference in Uncertainty in Artificial Intelligence (UAI01), pages 253–260, 2001.
Daphne Roller and Brian Milch. Multi-agent influence diagrams for representing and solving games. In Proceedings of Seventeenth International Joint Conference on Artificial Intelligence, pages 1027–1036, 2001.
Daphne Roller and Ronald Parr. Computing factored value functions for policies in structured MDPs. In Proceedings of the Sixteenth International Conference on Artificial Intelligence IJCAI-99, pages 1332–1339, 1999.
Michael L. Littman. Markov games as a framework for multi-agent reinforcement learning. In Proceedings of the 11th International Conference on Machine Learning (ML-94), pages 157–163, New Brunswick, NJ, 1994. Morgan Kaufmann.
Guillermo Owen. Game Theory: Second Edition. Academic Press, Orlando, Florida, 1982.
M. L. Puterman. Markov Decision Processes. John Wiley & Sons, New York, 1994.
D. Pynadath and M. Tambe. Multiagent teamwork: Analyzing the optimality and complexity of key theories and models. In In Proceedings of the First Conference on autonomous agents and multiagent systems (AAMAS-2002), 2002.
Satinder Singh and David Cohn. How to dynamically merge Markov decision processes. In Michael I. Jordan, Michael J. Kearns, and Sara A. Solla, editors, Advances in Neural Information Processing Systems, volume 10. The MIT Press, 1998.
Robert St-Aubin, Jesse Hoey, and Craig Boutilier. Apricodd: Approximate policy construction using decision diagrams. In NIPS, pages 1089–1095, 2000.
J. van der Wai. Stochastic dynamic programming. Mathematical Centre Tracts, 139, 1981.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2006 Springer Science+Business Media, Inc.
About this chapter
Cite this chapter
Dolgov, D.A., Durfee, E.H. (2006). The Effects of Locality and Asymmetry in Large-Scale Multiagent MDPs. In: Scerri, P., Vincent, R., Mailler, R. (eds) Coordination of Large-Scale Multiagent Systems. Springer, Boston, MA. https://doi.org/10.1007/0-387-27972-5_1
Download citation
DOI: https://doi.org/10.1007/0-387-27972-5_1
Publisher Name: Springer, Boston, MA
Print ISBN: 978-0-387-26193-5
Online ISBN: 978-0-387-27972-5
eBook Packages: Computer ScienceComputer Science (R0)