skip to main content
10.1145/1160633.1160681acmconferencesArticle/Chapter ViewAbstractPublication PagesaamasConference Proceedingsconference-collections
Article

Security in multiagent systems by policy randomization

Authors Info & Claims
Published:08 May 2006Publication History

ABSTRACT

Security in multiagent systems is commonly defined as the ability of the system to deal with intentional threats from other agents. This paper focuses on domains where such intentional threats are caused by unseen adversaries whose actions or payoffs are unknown. In such domains, action randomization can effectively deteriorate an adversary's capability to predict and exploit an agent/agent team's actions. Unfortunately, little attention has been paid to intentional randomization of agents' policies in single-agent or decentralized (PO)MDPs without significantly sacrificing rewards or breaking down coordination. This paper provides two key contributions to remedy this situation. First, it provides three novel algorithms, one based on a non-linear program and two based on linear programs (LP), to randomize single-agent policies, while attaining a certain level of expected reward. Second, it provides Rolling Down Randomization (RDR), a new algorithm that efficiently generates randomized policies for decentralized POMDPs via the single-agent LP method.

References

  1. E. Altman. Constrained Markov Decision Process. Chapman and Hall, 1999.Google ScholarGoogle Scholar
  2. R. Beard and T. McLain. Multiple uav cooperative search under collision avoidance and limited range communication constraints. In IEEE CDC, 2003.Google ScholarGoogle ScholarCross RefCross Ref
  3. D. Bernstein, S. Zilberstein, and N. Immerman. The complexity of decentralized control of MDPs. In UAI, 2000. Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. D. Dolgov and E. Durfee. Constructing optimal policies for agents with constrained architectures. Technical report, Univ of Michigan, CSE-TR-476-03, 2003.Google ScholarGoogle Scholar
  5. R. Emery-Montemerlo, G. Gordon, J. Schneider, and S. Thrun. Approximate solutions for partially observable stochastic games with common payoffs. In AAMAS, 2004. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. E. Hansen, D. Bernstein, and S. Zilberstein. Dynamic programming for partially observable stochastic games. In AAAI, 2004. Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. J. Hu and P. Wellman. Multiagent reinforcement learning: theoretical framework and an algorithm. In ICML, 1998. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. D. A. Huffman. A method for the construction of minimum redundancy codes. In Proc. IRE 40, 1952.Google ScholarGoogle Scholar
  9. T. Jaakkola, S. Singh, and M. Jordan. Reinforcement learning algorithm for partially observable markov decision problems. Advances in NIPS, 7, 1994.Google ScholarGoogle Scholar
  10. M. Littman. Markov games as a framework for multi-agent reinforcement learning. In ML, 1994.Google ScholarGoogle Scholar
  11. R. Nair, D. Pynadath, M. Yokoo, M. Tambe, and S. Marsella. Taming decentralized POMDPs: Towards efficient policy computation for multiagent settings. In IJCAI, 2003. Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. S. Paquet, L. Tobin, and B. Chaib-draa. An online POMDP algorithm for complex multiagent environments. In AAMAS, 2005. Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. P. Paruchuri, M. Tambe, F. Ordonez, and S. Kraus. Towards a formalization of teamwork with resource constraints. In AAMAS, 2004. Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. D. V. Pynadath and M. Tambe. The communicative multiagent team decision problem: Analyzing teamwork theories and models. JAIR, 16:389--423, 2002. Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. C. Shannon. A mathematical theory of communication. The Bell Labs Technical Journal, pages 379--457, 623, 656, 1948.Google ScholarGoogle Scholar
  16. S. Vavasis. Nonlinear optimization: Complexity issues. In University Press, New York, 1991. Google ScholarGoogle ScholarDigital LibraryDigital Library

Recommendations

Comments

Login options

Check if you have access through your login credentials or your institution to get full access on this article.

Sign in
  • Published in

    cover image ACM Conferences
    AAMAS '06: Proceedings of the fifth international joint conference on Autonomous agents and multiagent systems
    May 2006
    1631 pages
    ISBN:1595933034
    DOI:10.1145/1160633

    Copyright © 2006 ACM

    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    • Published: 8 May 2006

    Permissions

    Request permissions about this article.

    Request Permissions

    Check for updates

    Qualifiers

    • Article

    Acceptance Rates

    Overall Acceptance Rate1,155of5,036submissions,23%

PDF Format

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader