[go: up one dir, main page]
More Web Proxy on the site http://driver.im/ skip to main content
10.5555/2484920.2485132acmotherconferencesArticle/Chapter ViewAbstractPublication PagesaamasConference Proceedingsconference-collections
poster

Exploiting structure and utilizing agent-centric rewards to promote coordination in large multiagent systems

Published: 06 May 2013 Publication History

Abstract

A goal within the field of multiagent systems is to achieve scaling to large systems involving hundreds or thousands of agents. In such systems the communication requirements for agents as well as the individual agents' ability to make decisions both play critical roles in performance. We take an incremental step towards improving scalability in such systems by introducing a novel algorithm that conglomerates three well-known existing techniques to address both agent communication requirements as well as decision making within large multiagent systems. In particular, we couple a Factored-Action Factored Markov Decision Process (FA-FMDP) framework which exploits problem structure and establishes localized rewards for agents (reducing communication requirements) with reinforcement learning using agent-centric difference rewards which addresses agent decision making and promotes coordination by addressing the structural credit assignment problem. We demonstrate our algorithms performance compared to two other popular reward techniques (global, local) with up to 10,000 agents.

References

[1]
A. Agogino, C. HolmesParker, and K. Tumer. Evolving large scale uav communication system. In Proceedings of the Genetic and Evolutionary Computation Conference (GECCO), Philadelphia, PA, July 2012.
[2]
C. Boutilier, T. Dean, and S. Hanks. Decision-theoretic planning - structural assumptions and computational leverage. Journal of Artificial Intelligence Research (JAIR), 1999.
[3]
C. Guestrin, D. Koller, R. Parr, and S. Venkataraman. Efficient solution algorithms for factored mdps. Journal of Artificial Intelligence Research (JAIR), 2003.
[4]
L. Panait and S. Luke. Cooperative multi-agent learning - the state of the art. Journal of Autonomous Agents and MultiAgent Systems (JAAMAS), 2005.
[5]
A. Strehl, C. Diuk, and M. Littman. Efficient structure learning in factored-state mdps. Association for the Advancement of Artificial Intelligence (AAAI), 2007.

Index Terms

  1. Exploiting structure and utilizing agent-centric rewards to promote coordination in large multiagent systems

    Recommendations

    Comments

    Please enable JavaScript to view thecomments powered by Disqus.

    Information & Contributors

    Information

    Published In

    cover image ACM Other conferences
    AAMAS '13: Proceedings of the 2013 international conference on Autonomous agents and multi-agent systems
    May 2013
    1500 pages
    ISBN:9781450319935

    Sponsors

    • IFAAMAS

    In-Cooperation

    Publisher

    International Foundation for Autonomous Agents and Multiagent Systems

    Richland, SC

    Publication History

    Published: 06 May 2013

    Check for updates

    Author Tags

    1. factored MDPs
    2. reward shaping
    3. scalability

    Qualifiers

    • Poster

    Conference

    AAMAS '13
    Sponsor:

    Acceptance Rates

    AAMAS '13 Paper Acceptance Rate 140 of 599 submissions, 23%;
    Overall Acceptance Rate 1,155 of 5,036 submissions, 23%

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • 0
      Total Citations
    • 78
      Total Downloads
    • Downloads (Last 12 months)1
    • Downloads (Last 6 weeks)0
    Reflects downloads up to 18 Jan 2025

    Other Metrics

    Citations

    View Options

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media