Skip to main content

Heuristic Planning for Decentralized MDPs with Sparse Interactions

  • Chapter
Distributed Autonomous Robotic Systems

Part of the book series: Springer Tracts in Advanced Robotics ((STAR,volume 83))

Abstract

In this work, we explore how local interactions can simplify the process of decision-making in multiagent systems, particularly in multirobot problems. We review a recent decision-theoretic model for multiagent systems, the decentralized sparse-interaction Markov decision process (Dec-SIMDP), that explicitly distinguishes the situations in which the agents in the team must coordinate from those in which they can act independently. We situate this class of problems within different multiagent models, such as MMDPs and transition independent Dec-MDPs. We then contribute a new general approach that leverages the particular structure of Dec-SIMDPs to efficiently plan in this class of problems, and propose two algorithms based on this underlying approach. We pinpoint the main properties of our approach through illustrative examples in multirobot navigation domains with partial observability, and provide empirical comparisons between our algorithms and other existing algorithms for this class of problems. We show that our approach allows the robots to look ahead for possible interactions, planning to accommodate such interactions and thus overcome some of the limitations of previous methods.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 84.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 109.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Allen, M., Zilberstein, S.: Complexity of Decentralized Control: Special Cases. In: Adv. Neural. Information Proc. Systems, pp. 19–27 (2009)

    Google Scholar 

  2. Becker, R., Zilberstein, S., Lesser, V., Goldman, C.: Solving transition independent decentralized Markov decision processes. J. Artif. Intell. Res. 22, 423–455 (2004)

    MathSciNet  MATH  Google Scholar 

  3. Bernstein, D., Givan, R., Immerman, N., Zilberstein, S.: The complexity of decentralized control of Markov decision processes. Math. Oper. Res. 27(4), 819–840 (2002)

    Article  MathSciNet  MATH  Google Scholar 

  4. Bernstein, D., Amato, C., Zilberstein, S.: Policy iteration for decentralized control of Markov decision processes. J. Artif. Intell. Res. 34, 89–132 (2009)

    MathSciNet  MATH  Google Scholar 

  5. Gerkey, B., Matarić, M.: Sold!: Auction methods for multirobot coordination. IEEE T. Robot. Autom. 18(5), 758–768 (2002)

    Article  Google Scholar 

  6. Ghavamzadeh, M., Mahadevan, S., Makar, R.: Hierarchical multiagent reinforcement learning. J. Auton. Agent Multiag. 13(2), 197–229 (2006)

    Article  Google Scholar 

  7. Goldman, C., Zilberstein, S.: Decentralized control of cooperative systems: Categorization and complexity analysis. J. Artif. Intell. Res. 22, 143–174 (2004)

    MathSciNet  MATH  Google Scholar 

  8. Guestrin, C., Koller, D., Parr, R.: Multiagent planning with factored MDPs. In: Adv. Neural Information Proc. Systems, pp. 1523–1530 (2001)

    Google Scholar 

  9. Kearns, M., Littman, M., Singh, S.: Graphical models for game theory. In: Conf. Uncert. Artif. Intell., pp. 253–260 (2001)

    Google Scholar 

  10. Kok, J., Hoen, P., Bakker, B., Vlassis, N.: Utile coordination: Learning interdependencies among cooperative agents. In: IEEE Symp. Comput. Intell. Games, pp. 61–68 (2005)

    Google Scholar 

  11. Littman, M., Cassandra, A., Kaelbling, L.: Learning policies for partially observable environments: Scaling up. In: Int. Conf. Mach. Learn., pp. 362–370 (1995)

    Google Scholar 

  12. Madani, O., Hanks, S., Condon, A.: On the undecidability of probabilistic planning in infinite-horizon partially observable Markov decision problems. In: AAAI Conf. Artif. Intell., pp. 541–548 (1999)

    Google Scholar 

  13. Melo, F., Veloso, M.: Local Multiagent Coordination in Decentralized MDPs with Sparse Interactions. Tech. Report CMU-CS-10-133, CS Dep., Carnegie Mellon Univ. (2010)

    Google Scholar 

  14. Mostafa, H., Lesser, V.: Offline planning for communication by exploiting structured interactions in decentralized MDPs. Tech Rep. TR 2009-020, CS Dep., Univ. Massachusetts (2009)

    Google Scholar 

  15. Parker, L.: ALLIANCE: An architecture for fault-tolerant multirobot cooperation. IEEE T. Robot. Autom. 14(2), 220–240 (1998)

    Article  Google Scholar 

  16. Roth, M., Simmons, R., Veloso, M.: Exploiting factored representations for decentralized execution in multiagent teams. In: Int. Conf. Auton. Agent Multiag., pp. 469–475 (2007)

    Google Scholar 

  17. Spaan, M., Melo, F.: Interaction-driven Markov games for decentralized multiagent planning under uncertainty. In: Int. Conf. Auton. Agent Multiag., pp. 525–532 (2008)

    Google Scholar 

  18. Stone, P.: Layered learning in multiagent systems. PhD thesis, Carnegie Mellon Univ. (1998)

    Google Scholar 

  19. Varakantham, P., Kwak, J., Taylor, M., Marecki, J., Scerri, P., Tambe, M.: Exploiting coordination locales in distributed POMDPs via social model shaping. In: Int. Conf. Autom. Plan Scheduling, pp. 313–320 (2009)

    Google Scholar 

  20. Xin Jiang, A., Leyton-Brown, K., Bhat, N.: Action-graph games. Tech Rep. TR-2008-13, Univ. British Columbia (2008)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Francisco S. Melo .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2013 Springer-Verlag Berlin Heidelberg

About this chapter

Cite this chapter

Melo, F.S., Veloso, M. (2013). Heuristic Planning for Decentralized MDPs with Sparse Interactions. In: Martinoli, A., et al. Distributed Autonomous Robotic Systems. Springer Tracts in Advanced Robotics, vol 83. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-32723-0_24

Download citation

  • DOI: https://doi.org/10.1007/978-3-642-32723-0_24

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-642-32722-3

  • Online ISBN: 978-3-642-32723-0

  • eBook Packages: EngineeringEngineering (R0)

Publish with us

Policies and ethics