The complexity of Markov decision processes
Mathematics of Operations Research
Planning and acting in partially observable stochastic domains
Artificial Intelligence
Multi-agent policies: from centralized ones to decentralized ones
Proceedings of the first international joint conference on Autonomous agents and multiagent systems: part 3
Learning to Cooperate via Policy Search
UAI '00 Proceedings of the 16th Conference on Uncertainty in Artificial Intelligence
The Complexity of Decentralized Control of Markov Decision Processes
UAI '00 Proceedings of the 16th Conference on Uncertainty in Artificial Intelligence
Value-Directed Sampling Methods for POMDPs
UAI '01 Proceedings of the 17th Conference in Uncertainty in Artificial Intelligence
Transition-independent decentralized markov decision processes
AAMAS '03 Proceedings of the second international joint conference on Autonomous agents and multiagent systems
Approximate Solutions for Partially Observable Stochastic Games with Common Payoffs
AAMAS '04 Proceedings of the Third International Joint Conference on Autonomous Agents and Multiagent Systems - Volume 1
Communication for Improving Policy Computation in Distributed POMDPs
AAMAS '04 Proceedings of the Third International Joint Conference on Autonomous Agents and Multiagent Systems - Volume 3
Dynamic programming for partially observable stochastic games
AAAI'04 Proceedings of the 19th national conference on Artifical intelligence
The communicative multiagent team decision problem: analyzing teamwork theories and models
Journal of Artificial Intelligence Research
Decentralized control of cooperative systems: categorization and complexity analysis
Journal of Artificial Intelligence Research
Taming decentralized POMDPs: towards efficient policy computation for multiagent settings
IJCAI'03 Proceedings of the 18th international joint conference on Artificial intelligence
Exploiting factored representations for decentralized execution in multiagent teams
Proceedings of the 6th international joint conference on Autonomous agents and multiagent systems
Q-value functions for decentralized POMDPs
Proceedings of the 6th international joint conference on Autonomous agents and multiagent systems
A decentralized approach to cooperative situation assessment in multi-robot systems
Proceedings of the 7th international joint conference on Autonomous agents and multiagent systems - Volume 1
Reward shaping for valuing communications during multi-agent coordination
Proceedings of The 8th International Conference on Autonomous Agents and Multiagent Systems - Volume 1
Learning of coordination: exploiting sparse interactions in multiagent systems
Proceedings of The 8th International Conference on Autonomous Agents and Multiagent Systems - Volume 2
Optimal and approximate Q-value functions for decentralized POMDPs
Journal of Artificial Intelligence Research
Offline Planning for Communication by Exploiting Structured Interactions in Decentralized MDPs
WI-IAT '09 Proceedings of the 2009 IEEE/WIC/ACM International Joint Conference on Web Intelligence and Intelligent Agent Technology - Volume 02
Introducing Communication in Dis-POMDPs with Finite State Machines
WI-IAT '09 Proceedings of the 2009 IEEE/WIC/ACM International Joint Conference on Web Intelligence and Intelligent Agent Technology - Volume 02
Point-based policy generation for decentralized POMDPs
Proceedings of the 9th International Conference on Autonomous Agents and Multiagent Systems: volume 1 - Volume 1
Online planning for multi-agent systems with bounded communication
Artificial Intelligence
Two decades of multiagent teamwork research: past, present, and future
CARE@AI'09/CARE@IAT'10 Proceedings of the CARE@AI 2009 and CARE@IAT 2010 international conference on Collaborative agents - research and development
Teamwork in distributed POMDPs: execution-time coordination under model uncertainty
The 10th International Conference on Autonomous Agents and Multiagent Systems - Volume 3
Towards Addressing Model Uncertainty: Robust Execution-Time Coordination for Teamwork
WI-IAT '11 Proceedings of the 2011 IEEE/WIC/ACM International Conferences on Web Intelligence and Intelligent Agent Technology - Volume 02
Cooperative situation assessment in a maritime scenario
International Journal of Intelligent Systems
Modeling information exchange opportunities for effective human-computer teamwork
Artificial Intelligence
Incremental clustering and expansion for faster optimal planning in decentralized POMDPs
Journal of Artificial Intelligence Research
Hi-index | 0.00 |
Just as POMDPs have been used to reason explicitly about uncertainty in single-agent systems, there has been recent interest in using multi-agent POMDPs to coordinate teams of agents in the presence of uncertainty. Although multi-agent POMDPs are known to be highly intractable, communication at every time step transforms a multi-agent POMDP into a more tractable single-agent POMDP. In this paper, we present an approach that generates "centralized" policies for multi-agent POMDPs at plan-time by assuming the presence of free communication, and at run-time, handles the problem of limited communication resources by reasoning about the use of communication as needed for effective execution. This approach trades off the need to do some computation at execution-time for the ability to generate policies more tractably at plan-time. In our algorithm, each agent, at run-time, models the distribution of possible joint beliefs. Joint actions are selected over this distribution, ensuring that agents remain synchronized. Communication is used to integrate local observations into the team belief only when those observations would improve team performance. We show, both through a detailed example and with experimental results, that our approach allows for effective decentralized execution while avoiding unnecessary instances of communication.