Acting optimally in partially observable stochastic domains
AAAI'94 Proceedings of the twelfth national conference on Artificial intelligence (vol. 2)
Communication decisions in multi-agent cooperation: model and experiments
Proceedings of the fifth international conference on Autonomous agents
Planning, Learning and Coordination in Multiagent Decision Processes
Proceedings of the Sixth Conference on Theoretical Aspects of Rationality and Knowledge
Learning to Cooperate via Policy Search
UAI '00 Proceedings of the 16th Conference on Uncertainty in Artificial Intelligence
The Complexity of Decentralized Control of Markov Decision Processes
UAI '00 Proceedings of the 16th Conference on Uncertainty in Artificial Intelligence
The communicative multiagent team decision problem: analyzing teamwork theories and models
Journal of Artificial Intelligence Research
Taming decentralized POMDPs: towards efficient policy computation for multiagent settings
IJCAI'03 Proceedings of the 18th international joint conference on Artificial intelligence
Reasoning about joint beliefs for execution-time communication decisions
Proceedings of the fourth international joint conference on Autonomous agents and multiagent systems
Playing games in many possible worlds
EC '06 Proceedings of the 7th ACM conference on Electronic commerce
Decentralized planning under uncertainty for teams of communicating agents
AAMAS '06 Proceedings of the fifth international joint conference on Autonomous agents and multiagent systems
Security in multiagent systems by policy randomization
AAMAS '06 Proceedings of the fifth international joint conference on Autonomous agents and multiagent systems
Stochastic planning for weakly-coupled distributed agents
AAMAS '06 Proceedings of the fifth international joint conference on Autonomous agents and multiagent systems
Letting loose a SPIDER on a network of POMDPs: generating quality guaranteed policies
Proceedings of the 6th international joint conference on Autonomous agents and multiagent systems
Q-value functions for decentralized POMDPs
Proceedings of the 6th international joint conference on Autonomous agents and multiagent systems
Subjective approximate solutions for decentralized POMDPs
Proceedings of the 6th international joint conference on Autonomous agents and multiagent systems
Not all agents are equal: scaling up distributed POMDPs for agent networks
Proceedings of the 7th international joint conference on Autonomous agents and multiagent systems - Volume 1
Interaction-driven Markov games for decentralized multiagent planning under uncertainty
Proceedings of the 7th international joint conference on Autonomous agents and multiagent systems - Volume 1
Solving Large-Scale and Sparse-Reward DEC-POMDPs with Correlation-MDPs
RoboCup 2007: Robot Soccer World Cup XI
Lossless clustering of histories in decentralized POMDPs
Proceedings of The 8th International Conference on Autonomous Agents and Multiagent Systems - Volume 1
Decentralised dynamic task allocation: a practical game: theoretic approach
Proceedings of The 8th International Conference on Autonomous Agents and Multiagent Systems - Volume 2
Point-based dynamic programming for DEC-POMDPs
AAAI'06 proceedings of the 21st national conference on Artificial intelligence - Volume 2
Anytime coordination using separable bilinear programs
AAAI'07 Proceedings of the 22nd national conference on Artificial intelligence - Volume 1
Perseus: randomized point-based value iteration for POMDPs
Journal of Artificial Intelligence Research
Optimal and approximate Q-value functions for decentralized POMDPs
Journal of Artificial Intelligence Research
Policy iteration for decentralized control of Markov decision processes
Journal of Artificial Intelligence Research
Memory-bounded dynamic programming for DEC-POMDPs
IJCAI'07 Proceedings of the 20th international joint conference on Artifical intelligence
A bilinear programming approach for multiagent planning
Journal of Artificial Intelligence Research
Bounded policy iteration for decentralized POMDPs
IJCAI'05 Proceedings of the 19th international joint conference on Artificial intelligence
Planning for weakly-coupled partially observable stochastic games
IJCAI'05 Proceedings of the 19th international joint conference on Artificial intelligence
MDP based active localization for multiple robots
CASE'09 Proceedings of the fifth annual IEEE international conference on Automation science and engineering
Heuristic search for identical payoff Bayesian games
Proceedings of the 9th International Conference on Autonomous Agents and Multiagent Systems: volume 1 - Volume 1
Point-based policy generation for decentralized POMDPs
Proceedings of the 9th International Conference on Autonomous Agents and Multiagent Systems: volume 1 - Volume 1
Point-based backup for decentralized POMDPs: complexity and new algorithms
Proceedings of the 9th International Conference on Autonomous Agents and Multiagent Systems: volume 1 - Volume 1
Optimizing fixed-size stochastic controllers for POMDPs and decentralized POMDPs
Autonomous Agents and Multi-Agent Systems
Online planning for multi-agent systems with bounded communication
Artificial Intelligence
Decentralized MDPs with sparse interactions
Artificial Intelligence
Solving efficiently Decentralized MDPs with temporal and resource constraints
Autonomous Agents and Multi-Agent Systems
Scaling up optimal heuristic search in Dec-POMDPs via incremental expansion
IJCAI'11 Proceedings of the Twenty-Second international joint conference on Artificial Intelligence - Volume Volume Three
Heuristic search of multiagent influence space
Proceedings of the 11th International Conference on Autonomous Agents and Multiagent Systems - Volume 2
Approximate solutions for factored Dec-POMDPs with many agents
Proceedings of the 2013 international conference on Autonomous agents and multi-agent systems
Incremental clustering and expansion for faster optimal planning in decentralized POMDPs
Journal of Artificial Intelligence Research
Sufficient plan-time statistics for decentralized POMDPs
IJCAI'13 Proceedings of the Twenty-Third international joint conference on Artificial Intelligence
Hi-index | 0.00 |
Partially observable decentralized decision making in robot teams is fundamentally different from decision making in fully observable problems. Team members cannot simply apply single-agent solution techniques in parallel. Instead, we must turn to game theoretic frameworks to correctly model the problem. While partially observable stochastic games (POSGs) provide a solution model for decentralized robot teams, this model quickly becomes intractable. We propose an algorithm that approximates POSGs as a series of smaller, related Bayesian games, using heuristics such as QMDP to provide the future discounted value of actions. This algorithm trades off limited look-ahead in uncertainty for computational feasibility, and results in policies that are locally optimal with respect to the selected heuristic. Empirical results are provided for both a simple problem for which the full POSG can also be constructed, as well as more complex, robot-inspired, problems.