The Complexity of Decentralized Control of Markov Decision Processes
Mathematics of Operations Research
Learning to Cooperate via Policy Search
UAI '00 Proceedings of the 16th Conference on Uncertainty in Artificial Intelligence
Transition-independent decentralized markov decision processes
AAMAS '03 Proceedings of the second international joint conference on Autonomous agents and multiagent systems
Decentralized Markov Decision Processes with Event-Driven Interactions
AAMAS '04 Proceedings of the Third International Joint Conference on Autonomous Agents and Multiagent Systems - Volume 1
A polynomial algorithm for decentralized Markov decision processes with temporal constraints
Proceedings of the fourth international joint conference on Autonomous agents and multiagent systems
The communicative multiagent team decision problem: analyzing teamwork theories and models
Journal of Artificial Intelligence Research
Decentralized control of cooperative systems: categorization and complexity analysis
Journal of Artificial Intelligence Research
Taming decentralized POMDPs: towards efficient policy computation for multiagent settings
IJCAI'03 Proceedings of the 18th international joint conference on Artificial intelligence
A Q-decomposition and bounded RTDP approach to resource allocation
Proceedings of the 6th international joint conference on Autonomous agents and multiagent systems
Proceedings of the 6th international joint conference on Autonomous agents and multiagent systems
Planning with continuous resources for agent teams
Proceedings of The 8th International Conference on Autonomous Agents and Multiagent Systems - Volume 2
A Cooperative Distributed Problem Solving Technique for Large Markov Decision Processes
Proceedings of the 2006 conference on ECAI 2006: 17th European Conference on Artificial Intelligence August 29 -- September 1, 2006, Riva del Garda, Italy
Vector Valued Markov Decision Process for robot platooning
Proceedings of the 2008 conference on ECAI 2008: 18th European Conference on Artificial Intelligence
Optimal and approximate Q-value functions for decentralized POMDPs
Journal of Artificial Intelligence Research
Online planning for multi-agent systems with bounded communication
Artificial Intelligence
Hi-index | 0.00 |
Despite the significant progress to extend Markov Decision Processes (MDP) to cooperative multi-agent systems, developing approaches that can deal with realistic problems remains a serious challenge. Existing approaches that solve Decentralized Markov Decision Processes (DEC-MDPs) suffer from the fact that they can only solve relatively small problems without complex constraints on task execution. OC-DEC-MDP has been introduced to deal with large DEC-MDPs under resource and temporal constraints. However, the proposed algorithm to solve this class of DEC-MDPs has some limits: it suffers from overestimation of opportunity cost and restricts policy improvement to one sweep (or iteration). In this paper, we propose to overcome these limits by first introducing the notion of Expected Opportunity Cost to better assess the influence of a local decision of an agent on the others. We then describe an iterative version of the algorithm to incrementally improve the policies of agents leading to higher quality solutions in some settings. Experimental results are shown to support our claims.