Fast planning through planning graph analysis
Artificial Intelligence
Solving very large weakly coupled Markov decision processes
AAAI '98/IAAI '98 Proceedings of the fifteenth national/tenth conference on Artificial intelligence/Innovative applications of artificial intelligence
How to dynamically merge Markov decision processes
NIPS '97 Proceedings of the 1997 conference on Advances in neural information processing systems 10
A heuristic approach for solving decentralized-POMDP: assessment on the pursuit problem
Proceedings of the 2002 ACM symposium on Applied computing
The Complexity of Decentralized Control of Markov Decision Processes
Mathematics of Operations Research
Learning to Cooperate via Policy Search
UAI '00 Proceedings of the 16th Conference on Uncertainty in Artificial Intelligence
Probabilistic Planning in the Graphplan Framework
ECP '99 Proceedings of the 5th European Conference on Planning: Recent Advances in AI Planning
Piecewise linear value function approximation for factored MDPs
Eighteenth national conference on Artificial intelligence
Transition-independent decentralized markov decision processes
AAMAS '03 Proceedings of the second international joint conference on Autonomous agents and multiagent systems
Approximate Solutions for Partially Observable Stochastic Games with Common Payoffs
AAMAS '04 Proceedings of the Third International Joint Conference on Autonomous Agents and Multiagent Systems - Volume 1
Decentralized Markov Decision Processes with Event-Driven Interactions
AAMAS '04 Proceedings of the Third International Joint Conference on Autonomous Agents and Multiagent Systems - Volume 1
Communication for Improving Policy Computation in Distributed POMDPs
AAMAS '04 Proceedings of the Third International Joint Conference on Autonomous Agents and Multiagent Systems - Volume 3
Spoken dialogue management using probabilistic reasoning
ACL '00 Proceedings of the 38th Annual Meeting on Association for Computational Linguistics
Modeling task allocation using a decision theoretic model
Proceedings of the fourth international joint conference on Autonomous agents and multiagent systems
Analyzing Myopic Approaches for Multi-Agent Communication
IAT '05 Proceedings of the IEEE/WIC/ACM International Conference on Intelligent Agent Technology
Agent interaction in distributed POMDPs and its implications on complexity
AAMAS '06 Proceedings of the fifth international joint conference on Autonomous agents and multiagent systems
Distributed management of flexible times schedules
Proceedings of the 6th international joint conference on Autonomous agents and multiagent systems
Letting loose a SPIDER on a network of POMDPs: generating quality guaranteed policies
Proceedings of the 6th international joint conference on Autonomous agents and multiagent systems
Proceedings of the 6th international joint conference on Autonomous agents and multiagent systems
Predictability & criticality metrics for coordination in complex environments
Proceedings of the 7th international joint conference on Autonomous agents and multiagent systems - Volume 2
Dynamic programming for partially observable stochastic games
AAAI'04 Proceedings of the 19th national conference on Artifical intelligence
Networked distributed POMDPs: a synthesis of distributed constraint optimization and POMDPs
AAAI'05 Proceedings of the 20th national conference on Artificial intelligence - Volume 1
IJCAI'97 Proceedings of the Fifteenth international joint conference on Artifical intelligence - Volume 2
Decentralized control of cooperative systems: categorization and complexity analysis
Journal of Artificial Intelligence Research
Solving transition independent decentralized Markov decision processes
Journal of Artificial Intelligence Research
Memory-bounded dynamic programming for DEC-POMDPs
IJCAI'07 Proceedings of the 20th international joint conference on Artifical intelligence
Taming decentralized POMDPs: towards efficient policy computation for multiagent settings
IJCAI'03 Proceedings of the 18th international joint conference on Artificial intelligence
Bounded policy iteration for decentralized POMDPs
IJCAI'05 Proceedings of the 19th international joint conference on Artificial intelligence
Decomposition techniques for planning in stochastic domains
IJCAI'95 Proceedings of the 14th international joint conference on Artificial intelligence - Volume 2
Point-based policy generation for decentralized POMDPs
Proceedings of the 9th International Conference on Autonomous Agents and Multiagent Systems: volume 1 - Volume 1
Planning under continuous time and resource uncertainty: a challenge for AI
UAI'02 Proceedings of the Eighteenth conference on Uncertainty in artificial intelligence
Incremental clustering and expansion for faster optimal planning in decentralized POMDPs
Journal of Artificial Intelligence Research
Hi-index | 0.00 |
Optimizing the operation of cooperative multi-agent systems that can deal with large and realistic problems has become an important focal area of research in the multi-agent community. In this paper, we first present a new model, the OC-DEC-MDP (Opportunity Cost Decentralized Markov Decision Process), that allows us to represent large multi-agent decision problems with temporal and precedence constraints. Then, we propose polynomial algorithms to efficiently solve problems formalized by OC-DEC-MDPs. The problems we deal with consist of a set of agents that have to execute a set of tasks in a cooperative way. The agents cannot communicate during task execution and they must respect resource and temporal constraints. Our approach is based on Decentralized Markov Decision Processes (DEC-MDPs) and uses the concept of opportunity cost borrowed from economics to obtain approximate control policies. Experimental results show that our approach produces good quality solutions for complex problems which are out of reach of existing approaches.