Multiagent Systems: A Survey from a Machine Learning Perspective
Autonomous Robots
Sequential Optimality and Coordination in Multiagent Systems
IJCAI '99 Proceedings of the Sixteenth International Joint Conference on Artificial Intelligence
The Complexity of Decentralized Control of Markov Decision Processes
UAI '00 Proceedings of the 16th Conference on Uncertainty in Artificial Intelligence
Algorithms for sequential decision-making
Algorithms for sequential decision-making
Coevolutive planning in markov decision processes
Proceedings of the first international joint conference on Autonomous agents and multiagent systems: part 2
Minimizing communication cost in a distributed Bayesian network using a decentralized MDP
AAMAS '03 Proceedings of the second international joint conference on Autonomous agents and multiagent systems
Communication for Improving Policy Computation in Distributed POMDPs
AAMAS '04 Proceedings of the Third International Joint Conference on Autonomous Agents and Multiagent Systems - Volume 3
Interac-DEC-MDP: Towards the Use of Interactions in DEC-MDP
AAMAS '04 Proceedings of the Third International Joint Conference on Autonomous Agents and Multiagent Systems - Volume 3
Hybrid BDI-POMDP framework for multiagent teaming
Journal of Artificial Intelligence Research
Optimal and approximate Q-value functions for decentralized POMDPs
Journal of Artificial Intelligence Research
Taming decentralized POMDPs: towards efficient policy computation for multiagent settings
IJCAI'03 Proceedings of the 18th international joint conference on Artificial intelligence
An investigation into mathematical programming for finite horizon decentralized POMDPs
Journal of Artificial Intelligence Research
Solving efficiently Decentralized MDPs with temporal and resource constraints
Autonomous Agents and Multi-Agent Systems
Coordinating teams in uncertain environments: a hybrid BDI-POMDP approach
ProMAS'04 Proceedings of the Second international conference on Programming Multi-Agent Systems
An optimal best-first search algorithm for solving infinite horizon DEC-POMDPs
ECML'05 Proceedings of the 16th European conference on Machine Learning
ACTIDS: an active strategy for detecting and localizing network attacks
Proceedings of the 2013 ACM workshop on Artificial intelligence and security
Map partitioning to approximate an exploration strategy in mobile robotics
Multiagent and Grid Systems
Hi-index | 0.00 |
Defining the behaviour of a set of situated agents, such that a collaborative problem can be solved is a key issue in multi-agent systems. In this paper, we formulate this problem from the decision theoretic perspective using the framework of Decentralized Partially Observable Markov Decision Processes (DEC-POMDP). Formulating the coordination problem in this way provides a formal foundation for study of cooperation activities. But, as it has been recently shown solving DEC-POMDP is NEXP-complete and thus it is not a realistic approach for the design of agent cooperation policies. However, we demonstrate in this paper that it is not completely desperate. Indeed, we propose an heuristic approach for solving DEC-POMDP when agents are memory-less and when the global reward function can be broken up into a sum of local reward functions. We demonstrate experimentally on an example (the so-called pursuit problem) that this heuristic is efficient within a few iteration steps.