Solving very large weakly coupled Markov decision processes
AAAI '98/IAAI '98 Proceedings of the fifteenth national/tenth conference on Artificial intelligence/Innovative applications of artificial intelligence
How to dynamically merge Markov decision processes
NIPS '97 Proceedings of the 1997 conference on Advances in neural information processing systems 10
Neuro-Dynamic Programming
The Complexity of Decentralized Control of Markov Decision Processes
Mathematics of Operations Research
Implicit Imitation in Multiagent Reinforcement Learning
ICML '99 Proceedings of the Sixteenth International Conference on Machine Learning
Exact and approximate algorithms for partially observable markov decision processes
Exact and approximate algorithms for partially observable markov decision processes
Nash q-learning for general-sum stochastic games
The Journal of Machine Learning Research
Approximate Solutions for Partially Observable Stochastic Games with Common Payoffs
AAMAS '04 Proceedings of the Third International Joint Conference on Autonomous Agents and Multiagent Systems - Volume 1
Planning, learning and coordination in multiagent decision processes
TARK '96 Proceedings of the 6th conference on Theoretical aspects of rationality and knowledge
Exploiting factored representations for decentralized execution in multiagent teams
Proceedings of the 6th international joint conference on Autonomous agents and multiagent systems
Exploiting locality of interaction in factored Dec-POMDPs
Proceedings of the 7th international joint conference on Autonomous agents and multiagent systems - Volume 1
Networked distributed POMDPs: a synthesis of distributed constraint optimization and POMDPs
AAAI'05 Proceedings of the 20th national conference on Artificial intelligence - Volume 1
The communicative multiagent team decision problem: analyzing teamwork theories and models
Journal of Artificial Intelligence Research
Solving transition independent decentralized Markov decision processes
Journal of Artificial Intelligence Research
A framework for sequential planning in multi-agent settings
Journal of Artificial Intelligence Research
Optimal and approximate Q-value functions for decentralized POMDPs
Journal of Artificial Intelligence Research
Bounded policy iteration for decentralized POMDPs
IJCAI'05 Proceedings of the 19th international joint conference on Artificial intelligence
Lossless clustering of histories in decentralized POMDPs
Proceedings of The 8th International Conference on Autonomous Agents and Multiagent Systems - Volume 1
Learning of coordination: exploiting sparse interactions in multiagent systems
Proceedings of The 8th International Conference on Autonomous Agents and Multiagent Systems - Volume 2
Exploiting locality of interactions using a policy-gradient approach in multiagent learning
Proceedings of the 2008 conference on ECAI 2008: 18th European Conference on Artificial Intelligence
Optimal and approximate Q-value functions for decentralized POMDPs
Journal of Artificial Intelligence Research
Learning multi-agent state space representations
Proceedings of the 9th International Conference on Autonomous Agents and Multiagent Systems: volume 1 - Volume 1
Approximate planning for decentralized MDPs with sparse interactions
Proceedings of the 9th International Conference on Autonomous Agents and Multiagent Systems: volume 1 - Volume 1
Online planning for multi-agent systems with bounded communication
Artificial Intelligence
Decentralized MDPs with sparse interactions
Artificial Intelligence
Coordinated learning for loosely coupled agents with sparse interactions
AI'11 Proceedings of the 24th international conference on Advances in Artificial Intelligence
Continuous time planning for multiagent teams with temporal constraints
IJCAI'11 Proceedings of the Twenty-Second international joint conference on Artificial Intelligence - Volume Volume One
Distributed value functions for the coordination of decentralized decision makers
Proceedings of the 11th International Conference on Autonomous Agents and Multiagent Systems - Volume 3
Exploiting independent relationships in multiagent systems for coordinated learning
PRICAI'12 Proceedings of the 12th Pacific Rim international conference on Trends in Artificial Intelligence
QueryPOMDP: POMDP-based communication in multiagent systems
EUMAS'11 Proceedings of the 9th European conference on Multi-Agent Systems
Incremental clustering and expansion for faster optimal planning in decentralized POMDPs
Journal of Artificial Intelligence Research
Hi-index | 0.00 |
In this paper we propose interaction-driven Markov games (IDMGs), a new model for multiagent decision making under uncertainty. IDMGs aim at describing multiagent decision problems in which interaction among agents is a local phenomenon. To this purpose, we explicitly distinguish between situations in which agents should interact and situations in which they can afford to act independently. The agents are coupled through the joint rewards and joint transitions in the states in which they interact. The model combines several fundamental properties from transition-independent Dec-MDPs and weakly coupled MDPs while allowing to address, in several aspects, more general problems. We introduce a fast approximate solution method for planning in IDMGs, exploiting their particular structure, and we illustrate its successful application on several large multiagent tasks.