The dynamics of reinforcement learning in cooperative multiagent systems
AAAI '98/IAAI '98 Proceedings of the fifteenth national/tenth conference on Artificial intelligence/Innovative applications of artificial intelligence
Friend-or-Foe Q-learning in General-Sum Games
ICML '01 Proceedings of the Eighteenth International Conference on Machine Learning
Multiagent Reinforcement Learning: Theoretical Framework and an Algorithm
ICML '98 Proceedings of the Fifteenth International Conference on Machine Learning
Near-Optimal Reinforcement Learning in Polynominal Time
ICML '98 Proceedings of the Fifteenth International Conference on Machine Learning
R-max - a general polynomial time algorithm for near-optimal reinforcement learning
The Journal of Machine Learning Research
Rational and convergent learning in stochastic games
IJCAI'01 Proceedings of the 17th international joint conference on Artificial intelligence - Volume 2
Efficient learning equilibrium
Artificial Intelligence
Learning payoff functions in infinite games
Machine Learning
Exploring selfish reinforcement learning in repeated games with stochastic rewards
Autonomous Agents and Multi-Agent Systems
Joint Equilibrium Policy Search for Multi-Agent Scheduling Problems
MATES '08 Proceedings of the 6th German conference on Multiagent System Technologies
Recent Advances in Reinforcement Learning
Learning payoff functions in infinite games
IJCAI'05 Proceedings of the 19th international joint conference on Artificial intelligence
Computing equilibria in multiplayer stochastic games of imperfect information
IJCAI'09 Proceedings of the 21st international jont conference on Artifical intelligence
Replanning in domains with partial information and sensing actions
Journal of Artificial Intelligence Research
Predicting behavior in unstructured bargaining with a probability distribution
Journal of Artificial Intelligence Research
Hi-index | 0.00 |
In common-interest stochastic games all players receive an identical payoff. Players participating in such games must learn to coordinate with each other in order to receive the highest-possible value. A number of reinforcement learning algorithms have been proposed for this problem, and some have been shown to converge to good solutions in the limit. In this paper we show that using very simple model-based algorithms, much better (i.e., polynomial) convergence rates can be attained. Moreover, our model-based algorithms are guaranteed to converge to the optimal value, unlike many of the existing algorithms.