Introduction to Reinforcement Learning
Introduction to Reinforcement Learning
Efficient learning of multi-step best response
Proceedings of the fourth international joint conference on Autonomous agents and multiagent systems
Individual Q-Learning in Normal Form Games
SIAM Journal on Control and Optimization
Learning to cooperate in multi-agent social dilemmas
AAMAS '06 Proceedings of the fifth international joint conference on Autonomous agents and multiagent systems
If multi-agent learning is the answer, what is the question?
Artificial Intelligence
IAMwildCAT: The Winning Strategy for the TAC Market Design Competition
Proceedings of the 2008 conference on ECAI 2008: 18th European Conference on Artificial Intelligence
Planning against fictitious players in repeated normal form games
Proceedings of the 9th International Conference on Autonomous Agents and Multiagent Systems: volume 1 - Volume 1
The lemonade stand game competition: solving unsolvable games
ACM SIGecom Exchanges
Multiagent learning in the presence of memory-bounded agents
Autonomous Agents and Multi-Agent Systems
Hi-index | 0.00 |
We describe the winning strategy of the inaugural Lemonade Stand Game (LSG) Tournament. The LSG is a repeated symmetric 3--player constant--sum finite horizon game, in which a player chooses a location for their lemonade stand on an island with the aim of being as far as possible from its opponents. To receive a high utility in this game, our strategy, EA2, attempts to find a suitable partner with which to coordinate and exploit the third player. To do this, we classify the behaviour of our opponents using the history of joint interactions in order to identify the best player to coordinate with and how this coordination should be established. This approach is designed to be adaptive to various types of opponents such that coordination is almost always achieved, which yields consistently high utilities to our agent, as evidenced by the Tournament results and our subsequent experimental analysis. Our strategy models behaviours of its opponents, rather than situations of the game (e.g. game theoretic equilibrium or off equilibrium paths), which makes EA2 easy to generalize to many other games.