Online learning about other agents in a dynamic multiagent system
AGENTS '98 Proceedings of the second international conference on Autonomous agents
Multiagent learning using a variable learning rate
Artificial Intelligence
Introduction to Reinforcement Learning
Introduction to Reinforcement Learning
A polynomial-time nash equilibrium algorithm for repeated games
Proceedings of the 4th ACM conference on Electronic commerce
Best-Response Multiagent Learning in Non-Stationary Environments
AAMAS '04 Proceedings of the Third International Joint Conference on Autonomous Agents and Multiagent Systems - Volume 2
The Role of Reactivity in Multiagent Learning
AAMAS '04 Proceedings of the Third International Joint Conference on Autonomous Agents and Multiagent Systems - Volume 2
Learning to compete, compromise, and cooperate in repeated general-sum games
ICML '05 Proceedings of the 22nd international conference on Machine learning
Learning predictive state representations in dynamical systems without reset
ICML '05 Proceedings of the 22nd international conference on Machine learning
Repeated auctions with complementarities
AMEC'05 Proceedings of the 2005 international conference on Agent-Mediated Electronic Commerce: designing Trading Agents and Mechanisms
Strategic Foresighted Learning in Competitive Multi-Agent Games
Proceedings of the 2006 conference on ECAI 2006: 17th European Conference on Artificial Intelligence August 29 -- September 1, 2006, Riva del Garda, Italy
Hi-index | 0.00 |
Multi-agent learning is a growing area of research. An important topic is to formulate how an agent can learn a good policy in the face of adaptive, competitive opponents. Most research has focused on extensions of single agent learning techniques originally designed for agents in more static environments. These techniques however fail to incorporate a notion of the effect of own previous actions on the development of the policy of the other agents in the system. We argue that incorporation of this property is beneficial in competitive settings. In this paper, we present a novel algorithm to capture this notion, and present experimental results to validate our claims.