Case-based planning: viewing planning as a memory task
Case-based planning: viewing planning as a memory task
Artificial intelligence: a modern approach
Artificial intelligence: a modern approach
Introduction to Reinforcement Learning
Introduction to Reinforcement Learning
Sequential Instance-Based Learning
AI '98 Proceedings of the 12th Biennial Conference of the Canadian Society for Computational Studies of Intelligence on Advances in Artificial Intelligence
Problem-Solving Methods in Artificial Intelligence
Problem-Solving Methods in Artificial Intelligence
Retrieval, reuse, revision and retention in case-based reasoning
The Knowledge Engineering Review
Case-Based Planning and Execution for Real-Time Strategy Games
ICCBR '07 Proceedings of the 7th international conference on Case-Based Reasoning: Case-Based Reasoning Research and Development
ECCBR '08 Proceedings of the 9th European conference on Advances in Case-Based Reasoning
Case-Based Plan Adaptation: An Analysis and Review
IEEE Intelligent Systems
RETALIATE: learning winning policies in first-person shooter games
IAAI'07 Proceedings of the 19th national conference on Innovative applications of artificial intelligence - Volume 2
Transfer learning in real-time strategy games using hybrid CBR/RL
IJCAI'07 Proceedings of the 20th international joint conference on Artifical intelligence
The virtue of reward: performance, reinforcement and discovery in case-based reasoning
ICCBR'05 Proceedings of the 6th international conference on Case-Based Reasoning Research and Development
Learning to win: case-based plan selection in a real-time strategy game
ICCBR'05 Proceedings of the 6th international conference on Case-Based Reasoning Research and Development
Hi-index | 0.00 |
In this paper we study the topic of CBR systems learning from observations in which those observations can be represented as stochastic policies. We describe a general framework which encompasses three steps: (1) it observes agents performing actions, elicits stochastic policies representing the agents' strategies and retains these policies as cases. (2) The agent analyzes the environment and retrieves a suitable stochastic policy. (3) The agent then executes the retrieved stochastic policy, which results in the agent mimicking the previously observed agent. We implement our framework in a system called JuKeCB that observes and mimics players playing games. We present the results of three sets of experiments designed to evaluate our framework. The first experiment demonstrates that JuKeCB performs well when trained against a variety of fixed strategy opponents. The second experiment demonstrates that JuKeCB can also, after training, win against an opponent with a dynamic strategy. The final experiment demonstrates that JuKeCB can win against "new" opponents (i.e. opponents against which JuKeCB is untrained).