Technical Note: \cal Q-Learning
Machine Learning
Reinforcement learning with replacing eligibility traces
Machine Learning - Special issue on reinforcement learning
Artificial Intelligence Review - Special issue on lazy learning
Approximate nearest neighbors: towards removing the curse of dimensionality
STOC '98 Proceedings of the thirtieth annual ACM symposium on Theory of computing
Introduction to Reinforcement Learning
Introduction to Reinforcement Learning
Learning to Predict by the Methods of Temporal Differences
Machine Learning
Workshop summary: Results of the 2009 reinforcement learning competition
ICML '09 Proceedings of the 26th Annual International Conference on Machine Learning
The kNN-TD Reinforcement Learning Algorithm
IWINAC '09 Proceedings of the 3rd International Work-Conference on The Interplay Between Natural and Artificial Computation: Part I: Methods and Models in Artificial and Natural Computation. A Homage to Professor Mira's Scientific Legacy
Reinforcement learning: a survey
Journal of Artificial Intelligence Research
A k-NN based perception scheme for reinforcement learning
EUROCAST'07 Proceedings of the 11th international conference on Computer aided systems theory
Nearest neighbor pattern classification
IEEE Transactions on Information Theory
Self-teaching adaptive dynamic programming for Gomoku
Neurocomputing
A modular hierarchical reinforcement learning algorithm
ICIC'12 Proceedings of the 8th international conference on Intelligent Computing Theories and Applications
Hi-index | 0.03 |
The aim of this paper is to present (jointly) a series of robust high performance (award winning) implementations of reinforcement learning algorithms based on temporal-difference learning and weighted k- nearest neighbors for linear function approximation. These algorithms, named kNN@?TD(@l) methods, where rigorously tested at the Second and Third Annual Reinforcement Learning Competitions (RLC2008 and RCL2009) held in Helsinki and Montreal respectively, where the kNN@?TD(@l) method (JAMH team) won in the PolyAthlon 2008 domain, obtained the second place in 2009 and also the second place in the Mountain-Car 2008 domain showing that it is one of the state of the art general purpose reinforcement learning implementations. These algorithms are able to learn quickly, to generalize properly over continuous state spaces and also to be robust to a high degree of environmental noise. Furthermore, we describe a derivation of kNN@?TD(@l) algorithm for problems where the use of continuous actions have clear advantages over the use of fine grained discrete actions: the Ex reinforcement learning algorithm.