Linear least-squares algorithms for temporal difference learning
Machine Learning - Special issue on reinforcement learning
Artificial Intelligence Review - Special issue on lazy learning
Learning to Predict by the Methods of Temporal Differences
Machine Learning
Computing Factored Value Functions for Policies in Structured MDPs
IJCAI '99 Proceedings of the Sixteenth International Joint Conference on Artificial Intelligence
A decision-theoretic generalization of on-line learning and an application to boosting
EuroCOLT '95 Proceedings of the Second European Conference on Computational Learning Theory
Learning and value function approximation in complex decision processes
Learning and value function approximation in complex decision processes
Least-squares policy iteration
The Journal of Machine Learning Research
Automatic basis function construction for approximate dynamic programming and reinforcement learning
ICML '06 Proceedings of the 23rd international conference on Machine learning
Proceedings of the 25th international conference on Machine learning
Regularized Fitted Q-Iteration: Application to Planning
Recent Advances in Reinforcement Learning
Basis Expansion in Natural Actor Critic Methods
Recent Advances in Reinforcement Learning
Projected equation methods for approximate solution of large linear systems
Journal of Computational and Applied Mathematics
Regularization and feature selection in least-squares temporal difference learning
ICML '09 Proceedings of the 26th Annual International Conference on Machine Learning
Learning Representation and Control in Markov Decision Processes: New Frontiers
Foundations and Trends® in Machine Learning
Feature Selection for Value Function Approximation Using Bayesian Model Selection
ECML PKDD '09 Proceedings of the European Conference on Machine Learning and Knowledge Discovery in Databases: Part I
Regularized fitted Q-iteration for planning in continuous-space Markovian decision problems
ACC'09 Proceedings of the 2009 conference on American Control Conference
Model-based and model-free reinforcement learning for visual servoing
ICRA'09 Proceedings of the 2009 IEEE international conference on Robotics and Automation
Temporal-difference networks for dynamical systems with continuous observations and actions
UAI '09 Proceedings of the Twenty-Fifth Conference on Uncertainty in Artificial Intelligence
Basis function construction for hierarchical reinforcement learning
Proceedings of the 9th International Conference on Autonomous Agents and Multiagent Systems: volume 1 - Volume 1
ECML PKDD'10 Proceedings of the 2010 European conference on Machine learning and knowledge discovery in databases: Part I
Automatic induction of bellman-error features for probabilistic planning
Journal of Artificial Intelligence Research
Metric learning for reinforcement learning agents
The 10th International Conference on Autonomous Agents and Multiagent Systems - Volume 2
Basis function discovery using spectral clustering and bisimulation metrics
The 10th International Conference on Autonomous Agents and Multiagent Systems - Volume 3
Basis function discovery using spectral clustering and bisimulation metrics
ALA'11 Proceedings of the 11th international conference on Adaptive and Learning Agents
Automatic state abstraction from demonstration
IJCAI'11 Proceedings of the Twenty-Second international joint conference on Artificial Intelligence - Volume Volume Two
Construction of approximation spaces for reinforcement learning
The Journal of Machine Learning Research
Hi-index | 0.00 |
We analyze a simple, Bellman-error-based approach to generating basis functions for value-function approximation. We show that it generates orthogonal basis functions that provably tighten approximation error bounds. We also illustrate the use of this approach in the presence of noise on some sample problems.