Proceedings of the seventh international conference (1990) on Machine learning
A challenging set of control problems
Neural networks for control
Variable Resolution Discretization in Optimal Control
Machine Learning
Adaptive Resolution Model-Free Reinforcement Learning: Decision Boundary Partitioning
ICML '00 Proceedings of the Seventeenth International Conference on Machine Learning
Reinforcement learning for dynamic robotic systems
Reinforcement learning for dynamic robotic systems
Cognitive economy and the role of representation in on-line learning
Cognitive economy and the role of representation in on-line learning
Reinforcement learning: a survey
Journal of Artificial Intelligence Research
Agent-based consumer learning in e-commerce
International Journal of Networking and Virtual Organisations
Hi-index | 0.00 |
For an intelligent agent to be truly autonomous, it must be able to adapt its representation to the requirements of its task as it interacts with the world. Most current approaches to on-line feature extraction are ad hoc; in contrast, this paper derives principled criteria for representational adequacy by applying the psychological principle of cognitive economy to reinforcement learning. The criteria are principled because they are based on an analysis of the amount of reward the agent forfeits when it generalizes over states. This analysis shows that action-value errors are sometimes irrelevant, and that the agent may optimize its performance with limited cognitive resources by grouping together states whose differences do not matter in its task. The paper presents an algorithm based on this analysis, incorporating an active form of Q-learning and partitioning continuous state-spaces by merging and splitting Voronoi regions. The experiments illustrate a new methodology for testing and comparing representations by means of learning curves. Results from the puck-on-a-hill task demonstrate the algorithm's ability to learn effective representations, superior to those produced by some other, well-known, methods.