Efficient reinforcement learning
COLT '94 Proceedings of the seventh annual conference on Computational learning theory
An introduction to computational learning theory
An introduction to computational learning theory
Learning to act using real-time dynamic programming
Artificial Intelligence - Special volume on computational research on interaction and agency, part 1
Introduction to Reinforcement Learning
Introduction to Reinforcement Learning
Neuro-Dynamic Programming
Near-Optimal Reinforcement Learning in Polynomial Time
Machine Learning
Expected Mistake Bound Model for On-Line Reinforcement Learning
ICML '97 Proceedings of the Fourteenth International Conference on Machine Learning
Introduction to Probability Models, Ninth Edition
Introduction to Probability Models, Ninth Edition
The Journal of Machine Learning Research
Efficient PAC Learning for Episodic Tasks with Acyclic State Spaces
Discrete Event Dynamic Systems
Hi-index | 0.00 |
The work presented in this paper provides a practical, customized learning algorithm for reinforcement learning tasks that evolve episodically over acyclic state spaces. The presented results are motivated by the Optimal Disassembly Planning (ODP) problem described in [14], and they complement and enhance some earlier developments on this problem that were presented in [15]. In particular, the proposed algorithm is shown to be a substantial improvement of the original algorithm developed in [15], in terms of, both, the involved computational effort and the attained performance, where the latter is measured by the accumulated reward. The new algorithm also leads to a robust performance gain over the typical Q-learning implementations for the considered problem context.