A rapid sparsification method for kernel machines in approximate policy iteration

  • Authors:
  • Chunming Liu;Zhenhua Huang;Xin Xu;Lei Zuo;Jun Wu

  • Affiliations:
  • Institute of Automation, National University of Defense Technology, Changsha, P.R. China;Institute of Automation, National University of Defense Technology, Changsha, P.R. China;Institute of Automation, National University of Defense Technology, Changsha, P.R. China;Institute of Automation, National University of Defense Technology, Changsha, P.R. China;Institute of Automation, National University of Defense Technology, Changsha, P.R. China

  • Venue:
  • ISNN'12 Proceedings of the 9th international conference on Advances in Neural Networks - Volume Part I
  • Year:
  • 2012

Quantified Score

Hi-index 0.00

Visualization

Abstract

Recently approximate policy iteration (API) has received increasing attention due to its good convergence and generalization abilities in solving difficult reinforcement learning (RL) problems, e.g. least-squares policy iteration (LSPI) and its kernelized version (KLSPI). However, the sparsification of feature vectors, especially the kernel-based features, costs much computation and greatly influences the performance of API methods. In this paper, a novel rapid sparsification method is proposed for sparsifying kernel machines in API. In this method, the approximation error of a new feature vector is computed prior in the original space to decide if it is added to the current kernel dictionary, so the computational cost becomes a little higher when the collected samples are sparse, but remarkably lower when the collected samples are dense. Experimental results on the swing-up control of an double-link pendulum verify that the computational cost of the proposed algorithm is lower than that of the previous kernel-based API algorithm, and this performance becomes more and more obvious when the number of the collected samples increases and when the level of sparsification increases.