Principled methods for biasing reinforcement learning agents

  • Authors:
  • Zhi Li;Kun Hu;Zengrong Liu;Xueli Yu

  • Affiliations:
  • College of Computer Science and Technology, Taiyuan University of Technology, Taiyuan, Shanxi, China;College of Computer Science and Technology, Taiyuan University of Technology, Taiyuan, Shanxi, China;College of Computer Science and Technology, Taiyuan University of Technology, Taiyuan, Shanxi, China;College of Computer Science and Technology, Taiyuan University of Technology, Taiyuan, Shanxi, China

  • Venue:
  • AICI'11 Proceedings of the Third international conference on Artificial intelligence and computational intelligence - Volume Part II
  • Year:
  • 2011

Quantified Score

Hi-index 0.00

Visualization

Abstract

Reinforcement learning (RL) is a powerful technique for learning in domains where there is no instructive feedback but only evaluative feedback and is rapidly expanding in industrial and research fields. One of the main limitations of RL is the slowness in convergence. Thus, several methods have been proposed to speed up RL. They involve the incorporation of prior knowledge or bias into RL. In this paper, we present a new method for incorporating bias into RL. This method extends the choosing initial Q-values method proposed by Hailu G. and Sommer G. and one kind of learning mechanism is introduced into agent. This allows for much more specific information to guide the agent which action to choose and meanwhile it is helpful to reduce the state research space. So it improves the learning performance and speed up the convergence of the learning process greatly.