Heuristic search based exploration in reinforcement learning

  • Authors:
  • Ngo Anh Vien;Nguyen Hoang Viet;SeungGwan Lee;TaeChoong Chung

  • Affiliations:
  • Artificial Intelligence Lab, Department of Computer Engineering, School of Electronics and Information, Kyunghee University, Yongin, Gyeonggi, South Korea;Artificial Intelligence Lab, Department of Computer Engineering, School of Electronics and Information, Kyunghee University, Yongin, Gyeonggi, South Korea;Artificial Intelligence Lab, Department of Computer Engineering, School of Electronics and Information, Kyunghee University, Yongin, Gyeonggi, South Korea;Artificial Intelligence Lab, Department of Computer Engineering, School of Electronics and Information, Kyunghee University, Yongin, Gyeonggi, South Korea

  • Venue:
  • IWANN'07 Proceedings of the 9th international work conference on Artificial neural networks
  • Year:
  • 2007

Quantified Score

Hi-index 0.00

Visualization

Abstract

In this paper, we consider reinforcement learning in systems with unknown environment where the agent must trade off efficiently between: exploration(long-term optimization) and exploitation (short-term optimization). Ɛ-greedy algorithm is a method using near-greedy action selection rule. It behaves greedily (exploitation) most of the time, but every once in a while, say with small probability Ɛ (exploration), instead select an action at random. Many works already proved that random exploration drives the agent towards poorly modeled states. Therefore, this study evaluates the role of heuristic based exploration in reinforcement learning. We proposed three methods: neighborhood search based exploration, simulated annealing based exploration, and tabu search based exploration. All techniques follow the same rule "Explore the most unvisited state". In the simulation, these techniques are evaluated and compared on a discrete reinforcement learning task (robot navigation).