Asynchronous action-reward learning for nonstationary serial supply chain inventory control

  • Authors:
  • Chang Ouk Kim;Ick-Hyun Kwon;Jun-Geol Baek

  • Affiliations:
  • Department of Information and Industrial Engineering, Yonsei University, Seoul, Republic of Korea 120-749;Department of Civil and Environmental Engineering, University of Illinois at Urbana-Champaign, Urbana, USA 61801;Department of Business Administration, Kwangwoon University, Seoul, Republic of Korea 139-701

  • Venue:
  • Applied Intelligence
  • Year:
  • 2008

Quantified Score

Hi-index 0.00

Visualization

Abstract

Action-reward learning is a reinforcement learning method. In this machine learning approach, an agent interacts with non-deterministic control domain. The agent selects actions at decision epochs and the control domain gives rise to rewards with which the performance measures of the actions are updated. The objective of the agent is to select the future best actions based on the updated performance measures. In this paper, we develop an asynchronous action-reward learning model which updates the performance measures of actions faster than conventional action-reward learning. This learning model is suitable to apply to nonstationary control domain where the rewards for actions vary over time. Based on the asynchronous action-reward learning, two situation reactive inventory control models (centralized and decentralized models) are proposed for a two-stage serial supply chain with nonstationary customer demand. A simulation based experiment was performed to evaluate the performance of the proposed two models.