Using temporal-difference learning for multi-agent bargaining

  • Authors:
  • Shiu-li Huang;Fu-ren Lin

  • Affiliations:
  • Department of Information Management, Ming Chuan University, 333 Taoyuan, Taiwan, ROC;Graduate Institute of Technology Management, National Tsing Hua University, 300 Hsinchu, Taiwan, ROC

  • Venue:
  • Electronic Commerce Research and Applications
  • Year:
  • 2008

Quantified Score

Hi-index 0.00

Visualization

Abstract

This research treats a bargaining process as a Markov decision process, in which a bargaining agent's goal is to learn the optimal policy that maximizes the total rewards it receives over the process. Reinforcement learning is an effective method for agents to learn how to determine actions for any time steps in a Markov decision process. Temporal-difference (TD) learning is a fundamental method for solving the reinforcement learning problem, and it can tackle the temporal credit assignment problem. This research designs agents that apply TD-based reinforcement learning to deal with online bilateral bargaining with incomplete information. This research further evaluates the agents' bargaining performance in terms of the average payoff and settlement rate. The results show that agents using TD-based reinforcement learning are able to achieve good bargaining performance. This learning approach is sufficiently robust and convenient, hence it is suitable for online automated bargaining in electronic commerce.