On learning algorithms for nash equilibria

  • Authors:
  • Constantinos Daskalakis;Rafael Frongillo;Christos H. Papadimitriou;George Pierrakos;Gregory Valiant

  • Affiliations:
  • MIT;UC Berkeley;UC Berkeley;UC Berkeley;UC Berkeley

  • Venue:
  • SAGT'10 Proceedings of the Third international conference on Algorithmic game theory
  • Year:
  • 2010

Quantified Score

Hi-index 0.00

Visualization

Abstract

Can learning algorithms find a Nash equilibrium? This is a natural question for several reasons. Learning algorithms resemble the behavior of players in many naturally arising games, and thus results on the convergence or nonconvergence properties of such dynamics may inform our understanding of the applicability of Nash equilibria as a plausible solution concept in some settings. A second reason for asking this question is in the hope of being able to prove an impossibility result, not dependent on complexity assumptions, for computing Nash equilibria via a restricted class of reasonable algorithms. In this work, we begin to answer this question by considering the dynamics of the standard multiplicative weights update learning algorithms (which are known to converge to a Nash equilibrium for zero-sum games). We revisit a 3×3 game defined by Shapley [10] in the 1950s in order to establish that fictitious play does not converge in general games. For this simple game, we show via a potential function argument that in a variety of settings the multiplicative updates algorithm impressively fails to find the unique Nash equilibrium, in that the cumulative distributions of players produced by learning dynamics actually drift away from the equilibrium.