The price of approximate stability for scheduling selfish tasks on two links

  • Authors:
  • Eric Angel;Evripidis Bampis;Fanny Pascual

  • Affiliations:
  • IBISC CNRS FRE 2873, Université d'Évry-Val d'Essonne, Évry, France;IBISC CNRS FRE 2873, Université d'Évry-Val d'Essonne, Évry, France;IBISC CNRS FRE 2873, Université d'Évry-Val d'Essonne, Évry, France

  • Venue:
  • Euro-Par'06 Proceedings of the 12th international conference on Parallel Processing
  • Year:
  • 2006

Quantified Score

Hi-index 0.00

Visualization

Abstract

We consider a scheduling game, where a set of selfish agents (traffic loads) want to be routed in exactly one of the two parallel links of a system. Every agent aims to minimize her own completion time, while the social objective is the makespan, i.e. the time at which the last agent finishes her execution. We study the problem of optimizing the makespan under the constraint that the obtained schedule is a (pure) Nash equilibrium, i.e. a schedule in which no agent has incentive to unilaterally change her strategy (link). We consider a relaxation of the notion of equilibrium by considering α-approximate Nash equilibria where an agent does not have sufficient incentive (w.r.t. the value of α) to unilaterally change her strategy. Our main contribution is the study of the tradeoff between the approximation ratio for the makespan and the value of α. We first give an algorithm which provides a solution with an approximation ratio of $\frac{8}{7}$ for the makespan and which is a 3-approximate Nash equilibrium, provided that the local policy of each link is Longest Processing Time (LPT). Furthermore, we show that a slight modification of the classical Polynomial Time Approximation Scheme (PTAS) of Graham allows to obtain a schedule whose makespan is arbitrarily close to the optimum while keeping a constant value for α. Finally, we give bounds establishing relations between the value of α and the best possible value of the approximation ratio, provided that the local policies of the links are LPT.