Minimizing churn in distributed systems

  • Authors:
  • P. Brighten Godfrey;Scott Shenker;Ion Stoica

  • Affiliations:
  • UC Berkeley;UC Berkeley;UC Berkeley

  • Venue:
  • Proceedings of the 2006 conference on Applications, technologies, architectures, and protocols for computer communications
  • Year:
  • 2006

Quantified Score

Hi-index 0.06

Visualization

Abstract

A pervasive requirement of distributed systems is to deal with churn-change in the set of participating nodes due to joins, graceful leaves, and failures. A high churn rate can increase costs or decrease service quality. This paper studies how to reduce churn by selecting which subset of a set of available nodes to use.First, we provide a comparison of the performance of a range of different node selection strategies in five real-world traces. Among our findings is that the simple strategy of picking a uniform-random replacement whenever a node fails performs surprisingly well. We explain its performance through analysis in a stochastic model.Second, we show that a class of strategies, which we call "Preference List" strategies, arise commonly as a result of optimizing for a metric other than churn, and produce high churn relative to more randomized strategies under realistic node failure patterns. Using this insight, we demonstrate and explain differences in performance for designs that incorporate varying degrees of randomization. We give examples from a variety of protocols, including anycast, over-lay multicast, and distributed hash tables. In many cases, simply adding some randomization can go a long way towards reducing churn.