Tradeoffs in scalable data routing for deduplication clusters

  • Authors:
  • Wei Dong;Fred Douglis;Kai Li;Hugo Patterson;Sazzala Reddy;Philip Shilane

  • Affiliations:
  • Princeton University;EMC;Princeton University and EMC;EMC;EMC;EMC

  • Venue:
  • FAST'11 Proceedings of the 9th USENIX conference on File and stroage technologies
  • Year:
  • 2011

Quantified Score

Hi-index 0.00

Visualization

Abstract

As data have been growing rapidly in data centers, deduplication storage systems continuously face challenges in providing the corresponding throughputs and capacities necessary to move backup data within backup and recovery window times. One approach is to build a cluster deduplication storage system with multiple deduplication storage system nodes. The goal is to achieve scalable throughput and capacity using extremely high-throughput (e.g. 1.5 GB/s) nodes, with a minimal loss of compression ratio. The key technical issue is to route data intelligently at an appropriate granularity. We present a cluster-based deduplication system that can deduplicate with high throughput, support deduplication ratios comparable to that of a single system, and maintain a low variation in the storage utilization of individual nodes. In experiments with dozens of nodes, we examine tradeoffs between stateless data routing approaches with low overhead and stateful approaches that have higher overhead but avoid imbalances that can adversely affect deduplication effectiveness for some datasets in large clusters. The stateless approach has been deployed in a two-node commercial system that achieves 3 GB/s for multi-stream deduplication throughput and currently scales to 5.6 PB of storage (assuming 20X total compression).