Project Hoover: auto-scaling streaming map-reduce applications

  • Authors:
  • Rajalakshmi Ramesh;Liting Hu;Karsten Schwan

  • Affiliations:
  • Georgia Institute of Technology, Atlanta, USA;Georgia Institute of Technology, Atlanta, USA;Georgia Institute of Technology, Atlanta, USA

  • Venue:
  • Proceedings of the 2012 workshop on Management of big data systems
  • Year:
  • 2012

Quantified Score

Hi-index 0.00

Visualization

Abstract

Real-time data processing frameworks like S4 and Flume have become scalable and reliable solutions for acquiring, moving, and processing voluminous amounts of data continuously produced by large numbers of online sources. Yet these frameworks lack the elasticity to horizontally scale-up or scale-down their based on current rates of input events and desired event processing latencies. The Project Hoover middleware provides distributed methods for measuring, aggregating, and analyzing the performance of distributed Flume components, thereby enabling online configuration changes to meet varying processing demands. Experimental evaluations with a sample Flume data processing code show Hoover's approach to be capable of dynamically and continuously monitoring Flume performance, demonstrating that such data can be used to right-size the number of Flume collectors according to different log production rates.