The family of mapreduce and large-scale data processing systems

  • Authors:
  • Sherif Sakr;Anna Liu;Ayman G. Fayoumi

  • Affiliations:
  • NICTA and University of New South Wales, Sydney, Australia;NICTA and University of New South Wales, Sydney, Australia;King Abdulaziz University, Saudia Arabia

  • Venue:
  • ACM Computing Surveys (CSUR)
  • Year:
  • 2013

Quantified Score

Hi-index 0.00

Visualization

Abstract

In the last two decades, the continuous increase of computational power has produced an overwhelming flow of data which has called for a paradigm shift in the computing architecture and large-scale data processing mechanisms. MapReduce is a simple and powerful programming model that enables easy development of scalable parallel applications to process vast amounts of data on large clusters of commodity machines. It isolates the application from the details of running a distributed program such as issues on data distribution, scheduling, and fault tolerance. However, the original implementation of the MapReduce framework had some limitations that have been tackled by many research efforts in several followup works after its introduction. This article provides a comprehensive survey for a family of approaches and mechanisms of large-scale data processing mechanisms that have been implemented based on the original idea of the MapReduce framework and are currently gaining a lot of momentum in both research and industrial communities. We also cover a set of introduced systems that have been implemented to provide declarative programming interfaces on top of the MapReduce framework. In addition, we review several large-scale data processing systems that resemble some of the ideas of the MapReduce framework for different purposes and application scenarios. Finally, we discuss some of the future research directions for implementing the next generation of MapReduce-like solutions.