Main Memory-Based Algorithms for Efficient Parallel Aggregation for Temporal Databases

  • Authors:
  • Dengfeng Gao;Jose Alvin G. Gendrano;Bongki Moon;Richard Thomas Snodgrass;Minseok Park;Bruce C. Huang;Jim M. Rodrigue

  • Affiliations:
  • Computer Science Department, University of Arizona, Tuscon, AZ 85721-0077, USA;Computer Science Department, University of Arizona, Tuscon, AZ 85721-0077, USA;Computer Science Department, University of Arizona, Tuscon, AZ 85721-0077, USA;Computer Science Department, University of Arizona, Tuscon, AZ 85721-0077, USA. rts@cs.arizona.edu;Computer Science Department, University of Arizona, Tuscon, AZ 85721-0077, USA;Computer Science Department, University of Arizona, Tuscon, AZ 85721-0077, USA;Computer Science Department, University of Arizona, Tuscon, AZ 85721-0077, USA

  • Venue:
  • Distributed and Parallel Databases
  • Year:
  • 2004

Quantified Score

Hi-index 0.00

Visualization

Abstract

The ability to model the temporal dimension is essential to many applications. Furthermore, the rate of increase in database size and stringency of response time requirements has out-paced advancements in processor and mass storage technology, leading to the need for parallel temporal database management systems. In this paper, we introduce a variety of parallel temporal aggregation algorithms for the shared-nothing architecture; these algorithms are based on the sequential Aggregation Tree algorithm. We are particularly interested in developing parallel algorithms that can maximally exploit available memory to quickly compute large-scale temporal aggregates without intermediate disk writes and reads. Via an empirical study, we found that the number of processing nodes, the partitioning of the data, the placement of results, and the degree of data reduction effected by the aggregation impacted the performance of the algorithms. For distributed result placement, we discovered that Greedy Time Division Merge was the obvious choice. For centralized results and high data reduction, Pairwise Merge was preferred for a large number of processing nodes; for low data reduction, it only performed well up to 32 nodes. This led us to a centralized variant of Greedy Time Division Merge which was best for the remaining cases. We present a cost model that closely predicts the running time of Greedy Time Division Merge.