Impact of Admission and Cache Replacement Policies on Response Times of Jobs on Data Grids

  • Authors:
  • Ekow Otoo;Doron Rotem;Arie Shoshani

  • Affiliations:
  • Lawrence Berkeley National Laboratory, University of California, Berkeley 94720;Lawrence Berkeley National Laboratory, University of California, Berkeley 94720;Lawrence Berkeley National Laboratory, University of California, Berkeley 94720

  • Venue:
  • Cluster Computing
  • Year:
  • 2005

Quantified Score

Hi-index 0.00

Visualization

Abstract

Caching techniques have been used widely to improve the performance gaps of storage hierarchies in computing systems. Little is known about the impact of policies on the response times of jobs that access and process very large files in data grids, particularly when data and computations on the data have to be co-located on the same host. In data intensive applications that access large data files over wide area network environment, such as data-grids, the combination of policies for job servicing (or scheduling), caching and cache replacement can significantly impact the performance of grid jobs. We present preliminary results of a simulation study that combines an admission policy with a cache replacement policy when servicing jobs submitted to a storage resource manager.The results show that, in comparison to a first come first serve policy, the response times of jobs are significantly improved, for practical limits of disk cache sizes, when the jobs that are back-logged to access the same files are taken into consideration in scheduling the next file to be retrieved into the disk cache. Not only are the response times of jobs improved, but also the metric measures for caching policies, such as the hit ratio and the average cost per retrieval, are improved irrespective of the cache replacement policy used.