Exploring the cache design space for large scale CMPs

  • Authors:
  • Lisa Hsu;Ravi Iyer;Srihari Makineni;Steve Reinhardt;Donald Newell

  • Affiliations:
  • University of Michigan, Ann Arbor;Systems Technology Lab, Intel Corporation;Systems Technology Lab, Intel Corporation;University of Michigan, Ann Arbor;Systems Technology Lab, Intel Corporation

  • Venue:
  • ACM SIGARCH Computer Architecture News - Special issue: dasCMP'05
  • Year:
  • 2005

Quantified Score

Hi-index 0.00

Visualization

Abstract

With the advent of dual-core chips in the marketplace, small-scale CMP (chip multiprocessor) architectures are becoming commonplace. We expect a continuing trend of increasing the number of cores on a die to maximize the performance/power efficiency of a single chip. We believe an era of large-scale CMPs (LCMPs) with several tens to hundreds of cores is on the way, but as of now architects have little understanding of how best to build a cache hierarchy given such a large number of cores/threads to support. With this in mind, our initial goals are to prune the cache design space for LCMPs by characterizing basic server workload behavior in such an environment.In this paper, we describe the range of methodologies that we are developing to overcome the challenges of exploring the cache design space for LCMP platforms. We then focus on employing a trace-driven approach to characterizing one key server workload (OLTP) in both a homogeneous and a heterogeneous workload environment. We study the effect of increasing threads (from 1 to 128) on a three-level cache hierarchy with emphasis on second and third level caches. We study the effect of varying sizes at these cache levels and show the effects of threads contending for cache space, the effects of prefetching instruction addresses, and the effects of inclusion. We make initial observations and conclusions about the factors on which LCMP cache hierarchy design decisions should be based and discuss future work.