Multiprocessor cache design considerations

  • Authors:
  • R. L. Lee;P. C. Yew;D. H. Lawrie

  • Affiliations:
  • Center for Supercomputing Research and Development, University of Illinois at Urbana-Champaign, Urbana, IL;Center for Supercomputing Research and Development, University of Illinois at Urbana-Champaign, Urbana, IL;Center for Supercomputing Research and Development, University of Illinois at Urbana-Champaign, Urbana, IL

  • Venue:
  • ISCA '87 Proceedings of the 14th annual international symposium on Computer architecture
  • Year:
  • 1987

Quantified Score

Hi-index 0.01

Visualization

Abstract

In this paper, cache design is explored for large high-performance multiprocessors with hundreds or thousands of processors and memory modules interconnected by a pipe-lined multi-stage network. The majority of the multiprocessor cache studies in the literature exclusively focus on the issue of cache coherence enforcement. However, there are other characteristics unique to such multiprocessors which create an environment for cache performance that is very different from that of many uniprocessors.Multiprocessor conditions are identified and modeled, including, 1) the cost of a cache coherence enforcement scheme, 2) the effect of a high degree of overlap between cache miss services, 3) the cost of a pin limited data path between shared memory and caches, 4) the effect of a high degree of data prefetching, 5) the program behavior of a scientific workload as represented by 23 numerical subroutines, and 6) the parallel execution of programs. This model is used to show that the cache miss ratio is not a suitable performance measure in the multiprocessors of interest and to show that the optimal cache block size in such multiprocessors is much smaller than in many uniprocessors.