Using simple page placement policies to reduce the cost of cache fills in coherent shared-memory systems

  • Authors:
  • Michael Marchetti;Leonidas I. Kontothanassis;Ricardo Bianchini;Michael L. Scott

  • Affiliations:
  • -;-;-;-

  • Venue:
  • IPPS '95 Proceedings of the 9th International Symposium on Parallel Processing
  • Year:
  • 1995

Quantified Score

Hi-index 0.00

Visualization

Abstract

The cost of a cache miss depends heavily on the location of the main memory that backs the missing line. For certain applications, this cost is a major factor in overall performance. We report on the utility of OS-based page placement as a mechanism to increase the frequency with which cache fills access local memory in distributed shared memory multiprocessors. Even with the very simple policy of first-use placement, we find significant improvements over round-robin placement for many applications on both hardware- and software-coherent systems. For most of our applications, first-use placement allows 35 to 75 percent of cache fills to be performed locally, resulting in performance improvements of up to 40 percent with respect to round-robin placement. We were surprised to find no performance advantage in more sophisticated policies, including page migration and page replication. In fact, in many cases the performance of our applications suffered under these policies.