Cache Design for Eliminating the Address Translation Bottleneck and Reducing the Tag Area Cost

  • Authors:
  • Affiliations:
  • Venue:
  • ICCD '02 Proceedings of the 2002 IEEE International Conference on Computer Design: VLSI in Computers and Processors (ICCD'02)
  • Year:
  • 2002

Quantified Score

Hi-index 0.00

Visualization

Abstract

For the physical caches, the address translation delay can be partially masked, but it is hard to avoid completely. In this paper, we propose a cache partition architecture, called paged cache, which can not only mask the address translation delay completely but also reduce the tag area dramatically. In the paged cache, we divide the entire cache into a set of partitions, and each partition is dedicated to only one page cached in the TLB. By restricting the range in which the cached block can be placed, we can eliminate the total or partial tag depending on the partition size. In addition, because the paged cache can be accessed without waiting for the generation of physical address, i.e., the paged cache and the TLB are accessed in parallel, the extended cache access time can be reduced largely. We use SimpleScalar to simulate the SPEC2000 benchmarks and perform the HSPICE simulations (with a 0.18µm technology and 1.8V voltage supply) to evaluate the proposed architecture. Experimental results show that the paged cache is very effective in reducing tag area of the on-chip L1 caches, while the average extended cache access time can be improved dramatically.