Effects of Multithreading on Cache Performance

  • Authors:
  • Hantak Kwak;Ben Lee;Ali R. Hurson;Suk-Han Yoon;Woo-Jong Hahn

  • Affiliations:
  • Oregon State Univ., Corvallis;Oregon State Univ., Corvallis;Pennsylvania State Univ., University Park;Electronics and Telecommunications Research Institute, Taejon, Korea;Electronics and Telecommunications Research Institute, Taejon, Korea

  • Venue:
  • IEEE Transactions on Computers - Special issue on cache memory and related problems
  • Year:
  • 1999

Quantified Score

Hi-index 0.00

Visualization

Abstract

As the performance gap between processor and memory grows, memory latency becomes a major bottleneck in achieving high processor utilization. Multithreading has emerged as one of the most promising and exciting techniques used to tolerate memory latency by exploiting thread-level parallelism. The question, however, remains as to how effective multithreading is on tolerating memory latency. The performance of multithreading is not only affected by the overlapping of memory latency with useful computation, but also strongly depends on the cache behavior and the overhead of multithreading (e.g., thread management and context-switch costs). In particular, multithreading affects the behavior of caches, and, thus, the overall performance in a nontrivial fashion. To study these issues, this paper presents the Multithreaded Virtual Processor (MVP) model. MVP integrates the multithreaded programming paradigm and a modern superscalar processor with support for fast context switching and thread scheduling. Our studies with MVP show that, in general, the performance improvements are obtained not only by tolerating memory latency but also lower cache miss rates due to exploitation of data locality.However, multithreading creates an additional stress on the memory hierarchy caused by the interference among threads. Also, the dynamic behavior of multithreaded execution hinders the instruction locality that results in a high number of misses in the L1 instruction cache.