A shared cache for a chip multi vector processor

  • Authors:
  • Akihiro Musa;Yoshiei Sato;Takashi Soga;Koki Okabe;Ryusuke Egawa;Hiroyuki Takizawa;Hiroaki Kobayashi

  • Affiliations:
  • Tohoku University, Sendai, Japan/ NEC Corporation, Tokyo, Japan;Tohoku University, Sendai, Japan;NEC System technologies, Osaka, Japan;Tohoku University, Sendai, Japan;Tohoku University, Sendai, Japan;Tohoku University, Sendai, Japan;Tohoku University, Sendai, Japan

  • Venue:
  • Proceedings of the 9th workshop on MEmory performance: DEaling with Applications, systems and architecture
  • Year:
  • 2008

Quantified Score

Hi-index 0.00

Visualization

Abstract

This paper discusses the design of a chip multi vector processor (CMVP), especially examining the effects of an on-chip cache when the off-chip memory bandwidth is limited. As chip multiprocessors (CMPs) have become the mainstream in commodity scalar processors, the CMP architecture will be adopted to design of vector processors in the near future for harnessing a large number of transistors on a chip. To keep a higher sustained performance in execution of scientific and engineering applications, a vector processor (core) generally requires the ratio of the memory bandwidth to the arithmetic performance of at least 4 bytes/flop (B/FLOP). However, vector supercomputers have been encountering the memory wall problem due to the limited pin bandwidth. Therefore, we propose an on-chip shared cache to maintain the effective memory bandwidth for a CMVP. We evaluate the performance of the CMVP based on the NEC SX vector architecture using real scientific applications. Especially, we examine the caching effect on the sustained performance when the B/FLOP rate is decreased. The experimental results indicate that an 8 MB on-chip shared cache can improve the performance of a four-core CMVP by 15% to 40%, compared with that without the cache. This is because the shared cache can increase cache hit rates of multi-threads. Here, the shared cache employs a miss status handling registers, which has the potential for accelerating difference schemes in scientific and engineering applications. Moreover, we show that the 2 B/FLOP is enough for the CMVP to achieve a high scalability when the on-chip cache is employed.