Simultaneous continual flow pipeline architecture

  • Authors:
  • Komal Jothi;Mageda Sharafeddine;Haitham Akkary

  • Affiliations:
  • Department of Electrical and Computer Engineering, American University of Beirut, Lebanon;Department of Electrical and Computer Engineering, American University of Beirut, Lebanon;Department of Electrical and Computer Engineering, American University of Beirut, Lebanon

  • Venue:
  • ICCD '11 Proceedings of the 2011 IEEE 29th International Conference on Computer Design
  • Year:
  • 2011

Quantified Score

Hi-index 0.00

Visualization

Abstract

Since the introduction of the first industrial out-of-order superscalar processors in the 1990s, instruction buffers and cache sizes have kept increasing with every new generation of out-of-order cores. The motivation behind this continuous evolution has been performance of single-thread applications. Performance gains from larger instruction buffers and caches come at the expense of area, power, and complexity. We show that this is not the most energy efficient way to achieve performance. Instead, sizing the instruction buffers to the minimum size necessary for the common case of L1 data cache hits and using new latency-tolerant microarchitecture to handle loads that miss the L1 data cache, improves execution time and energy consumption on SpecCPU 2000 benchmarks by an average of 10% and 12% respectively, compared to a large superscalar baseline. Our non-blocking architecture outperforms other latency tolerant architectures, such as Continual Flow Pipelines, by up to 15% on the same SpecCPU 2000 benchmarks.