Trifecta: a nonspeculative scheme to exploit common, data-dependent subcritical paths

  • Authors:
  • Patrick Ndai;Nauman Rafique;Mithuna Thottethodi;Swaroop Ghosh;Swarup Bhunia;Kaushik Roy

  • Affiliations:
  • Electrical and Computer Engineering Department, Purdue University, West Lafayette, IN;Google, San Francisco, CA and Electrical and Computer Engineering Department, Purdue University, West Lafayette, IN;Electrical and Computer Engineering Department, Purdue University, West Lafayette, IN;Electrical and Computer Engineering Department, Purdue University, West Lafayette, IN;Electrical and Computer Engineering Department, Case Western Reserve University, Cleveland, OH;Electrical and Computer Engineering Department, Purdue University, West Lafayette, IN

  • Venue:
  • IEEE Transactions on Very Large Scale Integration (VLSI) Systems
  • Year:
  • 2010

Quantified Score

Hi-index 0.00

Visualization

Abstract

Pipelined processor cores are conventionally designed to accommodate the critical paths in the critical pipeline stage(s) in a single clock cycle, to ensure correctness. Such conservative design is wasteful in many cases since critical paths are rarely exercised. Thus, configuring the pipeline to operate correctly for rarely used critical paths targets the uncommon case instead of optimizing for the common case. In this study, we describe Trifecta--an architectural technique that completes common-case, subcritical path operations in a single cycle but uses two cycles when the critical path is exercised. This increases slack for both single- and two-cycle operations and offers a unique advantage under process variation. In contrast with existing mechanisms that trade power or performance for yield, Trifecta improves the yield while preserving performance and power. We applied this technique to the critical pipeline stages of a superscalar out-of-order (OoO) and a single issue in-order processor, namely instruction issue and execute, respectively. Our experiments show that the rare two-cycle operations result in a small decrease (5% for integer and 2% for floating-point benchmarks of SPEC2000) in instructions per cycle. However, the increased delay slack causes an improvement in yield-adjusted-throughput by 20% (12.7%) for an in-order (InO) processor configuration.