Micro-architecture evaluation using performance vectors

  • Authors:
  • Umesh Krishnaswamy;Isaac D. Scherson

  • Affiliations:
  • Dept. of Information & Computer Science, University of California, Irvine, CA;Dept. of Information & Computer Science, University of California, Irvine, CA

  • Venue:
  • Proceedings of the 1996 ACM SIGMETRICS international conference on Measurement and modeling of computer systems
  • Year:
  • 1996

Quantified Score

Hi-index 0.00

Visualization

Abstract

Benchmarking is a widely used approach to measure computer performance. Current use of benchmarks only provides running times to describe the performance of a tested system. Glancing through these execution times provides little or no information about system strengths and weaknesses. A novel benchmarking methodology is proposed to identify key performance parameters; the methodology is based on measuring performance vectors. A performance vector is a vector of ratings that represents delivered performance of primitive operations of a system. Measuring the performance vector of a system in a typical user workload can be a tough problem. We show how the performance vector falls out of an equation consisting of dynamic instruction counts and execution times of benchmarks. We present a non-linear approach for computing the performance vector. The efficacy of the methodology is ascertained by evaluating the micro-architecture of the Sun SuperSPARC superscalar processor using SPEC benchmarks. Results show interesting tradeoffs in the SuperSPARC and speak favorably of our methodology.