Precise regression benchmarking with random effects: improving mono benchmark results

  • Authors:
  • Tomas Kalibera;Petr Tuma

  • Affiliations:
  • Distributed Systems Research Group, Department of Software Engineering, Faculty of Mathematics and Physics, Charles University, Prague, Czech Republic;Distributed Systems Research Group, Department of Software Engineering, Faculty of Mathematics and Physics, Charles University, Prague, Czech Republic

  • Venue:
  • EPEW'06 Proceedings of the Third European conference on Formal Methods and Stochastic Models for Performance Evaluation
  • Year:
  • 2006

Quantified Score

Hi-index 0.00

Visualization

Abstract

Benchmarking as a method of assessing software performance is known to suffer from random fluctuations that distort the observed performance. In this paper, we focus on the fluctuations caused by compilation. We show that the design of a benchmarking experiment must reflect the existence of the fluctuations if the performance observed during the experiment is to be representative of reality We present a new statistical model of a benchmark experiment that reflects the presence of the fluctuations in compilation, execution and measurement. The model describes the observed performance and makes it possible to calculate the optimum dimensions of the experiment that yield the best precision within a given amount of time Using a variety of benchmarks, we evaluate the model within the context of regression benchmarking. We show that the model significantly decreases the number of erroneously detected performance changes in regression benchmarking