An Experiment Family to Investigate the Defect Detection Effect of Tool-Support for Requirements Inspection

  • Authors:
  • Michael Halling;Stefan Biffl;Paul Grünbacher

  • Affiliations:
  • -;-;-

  • Venue:
  • METRICS '03 Proceedings of the 9th International Symposium on Software Metrics
  • Year:
  • 2003

Quantified Score

Hi-index 0.00

Visualization

Abstract

The inspection of software products can help to find defects early in the development process and to gather valuable information on product quality. An inspection is rather resource-intensive and involves several tedious tasks like navigating, sorting, or checking. Tool support is thus hoped to increase effectiveness and efficiency. However, little empirical work is available that directly compares paper-based (i.e., manual) and tool-based software inspections. Existing reports on tool support for inspection generally tend to focus on code inspections while little can be found on requirements or design inspection. This paper reports on an experiment family: two experiments on paper-based inspection and a third experiment to empirically investigate the effect of tool support regarding defect detection effectiveness and inspection effort in an academic environment with 40 subjects. Main results of the experiment family are: (a) The effectiveness is similar for manual and tool-supported inspections; (b) the inspection effort and defect overlap decreased significantly with tool support, while (c) efficiency increased considerably with tool support.