Assessing test adequacy for black-box systems without specifications

  • Authors:
  • Neil Walkinshaw

  • Affiliations:
  • Department of Computer Science, The University of Leicester

  • Venue:
  • ICTSS'11 Proceedings of the 23rd IFIP WG 6.1 international conference on Testing software and systems
  • Year:
  • 2011
  • Model-Based testing and model inference

    ISoLA'12 Proceedings of the 5th international conference on Leveraging Applications of Formal Methods, Verification and Validation: technologies for mastering change - Volume Part I

Quantified Score

Hi-index 0.00

Visualization

Abstract

Testing a black-box system without recourse to a specification is difficult, because there is no basis for estimating how many tests will be required, or to assess how complete a given test set is. Several researchers have noted that there is a duality between these testing problems and the problem of inductive inference (learning a model of a hidden system from a given set of examples). It is impossible to tell how many examples will be required to infer an accurate model, and there is no basis for telling how complete a given set of examples is. These issues have been addressed in the domain of inductive inference by developing statistical techniques, where the accuracy of an inferred model is subject to a tolerable degree of error. This paper explores the application of these techniques to assess test sets of black-box systems. It shows how they can be used to reason in a statistically justified manner about the number of tests required to fully exercise a system without a specification, and how to provide a valid adequacy measure for black-box test sets in an applied context.