Using Model Counting to Find Optimal Distinguishing Tests

  • Authors:
  • Stefan Heinz;Martin Sachenbacher

  • Affiliations:
  • Zuse Institute Berlin, Berlin, Germany 14195;Institut für Informatik, Technische Universität München, Garching, Germany 85748

  • Venue:
  • CPAIOR '09 Proceedings of the 6th International Conference on Integration of AI and OR Techniques in Constraint Programming for Combinatorial Optimization Problems
  • Year:
  • 2009

Quantified Score

Hi-index 0.00

Visualization

Abstract

Testing is the process of stimulating a system with inputs in order to reveal hidden parts of the system state. In the case of non-deterministic systems, the difficulty arises that an input pattern can generate several possible outcomes. Some of these outcomes allow to distinguish between different hypotheses about the system state, while others do not. In this paper, we present a novel approach to find, for non-deterministic systems, modeled as constraints over variables, tests that allow to distinguish among the hypotheses as good as possible. The idea is to assess the quality of a test by determining the ratio of distinguishing (good) and not distinguishing (bad) outcomes. This measure refines previous notions proposed in the literature on model-based testing and can be computed using model counting techniques. We propose and analyze a greedy-type algorithm to solve this test optimization problem, using existing model counters as a building block. We give preliminary experimental results of our method, and discuss possible improvements.