Sound empirical evidence in software testing
Proceedings of the 34th International Conference on Software Engineering
Pairwise testing for software product lines: comparison of two approaches
Software Quality Control
Achieving scalable model-based testing through test case diversity
ACM Transactions on Software Engineering and Methodology (TOSEM)
An orchestrated survey of methodologies for automated software test case generation
Journal of Systems and Software
Test case selection for black-box regression testing of database applications
Information and Software Technology
Static test case prioritization using topic models
Empirical Software Engineering
Hi-index | 0.00 |
Our experience with applying model-based testing on industrial systems showed that the generated test suites are often too large and costly to execute given project deadlines and the limited resources for system testing on real platforms. In such industrial contexts, it is often the case that only a small subset of test cases can be run. In previous work, we proposed novel test case selection techniques that minimize the similarities among selected test cases and outperforms other selection alternatives. In this paper, our goal is to gain insights into why and under which conditions similarity-based selection techniques, and in particular our approach, can be expected to work. We investigate the properties of test suites with respect to similarities among fault revealing test cases. We thus identify the ideal situation in which a similarity-based selection works best, which is useful for devising more effective similarity functions. We also address the specific situation in which a test suite contains outliers, that is a small group of very different test cases, and show that it decreases the effectiveness of similarity-based selection. We then propose, and successfully evaluate based on two industrial systems, a solution based on rank scaling to alleviate this problem.