An industrial case study on quality impact prediction for evolving service-oriented software
Proceedings of the 33rd International Conference on Software Engineering
Software testing-resource allocation with operational profile
Proceedings of the 27th Annual ACM Symposium on Applied Computing
Testing techniques selection based on ODC fault types and software metrics
Journal of Systems and Software
Combining model checking and testing with an application to reliability prediction and distribution
Proceedings of the 2013 International Symposium on Software Testing and Analysis
Performance and reliability prediction for evolving service-oriented software systems
Empirical Software Engineering
Hi-index | 0.00 |
With software systems increasingly being employed in critical contexts, assuring high reliability levels for large, complex systems can incur huge verification costs. Existing standards usually assign predefined risk levels to components in the design phase, to provide some guidelines for the verification. It is a rough-grained assignment that does not consider the costs and does not provide sufficient modeling basis to let engineers quantitatively optimize resources usage. Software reliability allocation models partially address such issues, but they usually make so many assumptions on the input parameters that their application is difficult in practice. In this paper, we try to reduce this gap, proposing a reliability and testing resources allocation model that is able to provide solutions at various levels of detail, depending upon the information the engineer has about the system. The model aims to quantitatively identify the most critical components of software architecture in order to best assign the testing resources to them. A tool for the solution of the model is also developed. The model is applied to an empirical case study, a program developed for the European Space Agency, to verify model's prediction abilities and evaluate the impact of the parameter estimation errors on the prediction accuracy.