Software errors and complexity: an empirical investigation0
Communications of the ACM
The Detection of Fault-Prone Programs
IEEE Transactions on Software Engineering
Predicting Fault-Prone Software Modules in Telephone Switches
IEEE Transactions on Software Engineering
Predicting Fault Incidence Using Software Change History
IEEE Transactions on Software Engineering
Does Code Decay? Assessing the Evidence from Change Management Data
IEEE Transactions on Software Engineering
The distribution of faults in a large industrial software system
ISSTA '02 Proceedings of the 2002 ACM SIGSOFT international symposium on Software testing and analysis
An empirical evaluation of fault-proneness models
Proceedings of the 24th International Conference on Software Engineering
Reexamining the Fault Density-Component Size Connection
IEEE Software
A Metrics Suite for Object Oriented Design
IEEE Transactions on Software Engineering
Quantitative Analysis of Faults and Failures in a Complex Software System
IEEE Transactions on Software Engineering
An Empirical Analysis of Fault Persistence Through Software Releases
ISESE '03 Proceedings of the 2003 International Symposium on Empirical Software Engineering
ISSTA '04 Proceedings of the 2004 ACM SIGSOFT international symposium on Software testing and analysis
Robust Prediction of Fault-Proneness by Random Forests
ISSRE '04 Proceedings of the 15th International Symposium on Software Reliability Engineering
Predicting the Location and Number of Faults in Large Software Systems
IEEE Transactions on Software Engineering
Looking for bugs in all the right places
Proceedings of the 2006 international symposium on Software testing and analysis
Predicting fault-prone components in a java legacy system
Proceedings of the 2006 ACM/IEEE international symposium on Empirical software engineering
Proceedings of the 30th international conference on Software engineering
Exploring the relationship of history characteristics and defect count: an empirical study
DEFECTS '08 Proceedings of the 2008 workshop on Defects in large software systems
Can developer-module networks predict failures?
Proceedings of the 16th ACM SIGSOFT International Symposium on Foundations of software engineering
Predicting failures with developer networks and social network analysis
Proceedings of the 16th ACM SIGSOFT International Symposium on Foundations of software engineering
Misclassification cost-sensitive fault prediction models
PROMISE '09 Proceedings of the 5th International Conference on Predictor Models in Software Engineering
Journal of Systems and Software
Information and Software Technology
The usual suspects: a case study on delivered defects per developer
Proceedings of the 2010 ACM-IEEE International Symposium on Empirical Software Engineering and Measurement
Towards a software failure cost impact model for the customer: an analysis of an open source product
Proceedings of the 6th International Conference on Predictive Models in Software Engineering
Programmer-based fault prediction
Proceedings of the 6th International Conference on Predictive Models in Software Engineering
A framework for defect prediction in specific software project contexts
CEE-SET'08 Proceedings of the Third IFIP TC 2 Central and East European conference on Software engineering techniques
Factors characterizing reopened issues: a case study
Proceedings of the 8th International Conference on Predictive Models in Software Engineering
Studying volatility predictors in open source software
Proceedings of the ACM-IEEE international symposium on Empirical software engineering and measurement
Influence of confirmation biases of developers on software quality: an empirical study
Software Quality Control
An algorithmic approach to missing data problem in modeling human aspects in software development
Proceedings of the 9th International Conference on Predictive Models in Software Engineering
Hi-index | 0.00 |
We have been investigating different prediction models to identify which files of a large multi-release industrial software system are most likely to contain the largest numbers of faults in the next release. To make predictions we considered a number of different file characteristics and change information about the files, and have built fully-automatable models that do not require that the user have any statistical expertise. We now consider the effect of adding developer information as a prediction factor and assess the extent to which this affects the quality of the predictions.