Abstract

Most empirical disciplines promote the reuse and sharing of datasets, as it leads to greater possibility of replication. While this is increasingly the case in Empirical Software Engineering, some of the most popular bug-fix datasets are now known to be biased. This raises two significant concerns: first, that sample bias may lead to underperforming prediction models, and second, that the external validity of the studies based on biased datasets may be suspect. This issue has raised considerable consternation in the ESE literature in recent years. However, there is a confounding factor of these datasets that has not been examined carefully: size. Biased datasets are sampling only some of the data that could be sampled, and doing so in a biased fashion; but biased samples could be smaller, or larger. Smaller data sets in general provide less reliable bases for estimating models, and thus could lead to inferior model performance. In this setting, we ask the question, what affects performance more, bias, or size? We conduct a detailed, large-scale meta-analysis, using simulated datasets sampled with bias from a high-quality dataset which is relatively free of bias. Our results suggest that size always matters just as much bias direction, and in fact much more than bias direction when considering information-retrieval measures such as AUCROC and F-score. This indicates that at least for prediction models, even when dealing with sampling bias, simply finding larger samples can sometimes be sufficient. Our analysis also exposes the complexity of the bias issue, and raises further issues to be explored in the future.

Highlights

  • Detailed data on bugs are clearly crucial to empirical studies of software quality

  • Since we are introducing bias into our training sets, which may disturb the relationship between the training set and test set distributions, ridge regression provides insurance that we can be reasonably certain that the bias introduced by multicollinearity is not impacting the quality of our prediction models in the face of bias introduced by our experimental setup

  • This has led to widespread concerns, reported in several papers, that biased datasets would lead to under-performing, even misleading, prediction models of limited practical value

Read more

Summary

Introduction

Detailed data on bugs are clearly crucial to empirical studies of software quality Such data is generally collected in bug-fix datasets. The fix location is provided by a link to commits in the version control system These links identify the source code files involved in a bug report, as well as other details, such as the developer who committed the fix, the date and time, and the lines changed in the corresponding files. This is a rich source of historical data for building software quality prediction models that may yield improved understanding of the factors that affect software quality

Objectives
Methods
Results
Discussion
Conclusion

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.