Supervised vs unsupervised models: A holistic look at effort-aware just-in-time defect prediction
Effort-aware just-in-time (JIT) defect prediction aims at finding more defective software changes with limited code inspection cost. Traditionally, supervised models have been used; however, they require sufficient labelled training data, which is difficult to obtain, especially for new projects. Re...
Saved in:
Main Authors: | , , |
---|---|
Format: | text |
Language: | English |
Published: |
Institutional Knowledge at Singapore Management University
2017
|
Subjects: | |
Online Access: | https://ink.library.smu.edu.sg/sis_research/3920 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Singapore Management University |
Language: | English |
id |
sg-smu-ink.sis_research-4922 |
---|---|
record_format |
dspace |
spelling |
sg-smu-ink.sis_research-49222018-12-05T02:52:32Z Supervised vs unsupervised models: A holistic look at effort-aware just-in-time defect prediction HUANG, Qiao XIA, Xin LO, David Effort-aware just-in-time (JIT) defect prediction aims at finding more defective software changes with limited code inspection cost. Traditionally, supervised models have been used; however, they require sufficient labelled training data, which is difficult to obtain, especially for new projects. Recently, Yang et al. proposed an unsupervised model (LT) and applied it to projects with rich historical bug data. Interestingly, they reported that, under the same inspection cost (i.e., 20 percent of the total lines of code modified by all changes), it could find more defective changes than a state-of-the-art supervised model (i.e., EALR). This is surprising as supervised models that benefit from historical data are expected to perform better than unsupervised ones. Their finding suggests that previous studies on defect prediction had made a simple problem too complex. Considering the potential high impact of Yang et al.'s work, in this paper, we perform a replication study and present the following new findings: (1) Under the same inspection budget, LT requires developers to inspect a large number of changes necessitating many more context switches. (2) Although LT finds more defective changes, many highly ranked changes are false alarms. These initial false alarms may negatively impact practitioners' patience and confidence. (3) LT does not outperform EALR when the harmonic mean of Recall and Precision (i.e., F1-score) is considered. Aside from highlighting the above findings, we propose a simple but improved supervised model called CBS. When compared with EALR, CBS detects about 15% more defective changes and also significantly improves Precision and F1-score. When compared with LT, CBS achieves similar results in terms of Recall, but it significantly reduces context switches and false alarms before first success. Finally, we also discuss the implications of our findings for practitioners and researchers. 2017-09-01T07:00:00Z text https://ink.library.smu.edu.sg/sis_research/3920 info:doi/10.1109/ICSME.2017.51 Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Predictive models Inspection Measurement Computer bugs Analytical models Feature extraction Software Software Engineering |
institution |
Singapore Management University |
building |
SMU Libraries |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
SMU Libraries |
collection |
InK@SMU |
language |
English |
topic |
Predictive models Inspection Measurement Computer bugs Analytical models Feature extraction Software Software Engineering |
spellingShingle |
Predictive models Inspection Measurement Computer bugs Analytical models Feature extraction Software Software Engineering HUANG, Qiao XIA, Xin LO, David Supervised vs unsupervised models: A holistic look at effort-aware just-in-time defect prediction |
description |
Effort-aware just-in-time (JIT) defect prediction aims at finding more defective software changes with limited code inspection cost. Traditionally, supervised models have been used; however, they require sufficient labelled training data, which is difficult to obtain, especially for new projects. Recently, Yang et al. proposed an unsupervised model (LT) and applied it to projects with rich historical bug data. Interestingly, they reported that, under the same inspection cost (i.e., 20 percent of the total lines of code modified by all changes), it could find more defective changes than a state-of-the-art supervised model (i.e., EALR). This is surprising as supervised models that benefit from historical data are expected to perform better than unsupervised ones. Their finding suggests that previous studies on defect prediction had made a simple problem too complex. Considering the potential high impact of Yang et al.'s work, in this paper, we perform a replication study and present the following new findings: (1) Under the same inspection budget, LT requires developers to inspect a large number of changes necessitating many more context switches. (2) Although LT finds more defective changes, many highly ranked changes are false alarms. These initial false alarms may negatively impact practitioners' patience and confidence. (3) LT does not outperform EALR when the harmonic mean of Recall and Precision (i.e., F1-score) is considered. Aside from highlighting the above findings, we propose a simple but improved supervised model called CBS. When compared with EALR, CBS detects about 15% more defective changes and also significantly improves Precision and F1-score. When compared with LT, CBS achieves similar results in terms of Recall, but it significantly reduces context switches and false alarms before first success. Finally, we also discuss the implications of our findings for practitioners and researchers. |
format |
text |
author |
HUANG, Qiao XIA, Xin LO, David |
author_facet |
HUANG, Qiao XIA, Xin LO, David |
author_sort |
HUANG, Qiao |
title |
Supervised vs unsupervised models: A holistic look at effort-aware just-in-time defect prediction |
title_short |
Supervised vs unsupervised models: A holistic look at effort-aware just-in-time defect prediction |
title_full |
Supervised vs unsupervised models: A holistic look at effort-aware just-in-time defect prediction |
title_fullStr |
Supervised vs unsupervised models: A holistic look at effort-aware just-in-time defect prediction |
title_full_unstemmed |
Supervised vs unsupervised models: A holistic look at effort-aware just-in-time defect prediction |
title_sort |
supervised vs unsupervised models: a holistic look at effort-aware just-in-time defect prediction |
publisher |
Institutional Knowledge at Singapore Management University |
publishDate |
2017 |
url |
https://ink.library.smu.edu.sg/sis_research/3920 |
_version_ |
1770573936223846400 |