SOFTWARE DEFECT PREDICTION: ANALYSIS OF CLASS IMBALANCE AND PERFORMANCE STABILITY

dc.contributor.authorBalogun, Abdullateef Oluwagbemiga
dc.contributor.authorBasri, Shuib
dc.contributor.authorSaid, Jadid Abdulkadir
dc.contributor.authorAdeyemo, Victor Ebenezer
dc.contributor.authorImam, Abdullahi Abubakar
dc.contributor.authorBajeh, Amos Orenyi
dc.date.accessioned2020-01-30T10:54:47Z
dc.date.available2020-01-30T10:54:47Z
dc.date.issued2019-12
dc.description.abstractThe performance of prediction models in software defect prediction depends on the quality of datasets used for training such models. Class imbalance is one of data quality problems that affect prediction models. This has drawn the attention of researchers and many approaches have been developed to address this issue. In this study, an extensive empirical study is presented, which evaluates the performance stability of prediction models in SDP. Ten software defect datasets from NASA and PROMISE repositories with varying imbalance ratio (IR) values were used as the original datasets. New datasets are generated from the original datasets using undersampling (Random under Sampling: RUS) and oversampling (Synthetic Minority Oversampling Technique: SMOTE) methods with different IR values. The sampling techniques were based on the equal proportion (100%) of the increment (SMOTE) of minority class label or decrement (RUS) of the majority class label until each dataset is balanced. IR is the ratio of the defective instances to non-defective instances in a dataset. Each newly generated datasets with different IR values based on different sampling techniques were randomized before applying prediction models. Nine standard prediction models were used on the newly generated datasets. The performance of the prediction models was measured using the Area Under Curve (AUC) and Co-efficient of Variation (CV) is used to determine the performance stability. Firstly, experimental results showed that class imbalance had a negative effect on the performance of prediction models and the oversampling method (SMOTE) enhanced the performances of prediction models. Secondly, Oversampling method of balancing datasets is better than using Undersampling methods as the latter had poor performance as a result of the random deletion of useful instances in the datasets. Finally, among the prediction models used in this study, it appeared that Logistic Regression (LR) (RUS: 30.05; SMOTE: 33.51), Naïve Bayes (NB) (RUS: 34.18; SMOTE: 33.05), and Random Forest (RF) (RUS: 29.24; SMOTE: 64.25) with their respective CV values are more stable prediction models and they work well with imbalanced datasets.en_US
dc.identifier.issn1823-4690
dc.identifier.urihttp://hdl.handle.net/123456789/3595
dc.language.isoenen_US
dc.publisherSchool of Engineering, Taylor’s Universityen_US
dc.relation.ispartofseries14;6
dc.subjectSoftware defect predictionen_US
dc.subjectmachine learningen_US
dc.subjectclass imbalanceen_US
dc.subjectData qualityen_US
dc.titleSOFTWARE DEFECT PREDICTION: ANALYSIS OF CLASS IMBALANCE AND PERFORMANCE STABILITYen_US
dc.typeArticleen_US

Files

Original bundle
Now showing 1 - 1 of 1
No Thumbnail Available
Name:
14_6_16.pdf
Size:
617.06 KB
Format:
Adobe Portable Document Format
Description:
License bundle
Now showing 1 - 1 of 1
No Thumbnail Available
Name:
license.txt
Size:
1.69 KB
Format:
Item-specific license agreed upon to submission
Description:

Collections