Please use this identifier to cite or link to this item: https://elib.vku.udn.vn/handle/123456789/5786
Full metadata record
DC FieldValueLanguage
dc.contributor.authorDang, Thi Kim Ngan-
dc.contributor.authorDao, Khanh Duy-
dc.contributor.authorHa, Thi Minh Phuong-
dc.contributor.authorNguyen, Thanh Binh-
dc.date.accessioned2025-11-11T04:13:35Z-
dc.date.available2025-11-11T04:13:35Z-
dc.date.issued2025-06-
dc.identifier.issn1859-3526-
dc.identifier.urihttps://ictmag.ictvietnam.vn/cntt-tt/article/view/1374-
dc.identifier.urihttps://elib.vku.udn.vn/handle/123456789/5786-
dc.descriptionResearch, Development and Application on Information and Communication Technology; Tập 2025, số 1, tháng 6vi_VN
dc.description.abstractSoftware fault prediction (SFP) involves the identification of potentially fault-prone modules before the testing phase in the software development lifecycle. By predicting faults early in the development process, the SFP process enables software developers to focus their efforts on components that may contain faults, thereby enhancing the overall quality and reliability of the software. Machine learning and deep learning techniques have been widely applied to train SFP models. However, these approaches face several challenges, including irrelevant or redundant features, imbalanced datasets, overfitting, and complex model structures. The NASA dataset from the PROMISE repository is the most commonly used dataset for fault prediction. Recently, the BugHunter dataset with its substantially larger number of instances was explored to train the SFP models. In this study, we present the comparative study of three deep learning models, including Convolutional Neural Network (CNN), Recurrent Neural Network (RNN), Long Short-Term Memory (LSTM) and four machine learning models as K-Nearest Neighbors (KNN), Multilayer Perceptron (MLP), Adaptive Boosting (AdaBoost), Extreme Gradient Boosting (XGB) to investigate the performance of SFP models on the BugHunter dataset. We employ the Lasso method for feature selection and apply the Synthetic Minority Oversampling Technique (SMOTE) to address the issue of imbalanced data, aiming to enhance the accuracy of the results. The experimental findings reveal that CNN and RNN outperformed other machine learning models, achieving the best overall performance.vi_VN
dc.language.isoenvi_VN
dc.publisherThe Journal of Information and Communicationvi_VN
dc.subjectSoftware fault predictionvi_VN
dc.subjectmachine learningvi_VN
dc.subjectBugHunter datasetvi_VN
dc.titlePerformance Analysis of Deep Learning Models for Software Fault Prediction Using the BugHunter Datasetvi_VN
dc.title.alternativeĐánh giá các mô hình học sâu cho bài toán dự đoán lỗi phần mềm trên bộ dữ liệu BugHuntervi_VN
dc.typeWorking Papervi_VN
Appears in Collections:NĂM 2025

Files in This Item:

 Sign in to read



Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.