Citation
Asim Shahid, Muhammad and Alam, Muhammad Mansoor and Mohd Su'ud, Mazliham (2024) A fact based analysis of decision trees for improving reliability in cloud computing. PLOS ONE, 19 (12). e0311089. ISSN 1932-6203
Text
journal.pone.0311089.pdf - Published Version Restricted to Repository staff only Download (7MB) |
Abstract
The popularity of cloud computing (CC) has increased significantly in recent years due to its cost-effectiveness and simplified resource allocation. Owing to the exponential rise of cloud computing in the past decade, many corporations and businesses have moved to the cloud to ensure accessibility, scalability, and transparency. The proposed research involves comparing the accuracy and fault prediction of five machine learning algorithms: AdaBoostM1, Bagging, Decision Tree (J48), Deep Learning (Dl4jMLP), and Naive Bayes Tree (NB Tree). The results from secondary data analysis indicate that the Central Processing Unit CPU�Mem Multi classifier has the highest accuracy percentage and the least amount of fault pre�diction. This holds for the Decision Tree (J48) classifier with an accuracy rate of 89.71% for 80/20, 90.28% for 70/30, and 92.82% for 10-fold cross-validation. Additionally, the Hard Disk Drive HDD-Mono classifier has an accuracy rate of 90.35% for 80/20, 92.35% for 70/ 30, and 90.49% for 10-fold cross-validation. The AdaBoostM1 classifier was found to have the highest accuracy percentage and the least amount of fault prediction for the HDD Multi classifier with an accuracy rate of 93.63% for 80/20, 90.09% for 70/30, and 88.92% for 10- fold cross-validation. Finally, the CPU-Mem Mono classifier has an accuracy rate of 77.87% for 80/20, 77.01% for 70/30, and 77.06% for 10-fold cross-validation. Based on the primary data results, the Naive Bayes Tree (NB Tree) classifier is found to have the highest accuracy rate with less fault prediction of 97.05% for 80/20, 96.09% for 70/30, and 96.78% for 10 folds cross-validation. However, the algorithm complexity is not good, taking 1.01 seconds. On the other hand, the Decision Tree (J48) has the second-highest accuracy rate of 96.78%, 95.95%, and 96.78% for 80/20, 70/30, and 10-fold cross-validation, respectively. J48 also has less fault prediction but with a good algorithm complexity of 0.11 seconds. The difference in accuracy and less fault prediction between NB Tree and J48 is only 0.9%, but the difference in time complexity is 9 seconds. Based on the results, we have decided to make modifications to the Decision Tree (J48) algorithm. This method has been proposed as it offers the highest accuracy and less fault prediction errors, with 97.05% accuracy for the 80/ 20 split, 96.42% for the 70/30 split, and 97.07% for the 10-fold cross-validation.
Item Type: | Article |
---|---|
Subjects: | Q Science > QA Mathematics |
Divisions: | Faculty of Computing and Informatics (FCI) |
Depositing User: | Ms Nurul Iqtiani Ahmad |
Date Deposited: | 03 Jan 2025 01:30 |
Last Modified: | 03 Jan 2025 01:30 |
URII: | http://shdl.mmu.edu.my/id/eprint/13260 |
Downloads
Downloads per month over past year
Edit (login required) |