Optimizing Decision Tree Hyperparameters via Random Search for Accurate Heart Failure Risk Prediction

Authors

DOI:

https://doi.org/10.54082/jiki.312

Keywords:

Decision Tree, Heart Failure Prediction, Hyperparameter Optimization, Machine Learning, Random Search

Abstract

Heart failure remains one of the leading causes of mortality worldwide, highlighting the need for reliable early-detection models to support clinical decision-making. This study investigates the effect of Random Search–based hyperparameter optimization on a Decision Tree model for heart failure risk prediction using a clinical dataset comprising 918 samples and 11 demographic and cardiovascular features. Rather than introducing a novel optimization algorithm, this work focuses on analyzing model performance sensitivity to hyperparameter tuning in a real-world medical dataset. The baseline Decision Tree achieved an accuracy of 0.80. After Random Search optimization, accuracy improved to 0.84, while recall for the positive class increased from 0.83 to 0.90, indicating a notable reduction in false-negative predictions. The optimized configuration, characterized by a shallow tree depth and increased minimum samples per leaf, suggests improved generalization and reduced overfitting. Compared with related studies employing ensemble-based models and genetic optimization, the proposed approach achieves competitive performance using a simpler and more interpretable classifier. These findings demonstrate that systematic hyperparameter tuning can substantially enhance the clinical utility of conventional machine learning models. Practically, the improved recall supports the use of the optimized Decision Tree as a screening-oriented decision support tool, enabling earlier identification of high-risk patients while maintaining model transparency. This study highlights the importance of dataset-specific optimization and provides a foundation for future work involving ensemble methods and advanced optimization strategies to develop robust and clinically applicable heart failure prediction systems.

References

N. L. K. A. Arsani, N. P. D. S. Wahyuni, N. N. M. Agustini, and M. Budiawan, “Deteksi dini dan pencegahan penyakit kardiovaskular,” Proceeding Senadimas Undiksha, vol. 1, no. 1, pp. 663–668, 2022.

A. Čartolovni, A. Tomičić, and E. L. Mosler, “Ethical, legal, and social considerations of AI-based medical decision-support tools: A scoping review,” International Journal of Medical Informatics, vol. 161, Art. no. 104738, 2022.

Q. Xu, W. Xie, B. Liao, C. Hu, L. Qin, Z. Yang, et al., “Interpretability of clinical decision support systems based on artificial intelligence from technological and medical perspective: A systematic review,” Journal of Healthcare Engineering, vol. 2023, no. 1, Art. no. 9919269, 2023.

Q. An, S. Rahman, J. Zhou, and J. J. Kang, “A comprehensive review on machine learning in healthcare industry: classification, restrictions, opportunities and challenges,” Sensors, vol. 23, no. 9, Art. no. 4178, 2023.

Badawy, M., Ramadan, N., & Hefny, H. A. (2023). Healthcare predictive analytics using machine learning and deep learning techniques: a survey. Journal of Electrical Systems and Information Technology, 10(1), 40.

M. S. Gangadhar, K. V. S. Sai, S. H. S. Kumar, K. A. Kumar, M. Kavitha, and S. S. Aravinth, “Machine learning and deep learning techniques on accurate risk prediction of coronary heart disease,” in Proc. 7th Int. Conf. Computing Methodologies and Communication (ICCMC), Erode, India, 2023, pp. 227–232, doi: 10.1109/ICCMC56507.2023.10083756.

M. T. García-Ordás, M. Bayón-Gutiérrez, C. Benavides, et al., “Heart disease risk prediction using deep learning techniques with feature augmentation,” Multimedia Tools and Applications, vol. 82, pp. 31759–31773, 2023, doi: 10.1007/s11042-023-14817-z.

M. Ozcan and S. Peker, “A classification and regression tree algorithm for heart disease modeling and prediction,” Healthcare Analytics, vol. 3, p. 100130, 2023, doi: 10.1016/j.health.2022.100130.

S. Dalal, P. Goel, E. M. Onyema, A. Alharbi, A. Mahmoud, M. A. Algarni, and H. Awal, “Application of machine learning for cardiovascular disease risk prediction,” Computational Intelligence and Neuroscience, Art. no. 9418666, 12 pp., 2023, doi: 10.1155/2023/9418666.

R. Azzaz, M. Jahazi, S. E. Kahou, and E. Moosavi-Khoonsari, “Prediction of final phosphorus content of steel in a scrap-based electric arc furnace using artificial neural networks,” Metals, vol. 15, no. 1, Art. no. 62, 2025, doi: 10.3390/met15010062.

RL, M., & Mishra, A. K. (2022). Measuring financial performance of Indian manufacturing firms: application of decision tree algorithms. Measuring Business Excellence, 26(3), 288-307.

Soni, T., Gupta, D., & Uppal, M. (2024, November). Optimizing Heart Disease Prediction with Random Forest: Insights from the Kaggle Dataset. In 2024 4th International Conference on Advancement in Electronics & Communication Engineering (AECE) (pp. 741-744). IEEE.

Noh, Y. D., & Cho, K. C. (2022). Heart Disease Prediction Using Decision Tree With Kaggle Dataset. Journal of The Korea Society of Computer and Information, 27(5), 21-28.

T. Wongvorachan, S. He, and O. Bulut, “A comparison of undersampling, oversampling, and SMOTE methods for dealing with imbalanced classification in educational data mining,” Information, vol. 14, no. 1, Art. no. 54, 2023.

M. Salmi, D. Atif, D. Oliva, A. Abraham, and S. Ventura, “Handling imbalanced medical datasets: Review of a decade of research,” Artificial Intelligence Review, vol. 57, no. 10, Art. no. 273, 2024.

D. Rajput, W. J. Wang, and C. C. Chen, “Evaluation of a decided sample size in machine learning applications,” BMC Bioinformatics, vol. 24, no. 1, Art. no. 48, 2023.

E. Widad, E. Saida, and Y. Gahi, “Quality anomaly detection using predictive techniques: An extensive big data quality framework for reliable data analysis,” IEEE Access, vol. 11, pp. 103306–103318, 2023.

G. R. Hemanth and S. C. Raja, “Proposing suitable data imputation methods by adopting a stage-wise approach for various classes of smart meters missing data—Practical approach,” Expert Systems with Applications, vol. 187, Art. no. 115911, 2022.

V. R. Joseph and A. Vakayil, “SPlit: An optimal method for data splitting,” Technometrics, vol. 64, no. 2, pp. 166–176, 2022.

J. S. Kushwah, A. Kumar, S. Patel, R. Soni, A. Gawande, and S. Gupta, “Comparative study of regressor and classifier with decision tree using modern tools,” Materials Today: Proceedings, vol. 56, pp. 3571–3576, 2022.

Tariq, A., Yan, J., Gagnon, A. S., Riaz Khan, M., & Mumtaz, F. (2023). Mapping of cropland, cropping patterns and crop types by combining optical remote sensing images with decision tree classifier and random forest. Geo-Spatial Information Science, 26(3), 302-320.

Colledani, D., Anselmi, P., & Robusto, E. (2023). Machine learning-decision tree classifiers in psychiatric assessment: An application to the diagnosis of major depressive disorder. Psychiatry Research, 322, 115127.

Jo, N., Aghaei, S., Benson, J., Gomez, A., & Vayanos, P. (2023, August). Learning optimal fair decision trees: Trade-offs between interpretability, fairness, and accuracy. In Proceedings of the 2023 AAAI/ACM Conference on AI, Ethics, and Society (pp. 181-192).

Custode, L. L., & Iacca, G. (2023). Evolutionary learning of interpretable decision trees. IEEE Access, 11, 6169-6184.

Liu, Y. (2022). bsnsing: A decision tree induction method based on recursive optimal boolean rule composition. INFORMS Journal on Computing, 34(6), 2908-2929.

Blockeel, H., Devos, L., Frénay, B., Nanfack, G., & Nijssen, S. (2023). Decision trees: from efficient prediction to responsible AI. Frontiers in artificial intelligence, 6, 1124553.

X. Jiang, J. Zhang, X. Shi, and J. Cheng, “Learning the policy for mixed electric platoon control of automated and human-driven vehicles at signalized intersection: A random search approach,” IEEE Transactions on Intelligent Transportation Systems, vol. 24, no. 5, pp. 5131–5143, 2023.

N. Subaşı, “Comprehensive analysis of grid and randomized search on dataset performance,” European Journal of Engineering and Applied Sciences, vol. 7, no. 2, pp. 77–83, 2024.

K. Deligkaris, “Particle swarm optimization and random search for convolutional neural architecture search,” IEEE Access, vol. 12, pp. 91229–91241, 2024.

S. Suyahman and A. Hapsari, “VGG-based feature extraction for classifying traditional batik motifs using machine learning models,” Preservation, Digital Technology & Culture, 2025.

M. Azhari, Z. Situmorang, and R. Rosnelly, “Perbandingan akurasi, recall, dan presisi klasifikasi pada algoritma C4.5, random forest, SVM, dan naive Bayes,” Jurnal Media Informatika Budidarma, vol. 5, no. 2, p. 640, 2021.

P. R. Togatoropa, M. Sianturia, D. Simamoraa, and D. Silaena, “Optimizing random forest using genetic algorithm for heart disease classification,” Machine Learning, vol. 2, no. 3, Art. no. 4, 2022.

Y. Rimal, N. Sharma, and A. Alsadoon, “The accuracy of machine learning models relies on hyperparameter tuning: Student result classification using random forest, randomized search, grid search, Bayesian, genetic, and Optuna algorithms,” Multimedia Tools and Applications, vol. 83, no. 30, pp. 74349–74364, 2024.

Downloads

Published

2026-03-15

How to Cite

Suyahman, S. (2026). Optimizing Decision Tree Hyperparameters via Random Search for Accurate Heart Failure Risk Prediction. Jurnal Ilmu Komputer Dan Informatika, 5(2), 143–154. https://doi.org/10.54082/jiki.312