Next Article in Journal
Catheter Ablation for Atrial Fibrillation in Patients with Heart Failure: Current Evidence and Future Opportunities
Next Article in Special Issue
Mapping the Heartbeat of America with ChatGPT-4: Unpacking the Interplay of Social Vulnerability, Digital Literacy, and Cardiovascular Mortality in County Residency Choices
Previous Article in Journal
Indicators of HSV1 Infection, ECM–Receptor Interaction, and Chromatin Modulation in a Nuclear Family with Schizophrenia
Previous Article in Special Issue
Epigenetic Signatures in Hypertension
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Comparing Machine Learning Classifiers for Predicting Hospital Readmission of Heart Failure Patients in Rwanda

by
Theogene Rizinde
1,*,
Innocent Ngaruye
2 and
Nathan D. Cahill
3
1
College of Business and Economics, University of Rwanda, Kigali 4285, Rwanda
2
College of Science and Technology, University of Rwanda, Kigali 4285, Rwanda
3
School of Mathematics and Statistics, Rochester Institute of Technology, Rochester, NY 14623, USA
*
Author to whom correspondence should be addressed.
J. Pers. Med. 2023, 13(9), 1393; https://doi.org/10.3390/jpm13091393
Submission received: 2 August 2023 / Revised: 22 August 2023 / Accepted: 28 August 2023 / Published: 18 September 2023

Abstract

:
High rates of hospital readmission and the cost of treating heart failure (HF) are significant public health issues globally and in Rwanda. Using machine learning (ML) to predict which patients are at high risk for HF hospital readmission 20 days after their discharge has the potential to improve HF management by enabling early interventions and individualized treatment approaches. In this paper, we compared six different ML models for this task, including multi-layer perceptron (MLP), K-nearest neighbors (KNN), logistic regression (LR), decision trees (DT), random forests (RF), and support vector machines (SVM) with both linear and radial basis kernels. The outputs of the classifiers are compared using performance metrics including the area under the receiver operating characteristic curve (AUC), sensitivity, and specificity. We found that RF outperforms all the remaining models with an AUC of 94% while SVM, MLP, and KNN all yield 88% AUC. In contrast, DT performs poorly, with an AUC value of 57%. Hence, hospitals in Rwanda can benefit from using the RF classifier to determine which HF patients are at high risk of hospital readmission.

1. Introduction

Heart failure (HF) occurs when the heart becomes too weak or stiff to effectively pump blood to meet the body’s needs [1], resulting in a variety of health problems and high medical expenses. HF has a significant global impact, affecting millions of people worldwide despite medical advancements [2,3]. There is a shortage of HF data on certain patient populations [3]: Many studies focus on HF patients in the United States, but these studies might not be representative of HF patients in the rest of the world [4,5,6]. In Africa, HF is still a significant clinical and health concern, often manifesting as an urgent medical condition requiring prolonged hospital stays [7]. Comprehensive data on HF are lacking in Sub-Saharan Africa (SSA), with little information primarily obtained from urban hospitals [8,9,10]. SSA HF cases have high hospitalization rates and significant associated healthcare costs. Like other countries, HF is a serious public health emergency [11] in Rwanda. Non-communicable illnesses such as HF accounted for 34.7% of deaths in Rwanda in 2020. The need to address HF’s socio-economic effects is highlighted by the fact that it accounts for 5% to 10% of adult hospital admissions in SSA, with a similar trend in Rwanda [12].
Several methods for lowering HF-related hospital readmissions have been explored over the past three decades [13,14]. However, there is still a lack of widespread application machine learning (ML)/artificial intelligence (AI) techniques for anticipating readmissions due to HF, particularly in low- and middle-income countries [15]. ML classifiers enable computers to autonomously learn from data, recognize patterns, and predict outcomes from various inputs without explicitly being programmed [14]. The popularity of ML has grown across a variety of industries due to its outstanding ability to quickly analyze large datasets and reach complex conclusions, improving operations, data-driven decision-making, and innovation [16].
ML classifiers have been used in previous research to precisely predict outcomes like hospital readmission for HF and others [17,18]. However, issues with limited electronic health data integration and class imbalance in medical datasets are present in low- and middle-income countries [19]. Due to lack of data unique to specific countries, like health knowledge, cultural norms, and medical facilities, Rwanda lacks reliable and accurate predictive models to be used in clinical practice [17,20,21]. In this study, we use locally gathered comprehensive data on HF in Rwanda, and we explore a variety of ML classifiers to predict HF hospital readmission. We compare prediction performance of multi-layer perceptrons (MLP), logistic regression (LR), decision trees (DT), K-nearest neighbors (KNN), random forests (RF), and support vector machines (SVM). This study also seeks to pinpoint crucial factors influencing hospital readmissions for HF in Rwanda, providing knowledge and skills to enhance the management of HF. Efforts to accurately predict HF hospital readmission, and to identify high-risk HF patients in Rwanda, may improve HF management and result in better patient outcomes and cost savings [22,23].

2. Materials and Methods

This retrospective study collected data from medical records of HF patients who were hospitalized in Rwanda between 1 January 2008 and 31 December 2019. The records were obtained from seven hospitals that were able to treat HF in Rwanda. These include Rwandan Military Hospital (RMH), King Faisal Hospital (KFH), University Teaching Hospital of Butare (CHUB), University Teaching Hospital of Kigali (CHUK), Rwinkwavu Hospital (RWH), Kirehe Hospital (KIH), and Butaro Hospital (BUH). We extracted various features of interest from the patients’ medical records, including age, sex, district of residence, marital status, occupation, resting heart rate, blood pressure, history of hypertension and smoking, heart ultrasound results, risk factors for HF, number of hospitalization days, respiratory rate upon admission, slope, chest pain, cholesterol status, blood sugar, results of electrocardiography at rest, reason for discharge, presence of feces on admission, and past medical and family history.
We utilized Jupyter Notebook as the primary tool for building ML models. We installed Python 3.11.1 and essential packages such as pandas, seaborn, matplotlib, and scikit-learn, which come with built-in libraries and functions, to facilitate data manipulation, visualization, analysis, and construction of machine learning models. Jupyter Notebook was our preferred tool due to its user-friendly interface, advanced data cleaning capabilities, and fast implementation of modeling processes using the Python programming language.
In this study, we compared six ML classifiers, including multi-layer perceptron (MLP), K-nearest neighbors (KNN), logistic regression (LR), decision trees (DT), random forests (RF), and support vector machines (SVM). All classifiers were trained, tested, and compared using performance metrics including the area under the receiver operating characteristic curve (AUC), sensitivity, and specificity. Inputs for each classifier included the following: The MLP classifier contained 3 hidden layers with 128, 64, and 32 neurons in the hidden layers, and it used a sigmoid activation function. The KNN classifier assumed k = 5; the DT classifier used a maximum depth of 8 and a minimum of 20 samples per leaf. For splitting purposes, Gini impurity criteria were used. The RF classifier used a random forest ensemble with 100 trees, with a maximum depth of 10 and 10 samples per leaf. Then, the SVM classifier with a radial basis function kernel used a regularization parameter of 10. For the SVM with a linear kernel, the regularization parameter was 0.1; other hyperparameters were considered at their default values for simplicity.
First, we performed data collection, exploratory data analysis, and preprocessing to clean and prepare the data for further analysis. Here, we dropped all variables with 50% or more null values from the data frame. For most variables with less than 50% null entries, we filled in missing values using the KNN imputer algorithm. For the age variable, we filled in the missing values using the median since this variable appeared to have a uniform distribution. Second, the imbalance in the dataset was handled to ensure that the two classes of HF patients (0 = no hospital readmission; 1 = at least one hospital readmission within 20 days of hospital discharge) were balanced. To address the problem of dataset heterogeneity, we used the Synthetic Minority Oversampling Technique (SMOTE). SMOTE is an oversampling method that involves synthesizing new instances using the current data to oversample underrepresented groups [24,25]. Third, we extracted important features to reduce the dimensionality of the dataset as it contained over 60 features. This step ensured that only relevant features were included in the model development process, leading to improved model accuracy. We then split the dataset into training and testing sets to evaluate the performance of the model. According to the methodology of Dobbin and colleagues, we used 80% of the data for training and 20% for testing [26,27]. Fourth, we standardized the features for training and testing by shifting and scaling the data to have zero mean and unit standard deviation. Lastly, we trained the classifiers with the training set and evaluated using the testing set to determine the precision and accuracy of the models in predicting HF hospital readmission rates in Rwanda.
We used a confusion matrix as the evaluation metric to measure the performance of the algorithms on both the training and testing datasets. Furthermore, the receiver operating characteristic curve (ROC), the area under the ROC curve (AUC), accuracy, precision, recall, and F1-score metrics were utilized to plot, compare, and identify the best-performing model. The results of this evaluation provided the necessary information for drawing conclusions in line with the research objectives.
Though there was no direct contact with the HF patients while collecting HF data but with their respective files, ethical approval was provided by competent authorities including the Ministry of Health of Rwanda and the Rwandan Institutional Review Board and all concerned seven hospitals.

3. Results

3.1. Exploratory Data Analysis and Preprocessing

We loaded the dataset into the Python Jupyter Notebook, a Python environment where we conducted exploratory data analysis, preprocessing, and model building. The data frame shown in Table 1 contains information for the top five records, including 75 features.
The next step was to conduct exploratory data analysis. Upon inspection, we discovered that almost all the features in the data frame were of string data type, and the dataset contained a total of 4085 objects. To facilitate further analysis, visualization, and preprocessing, we converted the variables to numeric types and subdivided the entries with class labels.
One aspect of our analysis focused on the distribution of admitted HF patients’ age as illustrated in Figure 1.
As shown by Figure 1, patients in their late fifties (50 s) to late sixties (60 s) had the highest admission rates, followed by patients between the ages of zero (0) and early twenties. The least-admitted patients due to HF were aged 90 years and above.
Figure 2 show that the collected and pre-processed dataset was imbalanced, with class 0 accounting for 85% of the total dataset with 3469 records.
Training a model on this imbalanced dataset without addressing the imbalance could result in accurate predictions because the model will perform poorly on class 1, leading to inaccurate overall predictions. Therefore, the imbalance was addressed using the SMOTE oversampling algorithm. The results of the target class before and after handling the imbalance are presented in Table 2.

3.2. Important Features Selection

To optimize the prediction performance of our ML models, we need to mitigate overfitting by selecting important or influencing features. Using the ExtraTreesClassifier tool in python, random splits of all the observations were performed to avoid undesirable ML behavior. This is a score for the preprocessed dataset, displaying only the ten most significant features out of the 59 preprocessed features.
By order of importance, crucial features identified in the Figure 3 include the district of residence (district), shortness of breath (symptom_1), maximum diastolic blood pressure at rest (dbp_max_resrtbpr), maximum systolic blood pressure at rest (sbp_max_rerstbpr), maximum heart rate (maxhra), risk factor of decompensated heart failure called cardiac arrthym (risk_decom_5) alcohol intake (alcohol), sex of the HF patient (sex), number of days for the first hospitalization(hosp_days), and age(age) of the HF patient. These features play a critical role in determining the likelihood of a patient with HF to be readmitted to the hospital.

3.3. Model Building and Evaluation

3.3.1. Random Forest Classification

The confusion matrix shown by Table A1 in Appendix A indicates that the Randon Forest (RF) classification model achieved a 100% accuracy rate when predicting both classes 0 and 1 using training set. Additionally, the precision, recall, and f1-score metrics all achieved a score of 100%. The training set accuracy for this model is also 100%. On the other hand, while using the new data, i.e., the testing set, as shown by Table A2 in Appendix A, the confusion matrix reveals that the RF model correctly predicted 599 instances of class 0 but incorrectly predicted 109 instances of class 0.
For the new data or testing set, again the same Table A2 in Appendix A shows that the model incorrectly predicts 64 instances of class 1 and correctly predicts 616 instances of class 1. The precision of the model in class 0 is 89%, and in class 1, it is 84%. The recall score in class 0 is 84%, and in class 1, it is 89%. The f1-score is 0.87 in both class 0 and class 1. The accuracy of the RF model in the testing set is 87% as well.
Figure 4 depicts the AUC and ROC of the RF classifier. The figure shows that the AUC value is 0.94, and this implies that the model’s predictions are largely accurate and also that the probability of obtaining a wrong prediction from the model is 0.06. Although on a small scale, this wrong prediction is negligible; as the sample size increases, the wrong predictions are certain to take a toll on the model’s performance. The RF classifier has a low false-positive rate and a high true-positive rate if the AUC is high. This indicates that the RF classifier has a strong ability to categorize cases that are present (true positives) and that it also has a strong ability to classify situations that are absent (true negatives). This translates into the classifier having a high predictive accuracy, making it a useful tool for determining the pertinent features that go into the target variable.

3.3.2. SVM Classification Using Linear Kernel

The performance of the SVM classifier using a linear kernel was evaluated, and Table A3 in Appendix A presents the resulting confusion matrix and classification report for the training data. The SVM model with a linear kernel made 2149 correct predictions and 612 incorrect predictions for class 0, and for class 1, there were 2346 correct predictions and 443 incorrect predictions. The overall accuracy of the model in the training data is 81%, with a precision of 83% for class 0 and 79% for class 1.
On the other hand, using the testing data as presented in Table A4 in Appendix A, the SVM model with a linear kernel made 540 correct predictions and 168 incorrect predictions for class 0. For class 1, the model made 558 correct predictions and 122 incorrect predictions. The overall accuracy of the model in the testing set is 79%.
Figure 5 reveals that the AUC value is 0.88, which suggests that the classifier’s predictions are reasonably accurate and that there is a 0.12 chance of generating a prediction that is incorrect. Once the size of the dataset quadruples, this incorrect prediction will undoubtedly have an adverse effect on the model’s performance. It is crucial to keep in mind, though, that an AUC score of 0.88 does not imply that the SVM model is definitely the best model for the specific situation. Simply said, it indicates that the model is successful at classifying data according to the desired variable. Depending on the particular situation, various model performance measures may be assessed. In order to choose a model with confidence, it is vital to thoroughly compare various models using various performance indicators.

3.3.3. SVM Classification with Gaussian Radial Basis Function Kernel

Using the training data Table A5 in Appendix A shows the confusion matrix of the SVM model with a Gaussian radial basis function (RBF)kernel. It indicates that out of a total of 2761 predictions made, 2162 were valid, and 599 were false positives for class 0. Similarly, for class 1, the model made 2354 correct predictions and 435 false predictions. The overall prediction accuracy of the model is 81%.
The precision of the model in class 0 is 83%, and in class 1, it is 80%. On the other hand, using the testing data, Table A6 in Appendix A displays the confusion matrix and classification report for the SVM with Gaussian Radial Basis Function Kernal. The results in Table A6 in Appendix A indicate that the model correctly predicts 556 instances of class 0 but incorrectly predicts 152 instances of class 0. For class 1, the model correctly predicts 591 instances and made 89 false predictions. The overall accuracy of the model in the testing or new data is 83%. The precision of the model in class 0 is 86%, while in class 1, it is 80%.
Figure 6 demonstrates that the AUC value is 0.87, which suggests that the classifier’s predictions are reasonably accurate and that there is a 0.13 chance of generating a prediction that is incorrect.
Once the size of the dataset quadruples, this incorrect prediction will undoubtedly have an adverse effect on the model’s performance. It is crucial to keep in mind, though, that an AUC score of 0.88 does not imply that the SVM model is definitely the best model for the specific situation. Simply said, it indicates that the model is successful at classifying data according to the desired variable. Depending on the particular situation, various model performance measures may be assessed. In order to choose a classifier with confidence, it is vital to thoroughly compare various models using various performance indicators.

3.3.4. Evaluation of the KNN Classifier

The K-nearest neighbors (KNN) model evaluation on the training data is as shown in Table A7 in Appendix A.
It displays the confusion matrix and classification report of the KNN model on the training dataset. The model made 2319 correct predictions and 442 incorrect predictions for class 0, and for class 1, there were 2680 correct predictions and 109 incorrect predictions. The overall accuracy of the model in the training dataset is 90%. On the other hand, Table A8 in Appendix A presents the confusion matrix and classification report of the KNN model on the testing dataset.
The model’s accuracy in the testing data is 85%, with 554 correct predictions and 154 incorrect predictions for class 0. For class 1, the model made 621 correct predictions and 59 incorrect predictions.
Figure 7 shows that the AUC value for the KNN classifier is 0.88, which is nearly identical to that of the SVM (with linear kernel), and in the same vein, it signifies that the model’s predictions are largely accurate and that there is a 0.12 probability of receiving an inaccurate forecast of HF hospital readmission. The KNN model’s AUC of 0.88 further indicates that the model can accurately and confidently distinguish between positive and negative instances.

3.3.5. Evaluation of the DT Classifier

Table A9 in Appendix A, indicates that the DT classifier has a prediction accuracy of 54% in the training data.
The model performed poorly on class 0, as it incorrectly predicts almost the entire class 2503 and correctly predicts only 258 but performs relatively well on class 1 by correctly predicting almost the entire class of 2724 and incorrectly predicts only 65 instances. The precision, recall, and f1-score of the model on class 0 are 80%, 9%, and 17%, respectively. This indicates that the model did not learn much about class 0. The overall performance is 54%. Likewise, on the testing set, the model also performs poorly, as it did not learn well in either class 0 or class 1. Table A10 in Appendix A provides a detailed confusion matrix and classification report of the decision tree model on the testing set. The overall performance of this model on the testing data is 52%, which is the lowest in these trained models.
For the area under the ROC curve, Figure 8 shows an AUC of 0.57, which indicates that the DT classifier is not trustworthy enough to provide solid predictions of HF hospital readmission in Rwanda.
Simply said, the model is just slightly more accurate than speculating. It is not a reliable model for making critical judgments or predictions of HF hospital readmission in Rwanda. With an AUC of 0.57, the decision tree model cannot be trusted to make reliable predictions.

3.3.6. Evaluation of the Logistic Regression (LR) Classifier

Table A11 in Appendix Ashows the evaluation of the LR classifier on the training part of the dataset. The model correctly predicts 2056 instances of class 0, but incorrectly predicts 705 instances of the same class. Again, the model correctly predicts 2171 instances of class 1, but incorrectly predicts 618 instances of class 1. The overall accuracy of the model in the training set is 76%, with a precision of 77% for class 0 and 75% for class 1, as well as a recall of 74% for class 0 and 78% for class 1.
The performance of the logistic regression model on the testing set is illustrated in Table A12 in Appendix A. The findings revealed that the model achieves an accuracy score of 77% in its predictions. In addition, the same table indicates that the model makes 531 correct predictions on class 0 and 177 incorrect predictions, while for class 1, the model makes 532 correct predictions and 148 incorrect predictions. The precision of the model in class 0 is 78%, and in class 1, it is 75%. Then again, Figure 9 shows a LR classifier with an AUC of 0.81, which indicates that the model has a decent ability to distinguish between the readmitted and non-readmitted HF patients.
An AUC score of 0.5 is similar to random guessing, whereas an AUC score of 1 denotes a perfect classifier. As a result, an AUC of 0.81 shows that the model is reasonably effective at predicting the response variable and is better than random guessing. Here, the positive class denotes those who are prone to HF readmission, whereas the negative class denotes those who are not. An AUC of 0.81 means that it can accurately classify 81% of patients as subject to readmission; in other words, the model’s error rate of 19% indicates that it predicts both positive and negative classes with the same level of error.

3.3.7. Model Building Using Multilayer Perceptron Model

Table A13 in Appendix A presents the evaluation of the multilayer perceptron (MLP) model on the training set. The model correctly predicted 2476 instances of class 0 but incorrectly predicted 285 instances of class 0 as class 1. The model also correctly predicted 2615 instances of class 1 but incorrectly predicted 174 instances of class 1 as class 0. The overall accuracy of the model in the training set is 92%, with a precision of 93% for class 0 and 90% for class 1, as well as a recall of 90% for class 0 and 94% for class 1.
Furthermore, the performance of the MLP classifier on the testing set is illustrated in Table A14 in Appendix A. The model achieved an accuracy score of 82% in its predictions. The confusion matrix indicates that the model made 552 correct predictions on class 0 and 156 incorrect predictions, while for class 1, the model made 583 correct predictions and 97 incorrect predictions. The precision of the model in class 0 is 85%, and in class 1, it is 79%.
Figure 10 displays the MLP’s AUC and ROC. It illustrates the model’s AUC value of 0.88, which shows that predictions are only marginally accurate and that there is a 0.12 probability of receiving an inaccurate one.
An AUC of 0.88 produced by the MLP model is regarded as a reasonably good result, showing that the model can be used practically and has a respectable ability to distinguish between the positive and negative classes. However, by tweaking the model or including more elements in the data, more advancements might be achievable.

3.4. General Result Evaluation and Comparison

Each algorithm utilized to predict hospital readmission rates for HF in Rwanda has its own unique characteristics and tendencies. The review process aims to narrow down the choices to the most promising ones that can be implemented in Rwanda. During the evaluation process, the accuracy score and overall performance on both the training and testing datasets are considered when comparing the algorithms. Thus, the study concludes with evidence in the form of different results obtained from assessing all the models.
Figure 11 provides a comparison of the performance of all the models using the receiver operating characteristics (ROC) curve and the area under the curve (AUC) metric. From the graph, it is evident that the random forest model outperforms the other models, with an AUC score of 0.94. The KNN and SVM models follow with an AUC score of 0.88, while the logistic regression and decision tree models exhibit the worst performance, with AUC scores of 0.81 and 0.57, respectively.
The results in the Table 3 show that the area under the ROC curve for the decision trees model is 57%, which is the worst performance compared to the seven trained models. On the other hand, the random forest model outperformed all the trained models, with the area under the ROC curve of 94%. It is very important to mention that all four classifiers (SVM with linear kernel, SVM with Gaussian radial basis function kernel, KNN, and MLP) ranked second, with an excellent performance, having the area under the ROC curve of 88%; the LR classifier ranked third, with the area under the ROC curve of 81%.
The findings in the Figure 12 demonstrate that the choice of the model significantly affects how well the classification tasks are carried out. Therefore, it is very important to test various models and contrast the results to be able to identify the most accurate model that satisfies the dataset’s specific requirements.
Table 3, Figure 11 demonstrate that random forest outperforms other models, with an AUC of 94% and an evaluation accuracy of 87%. KNN and SVM models also show excellent performance, both having an AUC of 88%, followed by logistic regression, with an AUC of 84%. On the other hand, the decision tree model has the worst performance, with an AUC of 57%, similar to the baseline model.

4. Discussion

The purpose of this study was to investigate the applicability of multiple ML classifiers for predicting HF hospital readmission in Rwanda using locally gathered data. Our findings revealed notable insights into the effectiveness of these models. In terms of predicting hospital readmissions in the Rwandan context among the models evaluated, the random forest classifier emerged as the most promising option for predicting HF hospital readmission in Rwanda. In addition, the support vector machine, K-nearest neighbors, and multi-layer perceptron approaches all demonstrated admirable performance. Therefore, the results obtained in this study are consistent and reliable, which is supported by the fact that they align well with the study conducted in 2022 by Michailidis and colleagues [28]. This implies that the effectiveness of these classifiers can be, at least in part, generalized and is not limited to a particular dataset or context. On the other hand, it is crucial to note that the decision tree classifier only managed to achieve an area under the curve of 57%, and this performance is below average. This outcome differs significantly from the performance of the same model that has been used previously in the literature [29,30,31]. This might indicate that the decision tree classifier is not a good fit for the specific characteristics of the healthcare dataset from Rwanda.
Generally, the findings of this study as a whole highlight the potential of ML techniques in accurately predicting hospital readmission for HF patients in Rwanda. Nevertheless, system-specific challenges will need to be carefully considered in future studies due to the decision tree classifier’s poor performance, which calls into question its suitability. In fact, the use of these models can lead to better long-term health outcomes, reduced readmissions to hospitals, and enhanced patient care. These predictive capabilities might help healthcare professionals in Rwanda allocate resources more effectively and customize interventions to patients’ unique needs, thereby improving the standard of care. Nevertheless, it is important to recognize the limitation and challenges that come up when applying ML techniques to the Rwandan healthcare system. These classifiers’ performance and generalizability may be impacted by the particularities and complexities of the Rwandan healthcare system, including data availability, quality, and cultural considerations.

5. Conclusions and Recommendations

In order to improve the standard of care and health outcomes of HF in Rwanda, more research is still required. This can result in more accurate predictions, personalized treatment plans, and better utilization of healthcare resources. Future research could improve classifiers to better fit the local context, address data issues, and account for the unique constraints of the healthcare system. Addressing the identified knowledge gaps can aid in the development of more precise and relevant predictive models for HF readmission in Rwanda, which can have a positive impact on the country’s ability to manage and prevent HF hospital readmission.
While random forest classification shows the best performance, it is important to base actions on principles. The SVM approach also works well for this purpose. The study’s results indicate that ML techniques can accurately predict hospital readmission for HF patients in Rwanda, which can lead to improved care, fewer hospital readmissions, and better long-term health outcomes.
In order to help healthcare practitioners anticipate the possibility of readmission for specific patients, the classifiers can be used as decision support tools. By offering early interventions and personalized treatment plans, this technology can assist healthcare professionals in making the most use of their resources and improving patient outcomes. RF is a useful tool for anticipating HF readmissions, offering a strategy that can help patients obtain better results and make the most of healthcare resources. To minimize potential negative effects, such as over-reliance on the model’s predictions, which would impair clinical judgement and patient-centered treatment, healthcare practitioners must understand the model’s strengths and limits and ensure that it is used effectively.
To conclude, this study contributes to the growing body of literature on the application of the ML algorithm in medicine and suggests that ML has the potential to enhance HF management in Rwanda.

Author Contributions

T.R. and I.N. designed the study, participated in data collection, conducted data analysis, and drafted the manuscript. N.D.C. performed a critical supervision of the work for this study. All authors have read and agreed to the published version of the manuscript.

Funding

This study was funded by the government of Rwanda through National Council for Science and Technology (NCST), with the grant number NCST-NRIF/ERG-BATCH1/P04/2019. In addition, the University of Rwanda via the African Center of Excellence in Data Science funded this study.

Institutional Review Board Statement

The institutional review board and the National Health Committee approved this study. Then, the Ministry of Health in Rwanda authorized this study in the following statement: “Based on Rwandan Health Sector Research Policy and approval from the College of Medicine and Health Science Institutional Review Board Committee Ref: No.137/CMHS IRB/2019 of 28 March 2019 and National Health Research Committee Ref:NHRC/2019/PROT/009 of 23 February 2019, I am pleased to inform you that the Ministry of Health of Rwanda has granted authorization to conduct this research according to the approved research proposal”.

Informed Consent Statement

This study was retrospective and did not directly involve humans. Thus, an informed consent was not applicable.

Data Availability Statement

The dataset used in this study is available from the corresponding author upon request.

Acknowledgments

The administration of the University of Rwanda via its Single Unit Implementation is acknowledged for their support in hiring enumerators for this study and for helping in financial reporting.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Table A1. Confusion matrix and classification reports of the random forest (RF) classifier on the training set.
Table A1. Confusion matrix and classification reports of the random forest (RF) classifier on the training set.
[[2761    0]
 [   0 2789]]
1.0
precisionrecallf1-scoresupport
01.001.001.002761
11.001.001.002789
Accuracy 1.005550
Macro Avg1.001.001.005550
Weighted Avg1.001.001.005550
Table A2. Confusion matrix and classification reports of the random forest (RF) classifier on the testing set.
Table A2. Confusion matrix and classification reports of the random forest (RF) classifier on the testing set.
[[596 112]
  [ 72 608]]
Classification report
precisionrecallf1-scoresupport
00.890.840.87708
10.840.890.87680
Accuracy 0.871388
Macro Avg0.870.870.871388
Weighted Avg0.870.870.871388
Table A3. SVM (linear kernel) training data confusion matrix and classification report.
Table A3. SVM (linear kernel) training data confusion matrix and classification report.
[[2149  612]
  [ 443 2346]]
Classification report
precisionrecallf1-scoresupport
00.830.780.802761
10.790.840.822789
Accuracy 0.815550
Macro Avg0.810.810.815550
Weighted Avg0.810.810.815550
Table A4. SVM (linear kernel) testing data confusion matrix and classification report on the testing data.
Table A4. SVM (linear kernel) testing data confusion matrix and classification report on the testing data.
[[540  168]
  [ 122 558]]
Classification report
precisionrecallf1-scoresupport
00.820.760.79708
10.770.820.79680
Accuracy 0.791388
Macro Avg0.790.790.791388
Weighted Avg0.790.790.791388
Table A5. Confusion matrix and classification report for support vector machine with (RBF) on training data.
Table A5. Confusion matrix and classification report for support vector machine with (RBF) on training data.
[[2162  599]
  [ 435 2354]]
Classification report
precisionrecallf1-scoresupport
00.830.780.812761
10.800.840.822789
Accuracy 0.815550
Macro Avg0.810.810.815550
Weighted Avg0.810.810.815550
Table A6. Confusion matrix and classification report for SVM classifier with Gaussian (RBF) kernel on testing data.
Table A6. Confusion matrix and classification report for SVM classifier with Gaussian (RBF) kernel on testing data.
[[556 152]
  [ 89 591]]
Classification report
precisionrecallf1-scoresupport
00.860.790.82708
10.800.870.83680
Accuracy 0.831388
Macro Avg0.830.830.831388
Weighted Avg0.830.830.831388
Table A7. KNN training data confusion matrix and classification report.
Table A7. KNN training data confusion matrix and classification report.
[[2319  442]
  [ 109 2680]]
Classification report
precisionrecallf1-scoresupport
00.960.840.892761
10.860.960.912789
Accuracy 0.905550
Macro Avg0.910.910.915550
Weighted Avg0.910.910.915550
Table A8. KNN testing data confusion matrix and classification report.
Table A8. KNN testing data confusion matrix and classification report.
[[554  154]
  [ 59 621]]
Classification report
precisionrecallf1-scoresupport
00.900.780.84708
10.800.910.85680
Accuracy 0.851388
Macro Avg0.850.850.851388
Weighted Avg0.850.850.851388
Table A9. Decision tree confusion matrix and classification report on the training data.
Table A9. Decision tree confusion matrix and classification report on the training data.
[[258  2503]
  [ 65 2724]]
Classification report
precisionrecallf1-scoresupport
00.800.090.172761
10.520.980.682789
Accuracy 0.545550
Macro Avg0.660.540.425550
Weighted Avg0.660.540.425550
Table A10. Confusion matrix and classification report of decision tree on testing data.
Table A10. Confusion matrix and classification report of decision tree on testing data.
[[64 664]
  [ 27 653]]
Classification report
precisionrecallf1-scoresupport
00.700.090.16708
10.500.960.66680
Accuracy 0.521388
Macro Avg0.600.530.411388
Weighted Avg0.610.520.411388
Table A11. Logistic regression confusion matrix and classification report on training data.
Table A11. Logistic regression confusion matrix and classification report on training data.
[[2056 705]
  [ 618 2171]]
Classification report
precisionrecallf1-scoresupport
00.770.740.762761
10.750.780.772789
Accuracy 0.765550
Macro Avg0.760.760.765550
Weighted Avg0.760.760.765550
Table A12. Logistic regression confusion matrix and classification report on the testing data.
Table A12. Logistic regression confusion matrix and classification report on the testing data.
[[531 177]
  [ 148 532]]
Classification report
precisionrecallf1-scoresupport
00.780.750.77708
10.750.780.77680
Accuracy 0.771388
Macro Avg0.770.770.771388
Weighted Avg0.770.770.771388
Table A13. Multilayer perceptron confusion matrix and classification report on the training set.
Table A13. Multilayer perceptron confusion matrix and classification report on the training set.
[[2476 285]
  [ 174 2615]]
Classification report
precisionrecallf1-scoresupport
00.930.900.922761
10.900.940.922789
Accuracy 0.925550
Macro Avg0.920.920.925550
Weighted Avg0.920.920.925550
Table A14. Multilayer Perceptron confusion matrix and classification report on the testing set.
Table A14. Multilayer Perceptron confusion matrix and classification report on the testing set.
[[552 156]
  [ 96 583]]
Classification report
precisionrecallf1-scoresupport
00.850.780.81708
10.790.860.82680
Accuracy 0.821388
Macro Avg0.820.820.821388
Weighted Avg0.820.820.821388

References

  1. World Heart Federation Heart Failure. World Heart Federation. Available online: https://world-heart-federation.org/cvd-roadmaps/whf-global-roadmaps/heart-failure/ (accessed on 23 April 2023).
  2. Bragazzi, N.L.; Zhong, W.; Shu, J.; Abu Much, A.; Lotan, D.; Grupper, A.; Younis, A.; Dai, H. Burden of heart failure and underlying causes in 195 countries and territories from 1990 to 2017. Eur. J. Prev. Cardiol. 2021, 28, 1682–1690. [Google Scholar] [CrossRef] [PubMed]
  3. Awan, S.E.; Bennamoun, M.; Sohel, F.; Sanfilippo, F.M.; Dwivedi, G. Machine learning-based prediction of heart failure readmission or death: Implications of choosing the right model and the right metrics. ESC Heart Fail. 2019, 6, 428–435. [Google Scholar] [CrossRef]
  4. Sarker, I.H. Machine Learning: Algorithms, Real-World Applications and Research Directions. SN Comput. Sci. 2021, 2, 160. [Google Scholar] [CrossRef] [PubMed]
  5. Jiang, W.; Siddiqui, S.; Barnes, S.; Barouch, L.A.; Korley, F.; Martinez, D.A.; Toerper, M.; Cabral, S.; Hamrock, E.; Levin, S. Readmission risk trajectories for patients with heart failure using a dynamic prediction approach: Retrospective study. JMIR Public Health Surveill. 2019, 7, e14756. [Google Scholar] [CrossRef] [PubMed]
  6. Al-Omary, M.S.; Davies, A.J.; Evans, T.-J.; Bastian, B.; Fletcher, P.J.; Attia, J.; Boyle, A.J. Mortality and Readmission Following Hospitalisation for Heart Failure in Australia: A Systematic Review and Meta-Analysis. Hear. Lung Circ. 2018, 27, 917–927. [Google Scholar] [CrossRef]
  7. Savarese, G.; Becher, P.M.; Lund, L.H.; Seferovic, P.; Rosano, G.M.C.; Coats, A.J.S. Global burden of heart failure: A comprehensive and updated review of epidemiology. Cardiovasc. Res. 2022, 118, 3272–3287. [Google Scholar] [CrossRef]
  8. Gtif, I.; Bouzid, F.; Charfeddine, S.; Abid, L.; Kharrat, N. Heart failure disease: An African perspective. Arch. Cardiovasc. Dis. 2021, 114, 680–690. [Google Scholar] [CrossRef]
  9. Lippi, G.; Sanchis-Gomar, F. Global epidemiology and future trends of heart failure. AME Med. J. 2020, 5, 15. [Google Scholar] [CrossRef]
  10. Mandi, D.G.; Bamouni, J.; Yaméogo, R.A.; Naïbé, D.T.; Kaboré, E.; Kambiré, Y.; Kologo, K.J.; Millogo, G.R.C.; Zabsonré, P. Spectrum of heart failure in sub-Saharan Africa: Data from a tertiary hospital-based registry in the eastern center of Burkina Faso. Pan Afr. Med. J. 2020, 36, 30. [Google Scholar] [CrossRef]
  11. Niyibizi, J.B.; Okop, K.J.; Nganabashaka, J.P.; Umwali, G.; Rulisa, S.; Ntawuyirushintege, S.; Tumusiime, D.; Nyandwi, A.; Ntaganda, E.; Delobelle, P.; et al. Perceived cardiovascular disease risk and tailored communication strategies among rural and urban community dwellers in Rwanda: A qualitative study. BMC Public Health 2022, 22, 920. [Google Scholar] [CrossRef]
  12. NCD-Chapter-4-Heart-Failure.pdf.pdf. Available online: https://www.pih.org/sites/default/files/2017-07/NCD-Chapter-4-Heart-Failure.pdf.pdf (accessed on 24 April 2023).
  13. Shin, S.; Austin, P.C.; Ross, H.J.; Abdel-Qadir, H.; Freitas, C.; Tomlinson, G.; Chicco, D.; Mahendiran, M.; Lawler, P.R.; Billia, F.; et al. Machine learning vs. conventional statistical models for predicting heart failure readmission and mortality. ESC Heart Fail. 2021, 8, 106–115. [Google Scholar] [CrossRef] [PubMed]
  14. Cruz, J.A.; Wishart, D.S. Applications of machine learning in cancer prediction and prognosis. Cancer Inf. 2007, 2, 59–77. [Google Scholar] [CrossRef]
  15. Chawla, N.V.; Bowyer, K.W.; Hall, L.O.; Kegelmeyer, W.P. SMOTE: Synthetic Minority Over-sampling Technique. J. Artif. Intell. Res. 2002, 16, 321–357. [Google Scholar] [CrossRef]
  16. Uddin, S.; Khan, A.; Hossain, M.E.; Moni, M.A. Comparing different supervised machine learning algorithms for disease prediction. BMC Med. Inform. Decis. Mak. 2019, 19, 281. [Google Scholar] [CrossRef]
  17. Desai, R.J.; Wang, S.V.; Vaduganathan, M.; Evers, T.; Schneeweiss, S. Comparison of machine learning methods with traditional models for use of administrative claims with electronic medical records to predict heart failure outcomes. JAMA Netw. Open 2020, 3, e1918962. [Google Scholar] [CrossRef]
  18. de Ridder, D.; de Ridder, J.; Reinders, M.J.T. Pattern recognition in bioinformatics. Briefings Bioinform. 2013, 14, 633–647. [Google Scholar] [CrossRef]
  19. Bayati, M.; Braverman, M.; Gillam, M.; Mack, K.M.; Ruiz, G.; Smith, M.S.; Horvitz, E. Data-Driven Decisions for Reducing Readmissions for Heart Failure: General Methodology and Case Study. PLoS ONE 2014, 9, e109264. [Google Scholar] [CrossRef]
  20. Heidenreich, P.A.; Bozkurt, B.; Aguilar, D.; Allen, L.A.; Byun, J.-J.; Colvin, M.M.; Deswal, A.; Drazner, M.H.; Dunlay, S.M.; Evers, L.R.; et al. 2022 ACC/AHA/HFSA Guideline for the Management of Heart Failure: Executive Summary. J. Card. Fail. 2022, 28, 810–830. [Google Scholar] [CrossRef]
  21. Callender, T.; Woodward, M.; Roth, G.; Farzadfar, F.; Lemarie, J.-C.; Gicquel, S.; Atherton, J.; Rahimzadeh, S.; Ghaziani, M.; Shaikh, M.; et al. Heart Failure Care in Low- and Middle-Income Countries: A Systematic Review and Meta-Analysis. PLoS Med. 2014, 11, e1001699. [Google Scholar] [CrossRef]
  22. ACCF/AHA Task Force Members. 2013 ACCF/AHA Guideline for the Management of Heart Failure: A Report of the American College of Cardiology Foundation/American Heart Association Task Force on Practice Guidelines. J. Am. Coll. Cardiol. 2013, 62, e147–e239. [Google Scholar]
  23. Szymanski, P.Z.; Badri, M.; Mayosi, B.M. Clinical characteristics and causes of heart failure, adherence to treatment guidelines, and mortality of patients with acute heart failure: Experience at Groote Schuur Hospital, Cape Town, South Africa. S. Afr. Med. J. 2018, 108, 94–98. [Google Scholar] [CrossRef] [PubMed]
  24. Xu, Z.; Shen, D.; Kou, Y.; Nie, T. A Synthetic Minority Oversampling Technique Based on Gaussian Mixture Model Filtering for Imbalanced Data Classification. IEEE Trans. Neural Netw. Learn. Syst. 2022. Online ahead of print. [Google Scholar] [CrossRef]
  25. Zhao, Y.; Wood, E.P.; Mirin, N.; Cook, S.H.; Chunara, R. Social determinants in machine learning cardiovascular disease prediction models: A systematic review. Am. J. Prev. Med. 2021, 61, 596–605. [Google Scholar] [CrossRef] [PubMed]
  26. Dobbin, K.K.; Simon, R.M. Optimally splitting cases for training and testing high dimensional classifiers. BMC Med. Genom. 2011, 4, 31. [Google Scholar] [CrossRef] [PubMed]
  27. Guo, A.; Pasque, M.; Loh, F.; Mann, D.L.; Payne, P.R.O. Heart failure diagnosis, readmission, and mortality prediction using machine learning and artificial intelligence models. Curr. Epidemiol. Rep. 2020, 7, 212–219. [Google Scholar] [CrossRef]
  28. Michailidis, P.; Dimitriadou, A.; Papadimitriou, T.; Gogas, P. Forecasting Hospital Readmissions with Machine Learning. Healthcare 2022, 10, 981. [Google Scholar] [CrossRef]
  29. Purushottam; Saxena, K.; Sharma, R. Efficient heart disease prediction system using decision tree. In Proceedings of the International Conference on Computing, Communication & Automation, Greater Noida, India, 15–16 May 2015; pp. 72–77. [Google Scholar]
  30. Abdar, M. Using Decision Trees in Data Mining for Predicting Factors Influencing of Heart Disease. Carpathian J. Electron. Comput. Eng. 2015, 8, 31–36. [Google Scholar]
  31. Maheswari, S.; Pitchai, R. Heart Disease Prediction System Using Decision Tree and Naive Bayes Algorithm. Curr. Med. Imaging Rev. 2019, 15, 712–717. [Google Scholar] [CrossRef]
Figure 1. Admitted patients per age.
Figure 1. Admitted patients per age.
Jpm 13 01393 g001
Figure 2. Imbalances in the target class before using SMOTE in handling the imbalance.
Figure 2. Imbalances in the target class before using SMOTE in handling the imbalance.
Jpm 13 01393 g002
Figure 3. Graph of features and their importance.
Figure 3. Graph of features and their importance.
Jpm 13 01393 g003
Figure 4. ROC curve for RF classifier.
Figure 4. ROC curve for RF classifier.
Jpm 13 01393 g004
Figure 5. ROC curve for SVM classification using the linear kernel.
Figure 5. ROC curve for SVM classification using the linear kernel.
Jpm 13 01393 g005
Figure 6. ROC curve for SVM classifier with Gaussian radial basis function kernel.
Figure 6. ROC curve for SVM classifier with Gaussian radial basis function kernel.
Jpm 13 01393 g006
Figure 7. ROC curve for KNN classifier.
Figure 7. ROC curve for KNN classifier.
Jpm 13 01393 g007
Figure 8. ROC curve for DT classifier.
Figure 8. ROC curve for DT classifier.
Jpm 13 01393 g008
Figure 9. ROC curve for the LR classifier.
Figure 9. ROC curve for the LR classifier.
Jpm 13 01393 g009
Figure 10. ROC for the MLP classifier.
Figure 10. ROC for the MLP classifier.
Jpm 13 01393 g010
Figure 11. Receiver operating characteristics and area under the curve (ROC and AUC).
Figure 11. Receiver operating characteristics and area under the curve (ROC and AUC).
Jpm 13 01393 g011
Figure 12. Bar plot comparison of model performance from Table 3.
Figure 12. Bar plot comparison of model performance from Table 3.
Jpm 13 01393 g012
Table 1. Sample dataset (5 rows × 75 columns).
Table 1. Sample dataset (5 rows × 75 columns).
S/NAgeSexAlcoholRisk_Decom_5Hosp_DaysMax_HraSbp_MaxrestDbp_Maxrest
0600031289462
101002115610959
210005317011388
3101001411113372
4150003514212090
Table 2. Results of the target class before and after handling the class imbalance.
Table 2. Results of the target class before and after handling the class imbalance.
ClassNumber of Classes before BalancingNumber of Classes after Balancing
0 (no readmission)34693469
1 (at least one readmission)6143469
TOTAL40836938
Table 3. Tabular summary of algorithm performance.
Table 3. Tabular summary of algorithm performance.
Accuracy %Precision %Recall %F1_Score %ROC_AUC %
Algorithm
Random Forest8784898794
SVM_RBF7977827988
SVM_Linear7473767488
KNN8580918588
MLP8279868288
Logistic Regression7574757481
Decision Tree5250966657
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Rizinde, T.; Ngaruye, I.; Cahill, N.D. Comparing Machine Learning Classifiers for Predicting Hospital Readmission of Heart Failure Patients in Rwanda. J. Pers. Med. 2023, 13, 1393. https://doi.org/10.3390/jpm13091393

AMA Style

Rizinde T, Ngaruye I, Cahill ND. Comparing Machine Learning Classifiers for Predicting Hospital Readmission of Heart Failure Patients in Rwanda. Journal of Personalized Medicine. 2023; 13(9):1393. https://doi.org/10.3390/jpm13091393

Chicago/Turabian Style

Rizinde, Theogene, Innocent Ngaruye, and Nathan D. Cahill. 2023. "Comparing Machine Learning Classifiers for Predicting Hospital Readmission of Heart Failure Patients in Rwanda" Journal of Personalized Medicine 13, no. 9: 1393. https://doi.org/10.3390/jpm13091393

APA Style

Rizinde, T., Ngaruye, I., & Cahill, N. D. (2023). Comparing Machine Learning Classifiers for Predicting Hospital Readmission of Heart Failure Patients in Rwanda. Journal of Personalized Medicine, 13(9), 1393. https://doi.org/10.3390/jpm13091393

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop