Next Article in Journal
Body Mapping in a Drug and Alcohol Treatment Program: Eliciting New Identity and Experience
Next Article in Special Issue
Does Exposure of Lead and Cadmium Affect the Endometriosis?
Previous Article in Journal
Anaerobic Ammonium Oxidation Bacteria in a Freshwater Recirculating Pond Aquaculture System
Previous Article in Special Issue
The Association between Nonalcoholic Fatty Liver Disease and Stroke: Results from the Korean Genome and Epidemiology Study (KoGES)
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

The Comprehensive Machine Learning Analytics for Heart Failure

1
Institute of Public Health, School of Medicine, National Yang-Ming University, Taipei 112, Taiwan
2
Institute of Public Health, School of Medicine, National Yang Ming Chiao Tung University, Hsinchu 300, Taiwan
3
Center for Evidence-Based Medicine, Veteran General Hospital, Taipei 112, Taiwan
4
Department of Medicine, School of Medicine, National Yang Ming Chiao Tung University, Hsinchu 300, Taiwan
*
Authors to whom correspondence should be addressed.
Int. J. Environ. Res. Public Health 2021, 18(9), 4943; https://doi.org/10.3390/ijerph18094943
Submission received: 26 March 2021 / Revised: 1 May 2021 / Accepted: 4 May 2021 / Published: 6 May 2021
(This article belongs to the Special Issue Big Data for Public Health Research and Practice)

Abstract

:
Background: Early detection of heart failure is the basis for better medical treatment and prognosis. Over the last decades, both prevalence and incidence rates of heart failure have increased worldwide, resulting in a significant global public health issue. However, an early diagnosis is not an easy task because symptoms of heart failure are usually non-specific. Therefore, this study aims to develop a risk prediction model for incident heart failure through a machine learning-based predictive model. Although African Americans have a higher risk of incident heart failure among all populations, few studies have developed a heart failure risk prediction model for African Americans. Methods: This research implemented the Least Absolute Shrinkage and Selection Operator (LASSO) logistic regression, support vector machine, random forest, and Extreme Gradient Boosting (XGBoost) to establish the Jackson Heart Study’s predictive model. In the analysis of real data, missing data are problematic when building a predictive model. Here, we evaluate predictors’ inclusion with various missing rates and different missing imputation strategies to discover the optimal analytics. Results: According to hundreds of models that we examined, the best predictive model was the XGBoost that included variables with a missing rate of less than 30 percent, and we imputed missing values by non-parametric random forest imputation. The optimal XGBoost machine demonstrated an Area Under Curve (AUC) of 0.8409 to predict heart failure for the Jackson Heart Study. Conclusion: This research identifies variations of diabetes medication as the most crucial risk factor for heart failure compared to the complete cases approach that failed to discover this phenomenon.

1. Introduction

The incidence and prevalence of heart failure have been increasing in recent years, and this phenomenon has become a serious global public health issue [1]. Research has estimated that more than 5.8 million people are suffering from heart failure in the United States, and this is increasing at a rate of at least 550,000 per year, while in the world, more than 23 million people are suffering from heart failure [2]. Other studies also pointed out that African Americans have a higher incidence of heart failure than other races [3,4], and even before the age of 50, the risk of heart failure of African Americans is 20 times that of whites [4].
Machine learning is equipped with algorithms that improve performance with experience [5]. A machine learns if its performance at a task improves with experience [6]. In constructing a model using machine learning techniques, the first step divides the entire data into two parts: the training data and the testing data. The training data refers to the dataset used to train the machine learning hyperparameters. The optimized model parameters are estimated based on the training datasets. The testing data is independent of the training data. The testing data will not participate in the training process but will only be used to evaluate the trained model. To avoid overfitting [7], K-Fold Cross Validation (CV) is a common strategy. The CV errors based on accuracy, Mean Square Error (MSE), and F1 Score are adopted [8].
Although it has been indicated that African Americans have a higher risk of incident heart failure among all populations [3], few studies have developed a risk prediction model for heart failure in African Americans. Therefore, this study utilized the Jackson Heart Study (JHS), the most prominent African American research database in the United States, and machine learning methods to construct comprehensive predictive models for heart failure. Our large-scale analysis structure aimed to discover missing and yet important information in such a big-data approach.
Random forest can provide feature importance through decision trees [9]. Breiman discussed the complete algorithm in 2001 [10], a type of ensemble method that collects multiple weak classifiers that produce a robust classifier [11]. The Classification and Regression Tree (CART) is a decision tree for predictive classification and continuous value [12] that adopts the binary division rule. Each time a division generates two branches, and the Gini classification method determines which branch is the best. Extreme Gradient Boosting (XGBoost) is a further extension from GBDT (Gradient Boosted Decision Tree) [13].
After missing data are correctly imputed, four methods, including (1) the Least Absolute Shrinkage and Selection Operator (LASSO) logistic regression, (2) Support Vector Machine (SVM) [6,7], (3) random forest, and (4) Extreme Gradient Boosting (XGBoost), would be constructed for a predictive model for heart failure based on the African American population. The inclusion of the four strategies considers the essential aspects of the analysis concepts. LASSO logistic regression is the conventional statistical approach with a penalty term λ j = 1 p | β j |   that avoids overfitting in regression models. The SVM is a machine learning tool dealing with classification problems. The SVM classifies subjects according to the separating hyperplane, which is defined as ω T x i + b = 0 , where ω T x i + b 1 , y i = 1 and ω T x i + b 1 , y i = 1 . The last two methods are tree-based machine learning. However, the random forest and XGBoost are based on different mathematical optimizations. Random forest is an ensemble learning method that constructs a multitude of decision trees at training time. However, the XGBoost depends on gradient boosting trees, and the objection function is defined as O b j ( t ) = j = 1 T [ ( G j ) w j + 1 2 ( H j + λ ) w j 2 ] + γ T , where G j = i I j g i is the sum of the first derivative of the loss function in the jth leaf and H j = i I j h i represents similar calculations for the second derivative.
As a result, we examined 112 scenarios, and Figure 1 displays the analysis plan. To date, this research had the most extensive models, with seven missing patterns for quality control, four missing imputation strategies, and four power machine learning models. We aimed to discover novel indications for heart failure from such a big scale of analysis models.

2. Materials and Methods

Following the previous work [14], the JHS initially included 3883 people. After excluding those who had experienced heart failures and those who were unsure whether they had heart failures or not at the baseline, 3327 people remained in the study. At the end of the study, 246 incident heart failure cases were identified. Table 1 shows the descriptive statistics. Subjects with heart failures accounted for 7.4% of the overall study sample.
Regarding the missing data, variables in the database were classified according to the missing rates, which were <1%, <3%, <5%, <10%, <20%, <30%, and <40%. The missing rate was specific to the variable but not the proportion of missing data in the whole dataset. The column (<40%) means that this dataset contained all variables with missing rates less than 40%. Therefore, more variables were included in the analyses compared to other missing rates. Four imputation techniques dealt with the missing values, including (1) complete cases, (2) simple imputation, (3) the K Nearest-Neighbor (KNN) interpolation method [15], and (4) random forest imputation. More details of the imputation strategy and the statistical properties are available in a former article [16]. Besides, the random forest interpolation method is a machine learning-based imputation of the improved “MissForest” interpolation method [17].
After missing values were imputed, the next step was to generate dummy variables for the unordered category variables. This research adopted the Python language with “one-hot encoding” for categorical variables. We normalized continuous variables according to Min-Max normalization. The formula is x = x min ( x ) max ( x ) min ( x ) . The Min-Max normalization limited the range of each variable in the dataset to [0, 1]. In this way, each variable had the same scale, which could improve the predictive performance by avoiding an extremely skewed distribution. Finally, we divided the dataset into the training set (70%) and the testing set (30%). Figure 2 demonstrates the flow chart of the entire data processing.
Note that subjects with heart failures in this study population accounted for 7.4% of the overall sample. For machine learning models, this is an imbalance problem that introduces biases when building the model. We controlled the “class weight” in the LASSO logistic regression, SVM, and random forest to avoid potential biases. The “class weight” parameter in this paper is “balanced,” and the formula is: w j = n k n j , where “ w j ” represents the weight of the jth category of the sample, “n” represents the sample size, “k” represents the number of categories, and “nj” represents the sample size of the jth category. In the XGBoost, increasing the weight by adjusting the parameter “scale_pos_weight” could increase the predictive ability. The “scale_pos_weight” parameter finds the most appropriate value through K-fold cross-validation, and its range is between 1 and 10.
The logistic regression model is estimated by maximum likelihood [18], but we implemented the Least Absolute Shrinkage and Selection Operator (LASSO)-based model in this research [19]. The SVM [20] with various kernel functions were considered, including the linear kernel, polynomial kernel, and radial basis function kernel. Ten-fold cross-validation found the most suitable kernel function to establish the optimal SVM to predict heart failure.
This research’s XGBoost model is based on 10-fold cross-validation to find the best parameters to establish a heart failure prediction model. The parameters and ranges are as follows: “scale_pos_weight”: range 1~10. “n_estimators”: initial training 90~150, final training 500~3500, “max_depth”: range 2~10. “min_child_weight”: range 1~10. “Gamma”: range 0~5. “Subsample”: range 0.3~1. “colsample_bytree”: range 0.3~1. “reg_lambda”: range 10 3 ~ 10 2 . “reg_alpha”: range 10 3 ~ 10 2 . “learning_rate”: range 0.01~0.1. As there are too many XGBoost parameters, it is time-consuming to train all the parameters simultaneously. Therefore, our procedure adopted training one set of two sets of parameters and gradually trained the XGBoost model (Figure 3).

3. Results

If the analyses excluded missing data, Figure 4 shows the corresponding results, where the Y-axis is Area Under Curve (AUC), and the X-axis is the proportion of the missing value of the data. The predictive ability of XGBoost was better than the other three in most scenarios, and the best AUC was 0.808 when the missing rates of included variables were less than 3%. The best AUC of random forest was 0.8003 when the missing rates were less than 3%. The best AUC of SVM was 0.7892 when the missing rates were less than 3%. The best AUC of LASSO logistic regression was 0.7676 when the missing rates were less than 3%.
In summary, the best overall AUC performance occurs when the missing rates of included variables are less than 3%. When we included more variables in the analysis, we also excluded more subjects from the analysis due to missing values. As a result, the AUC decreased.
If the simple interpolation method is adopted, Figure 5 reveals the corresponding results. Similarly, XGBoost was the best performer, and the highest AUC was 0.8239 when the missing rates were less than 30%. The predictive abilities of random forest and LASSO logistic regression are very similar. The best AUC of random forest was 0.813, and LASSO logistic regression demonstrated a value of 0.814. Lastly, the AUC of SVM was 0.79 on average.
In summary, the best overall AUC performance was when the proportion of missing values of the included variables was less than 30%. If we further included the variables whose missing value ratio was less than 40%, the AUC decreased. Therefore, if a follow-up researcher wants to use a simple interpolation method to establish a prediction model, the suggestion from the results of this paper is to include variables whose proportion of missing values is less than 30%. The prediction performance will be relatively good, of which the best performance was the XGBoost prediction model, which had an AUC of 0.8239.
If the KNN interpolation method is adopted, Figure 6 shows the corresponding results. Among all scenarios, XGBoost remained the best performer, and the highest AUC was 0.8369 if the missing rates were less than 30%. Random forest and LASSO logistic revealed similar performance. The best AUC of random forest was 08136 when the missing rates were less than 30%, and LASSO logistic regression was 0.8157 when the missing rates were less than 20%. Finally, the AUC of SVM was 0.79 if the missing rates were less than 30%. In summary, the best AUC occurs when the missing rates are less than 30%.
If the random forest interpolation method is adopted, Figure 7 shows the corresponding results. The XGBoost outperformed the other three methods in all analyses, and the best AUC was 0.8409 when the missing rates were less than 30%. Random forest and LASSO logic regression have similar predictive abilities. The best AUC of random forest was 0.8191 when the missing rates were less than 30%, while LASSO logic regression was 0.8166 when the missing rates were less than 20%. Lastly, the AUC of SVM was still about 0.79. In summary, the best AUC occurs when the missing rates are less than 30%.
According to these results, our analysis plan is an excellent example for future studies in various health outcomes. We conclude that even when variables have a missing rate as high as 30%, as long as we adopt the random forest imputation before the analysis the XGBoost provides the best predictive ability.

4. Conclusions

In this research, the largest African American database in the United States, the JHS, was comprehensively examined by four different machine learning techniques to establish a heart failure predictive model. We also considered the impact of different missing rates and missing data imputation strategies on the predictive ability to discover the best analytic model. Besides, this work provides a series of guidelines for researchers interested in the JHS for future studies.
Based on these results, we conclude that (1) the XGBoost has the best performance in most situations to predict heart failure, especially when we implemented the random forest interpolation with the missing rates at the category of less than 30%. (2) The predictive performance using imputed data would be better than the models without imputations. Even the simplest imputation outperforms the complete cases scenario. (3) When building a predictive model, including too many variables (missing rates less than 40%) results in a worse prediction. The results revealed that in the analysis of complete cases, including variables with a missing rate of 3%, the predictive model performs better. Regarding the imputed data, it can tolerate up to 30% of the missing rates.
Finally, our work uses the feature importance function in random forest and XGBoost to discover which variables significantly contribute to predicting heart failures. Feature importance scores can provide insight and guidance in building the predictive model. We could use the importance scores to delete variables with the lowest scores and keep the variables with the highest scores.
Since the XGBoost performs better than random forest, we did not show the results of random forest. Table 2, Table 3, Table 4 and Table 5 display details of the XGBoost. If we excluded missing data from the analysis, risk factors could not be identified consistently (Table 2). In contrast, if the analyses incorporated the imputed data, diabetes medication (DMmeds) was the most crucial feature suggested by the XGBoost results regardless of the imputation strategy. In Table 2, for data missing <5%, the feature importance of diabetes medication is 0.0492 compared to the age variable with 0.0290. Thus, diabetes medication is almost two times more important when building the predictive model (0.0492/0.029). DMmeds is a dichotomous variable. In this research, 15.25% of the population had a value of 1 with at least one diabetes medication. In total, 84.75% were without diabetes medication. Therefore, variations of diabetes medication were the most influential risk factor for heart failure.
Furthermore, most of the findings revealed that heart failure was related to age, sex, diabetes, kidney function, and past heart disease history. Previous studies also pointed out that these variables are risk factors for heart failure [21,22,23], suggesting that our prediction models established by machine learning could discover crucial risk factors. In particular, a former study indicated significant predictors, including age, sex, body mass index, diabetes mellitus, systolic blood pressure, creatinine, serum albumin or total protein, LV hypertrophy, and coronary artery disease [24]. Compared to the simple AHEAD scoring system [25], the other AHEAD-U scoring system proposed by Chen et al. [26], and the most recent HANBAH score by Guo et al. [27], the most crucial feature importance variables determined by the XGBoost are similar to these variables used in the scoring system. However, variations of diabetes medications have never been reported previously. Therefore, the XGBoost discovered a novel influential risk factor and other similar factors that match previous studies.
In summary, the results suggest using variables with missing rates of less than 30% and adopting the Random Forest imputations when building a heart failure predictive model using the JHS database. Finally, we recommend that the XGBoost build the risk model, and the highest predictive accuracy is 0.8409. Most importantly, this research identifies variations of diabetes medication as the most crucial risk factor for heart failure, regardless of the missing data imputation strategy, since the variable of diabetes medications has the highest feater importance. In contrast, the commonly adopted approach that only uses complete cases failed to provide such novel findings.
The variable “diabetes medication” indicates that patients had diabetes, which is different from the conventional understanding of diabetic treatment [28]. The superficial understanding suggests that patients receiving diabetic treatment had a worse prognosis than those without treatment. It can be misunderstood that diabetes medication is the most crucial risk factor for heart failure. In this research, we interpret the variable “diabetes medication” as variations of diabetes medication since this variable has multiple levels, and we treated it as a dummy variable in the analysis.
In previous research [29,30], the authors used five methods to predict heart failure, including logistic regression, decision tree, random forest, simple Bayesian classifier, and support vector machine (SVM). The source of the data is from the Cleveland database. The number of samples included was 303, and the number of variables included was 14. For the data’s missing values, the study used the K-Nearest Neighbor (KNN) interpolation method. In the final predictive results, the study adopted accuracy to conclude that the five methods have good prediction accuracy (87.36%, 93.19%, 89.14%, 87.27%, 92.30%), and the decision tree performed best.
Later, the random survival forest [31] built a predictive model. The database was Action to Control Cardiovascular Risk in Diabetes (ACCORD). There were 8756 samples included. After excluding variables with a missing value greater than 10%, the number of variables included was 109. Regarding the missing values, this research imputed missing data using random forest interpolation. The C-index evaluated the final predictive results, and the random survival forest yielded a value of 0.74.
Another study implemented three methods, including logistic regression, SVM, and Ada-Boost, to predict heart failure [32]. They came from the Geisinger Clinic Electronic Health Record system (Electronic Health Record). The sample size was 4489, and the number of variables was 179. For the missing values, this study used the missing indicator method. For model optimization and selection of parameters, this study adopted a 10-fold CV. The study used the Area Under the Curve (AUC) to compare the final predictive results. The Receiver Operating Characteristic (ROC) curve is an excellent diagnostic tool [33]. The Youden Index is also a popular choice [34] defined as   argmax ( true   positive   rate   + true   negative   rate 1 ,   1 ) . The higher Area Under the Curve (AUC) [35] also indicates a better prediction [36]. Previous research suggested the interpretations of different cutoffs of AUC higher than 0.5 [37]. The logistic regression results were similar to the Ada-Boost with an AUC of 0.77, and the SVM was relatively unsatisfactory with an AUC of 0.62.
In future development, one could further discuss the following aspects. The problem of imbalanced data explains the weight of a small number of samples to be adjusted. There are many methods to deal with this issue, such as undersampling, which reduces multiple types of samples, or oversampling (SMOTE [38]) to generate more small samples, such that a ratio of normal and diseased samples of 1:1.
Other methods can build predictive models, such as the artificial neural network, naive Bayes, or other ensemble algorithms. These areas are essential topics in the future.

Author Contributions

C.-Y.G. proposed the study design; C.-Y.G. supervised statistical analyses; C.-Y.G. drafted the manuscript; M.-Y.W. carried out analysis and generated results; H.-M.C. secured funding support; H.-M.C. supervised the project and manuscript edits. All authors have read and agreed to the published version of the manuscript.

Funding

The Jackson Heart Study (JHS) is supported and conducted in collaboration with Jackson State University (HHSN268201800013I), Tougaloo College (HHSN268201800014I), the Mississippi State Department of Health (HHSN268201800015I/HHSN26800001), and the University of Mississippi Medical Center (HHSN268201800010I, HHSN268201800011I, and HHSN268201800012I) contracts from the National Heart, Lung, and Blood Institute (NHLBI) and the National Institute for Minority Health and Health Disparities (NIMHD). The authors also wish to thank the staff and participants of the JHS. Grants were received from the Ministry of Health and Welfare (MOHW104-TDU-B-211-113-003, MOHW106-TDU-B-211-113001) and an intramural grant from National Yang-Ming University (E107F-M01-0501). This work is particularly supported by “Yin Yen-Liang Foundation Development and Construction Plan” of the School of Medicine, National Yang-Ming University.

Institutional Review Board Statement

The study has been approved by the IRB of Taipei Veterans General Hospital.

Informed Consent Statement

Not applicable.

Data Availability Statement

This manuscript was prepared using the database of Jackson Heart Study with materials obtained from the NHLBI Biologic Specimen and Data Repository Information Coordinating Center and does not necessarily reflect the opinions or views of the Jackson Heart Study or the NHLBI. The views expressed in this manuscript are those of the authors and do not necessarily represent the views of the National Heart, Lung, and Blood Institute; the National Institutes of Health; or the U.S. Department of Health and Human Services.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Table A1. Coding book for variables included in this research.
Table A1. Coding book for variables included in this research.
Variable NameVariable TypesVariable Description
1. Demographics
ageContinuousAge in Years
sexCategoricalParticipant Sex
alcCategoricalAlcohol drinking in the past 12 months (Y/N)
alcwContinuousAverage number of drinks per week
currentSmokerCategoricalSelf-Reported Cigarette Smoking Status
everSmokerCategoricalSelf-Reported History of Cigarette Smoking
2. Anthropometrics
weightContinuousWeight (kg)
heightContinuousHeight (cm)
BMIContinuousBody Mass Index (kg/m2)
waistContinuousWaist Circumference (cm)
neckContinuousNeck Circumference (cm)
bsaContinuousCalculated Body Surface Area (m2)
obesity3catCategoricalIdeal Health: BMI < 25 (Normal)
Intermediate Health: 25 ≤ BMI < 30 (Overweight)
Poor Health: BMI ≥ 30 (Obese)
3. Medications
medAcctCategoricalMedication Accountability
BPmedsSelfCategoricalSelf-Reported Blood Pressure Medication Status (Y/N)
BPmedsCategoricalBlood Pressure Medication Status (Y/N)
DMmedsInsCategoricalDiabetic Insulin Medication Status (Y/N)
DMmedTypeCategoricalDiabetes Medication Type
dmMedsSelfCategoricalDefined as Yes (Treated), if the participant reported being on diabetic
DMmedsCategoricalDiabetic Medication Status (Y/N)
statinMedsSelfCategoricalDefined as Yes (Treated), if the participant reported being on statin medication.
statinMedsCategoricalStatin Medication Status (Y/N)
hrtMedsSelfEverCategoricalSelf Reported HRT Medication Status (Y/N)
hrtMedsSelfCategoricalSelf Reported Current HRT Medication Status (Y/N)
hrtMedsCategoricalHRT Medication Status (Y/N)
betaBlkMedsCategoricalBeta Blocker Medication Status (Y/N)
calBlkMedsCategoricalCalcium Channel Blocker Medication Status (Y/N)
diureticMedsCategoricalDiuretic Medication Status (Y/N)
antiArythMedsSelfCategoricalDefined as Yes (Treated), if the participant reported being on antiarrhythmic medication.
antiArythMedsCategoricalAntiarrhythmic Medication Status (Y/N)
4. Hypertension
sbpContinuousSystolic Blood Pressure (mmHg)
dbpContinuousDiastolic Blood Pressure (mmHg)
BPjnc7CategoricalJNC 7 BP Classification
HTNCategoricalHypertension Status
ABIContinuousAnkle Brachial Index
5. Diabetes
FPGContinuousFasting Plasma Glucose Level (mg/dL)
FPG3catCategoricalFasting Plasma Glucose Categorization
HbA1cContinuousNGSP Hemoglobin HbA1c (%)
HbA1c3catCategoricalNGSP Hemoglobin HbA1c (%) Categorization
HbA1cIFCCContinuousIFCC Hemoglobin HbA1c in SI units (mmol/mol)
HbA1cIFCC3catCategoricalIFCC Hemoglobin HbA1c in SI units (mmol/mol) Categorization
fastingInsulinContinuousFasting Insulin (Plasma IU/mL)
HOMA-BContinuousHOMA-B
HOMA-IRContinuousHOMA-IR
DiabetesCategoricalDiabetes Status (ADA 2010)
diab3catCategoricalDiabetes Categorization
6. Lipids
ldlContinuousFasting LDL Cholesterol Level (mg/dL)
ldl5catCategoricalFasting LDL Categorization
hdlContinuousFasting HDL Cholesterol Level (mg/dL)
hdl3catCategoricalFasting HDL Categorization
trigsContinuousFasting Triglyceride Level (mg/dL)
trigs4catCategoricalFasting Triglyceride Categorization
totCholContinuousFasting Total Cholesterol (mg/dL)
7. Biomarkers
hsCRPContinuousHigh Sensitivity C-Reactive Protein
(Serum mg/dL)
endothelinContinuousEndothelin-1
(Serum pg/mL)
sCortContinuousConcentration of Cortisol Levels
(Serum µg/dL)
reninRIAContinuousRenin Activity RIA (Plasma ng/mL/hr)
reninIRMAContinuousRenin Mass IRMA (Plasma pg/mL)
aldosteroneContinuous“Concentration of Aldosterone
leptinContinuous(Serum ng/dL)”
adiponectinContinuousConcentration of Leptin (Serum ng/mL)
8. Renal
SCrCCContinuousCC Calibrated Serum Creatinine (mg/dL)
SCrIDMSContinuousIDMS Tracebale Serum Creatinine (mg/dL)
eGFRmdrdContinuouseGFR MDRD
eGFRckdepiContinuouseGFR CKD-Epi
CreatinineU24hrContinuous24-hour urine creatinine (g/24hr)
CreatinineUSpotContinuousRandom spot urine creatinine (mg/dL)
AlbuminUSpotContinuousRandom spot urine albumin (mg/dL)
AlbuminU24hrContinuous24-hour urine albumin (mg/24hr)
DialysisEverCategoricalSelf-reported dialysis
DialysisDurationContinuousSelf-reported duration on dialysis (years)
CKDHxCategoricalChronic Kidney Disease History
9. Respiratory
asthmaCategoricalPhysician-Diagnosed Asthma
maneuversContinuousSuccessful Spirometry Maneuvers
FVCContinuousForced Vital Capacity (L)
FEV1ContinuousForced Expiratory Volume in 1 s (L)
FEV6ContinuousForced Expiratory Volume in 6 s (L)
FEV1PPContinuousFEV1 % Predicted
FVCPPContinuousFVC % Predicted
10. Echocardiogram
LVMechoContinuousLeft Ventricular Mass (g) from Echo
LVMindexContinuousLeft Ventricular Mass Indexed by Height(m)^2.7
LVHCategoricalLeft Ventricular Hypertrophy
EFContinuousEjection Fraction
EF3catCategoricalEjection Fraction Categorization
DiastLVdiaContinuousDiastolic LV Diameter (mm)
SystLVdiaContinuousSystolic LV Diameter (mm)
FSCategoricalFractional Shortening
RWTContinuousRelative Wall Thickness
11. Electrocardiogram
ConductionDefectCategoricalConduction Defect
MajorScarAntCategoricalAnterior QnQs Major Scar
MinorScarAntCategoricalAnterior QnQs Minor Scar
RepolarAntCategoricalAnterior Repolarization Abnormality
MIAntCategoricalAnterior ECG defined MI
MajorScarPostCategoricalPosterior QnQs Major Scar
MinorScarPostCategoricalPosterior QnQs Minor Scar
RepolarPostCategoricalPosterior Repolarization Abnormality
MIPostCategoricalPosterior ECG defined MI
MajorScarAntLatCategoricalAnterolateral QnQs Major Scar
MinorScarAntLatCategoricalAnterolateral QnQs Minor Scar
RepolarAntLatCategoricalAnterolateral Repolarization Abnormality
MIAntLatCategoricalAnterolateral ECG defined MI
MIecgCategoricalECG determined MI
ecgHRContinuousHeart Rate (bpm)
AfibCategoricalAtrial Fibrillation
AflutterCategoricalAtrial Flutter
QRSContinuousQRS Interval (msec)
QTContinuousQT Interval (msec)
QTcFramContinuousFramingham Corrected QT Interval (msec)
QTcBazContinuousBazett Corrected QT Interval (msec)
QTcHodContinuousHodge Corrected QT Interval (msec)
QTcFridContinuousFridericia Corrected QT Interval (msec)
CVContinuousCornell Voltage (microvolts)
LVHcvCategoricalCornell Voltage Criteria
12. Stroke History
speechLossEverCategoricalHistory of Speech Loss
visionLossEverCategoricalHistory of Sudden Loss of Vision
doubleVisionEverCategoricalHistory of Double Vision
numbnessEverCategoricalHistory of Numbness
paralysisEverCategoricalHistory of Paralysis
dizzynessEverCategoricalHistory of Dizziness
strokeHxCategoricalHistory of Stroke
13. CVD History
MIHxCategoricalSelf-Reported History of MI
CardiacProcHxCategoricalSelf-Reported history of Cardiac Procedures
CHDHxCategoricalCoronary Heart Disease Status/History
CarotidAngioHxCategoricalSelf-Reported history of Carotid Angioplasty
CVDHxCategoricalCardiovascular Disease History
14. Healthcare Access
InsuredCategoricalVisit 1 Health Insurance Status
15. Psychosocial
IncomeCategoricalIncome Status
occupationCategoricalOccupational Status
edu3catCategoricalEducation Attainment Categorization
HSgradCategoricalHigh School Graduate
dailyDiscrContinuousEveryday Discrimination Experiences
lifetimeDiscrmContinuousMajor Life Events Discrimination
discrmBurdenContinuousDiscrimination Burden
depressionContinuousTotal Depressive Symptoms Score
weeklyStressContinuousTotal Weekly Stress Score
perceivedStressContinuousTotal Global Stress Score
16. Life’s Simple 7
SMK3catCategoricalAHA Smoking Categorization
idealHealthSMKCategoricalIndicator for Ideal Health via Smoking Status
BMI3catCategoricalAHA BMI Categorization
idealHealthBMICategoricalIndicator for Ideal Health via BMI
PA3catCategoricalAHA Physical Activity Categorization
idealHealthPACategoricalIndicator for Ideal Health via Physical Activity
nutrition3catCategoricalAHA Nutrition Categorization
idealHealthNutritionCategoricalIndicator for Ideal Health via Nutrition
totChol3catCategoricalAHA Total Cholesterol Categorization
idealHealthCholCategoricalIndicator for Ideal Health via Total Cholesterol
BP3catCategoricalAHA BP Categorization
idealHealthBPCategoricalIndicator for Ideal Health via BP
glucose3catCategoricalAHA Glucose Categorization
idealHealthDMCategoricalIndicator for Ideal Health via Glucose
17. Nutrition
vitaminD2Continuous25(OH) Vitamin D2 (ng/mL)
vitaminD3Continuous25(OH) Vitamin D3 (ng/mL)
vitaminD3epimerContinuousep-25(OH) Vitamin D3 (ng/mL)
darkgrnVegContinuousDark-green Vegetables
eggsContinuousEggs
fishContinuousFish
18. Physical Activity
sportIndexContinuousSport Index
hyIndexContinuousHome/Yard Index
activeIndexContinuousActive Living Index
19. Risk Scores
frs_chdtenyrriskContinuousFramingham Risk Score-Coronary Heart Disease
frs_cvdtenyrriskContinuousFramingham Risk Score-Cardiovascular Disease
frs_atpiii_tenyrriskContinuousFramingham Risk Score-Adult Treatment Panel (III)—Coronary Heart Disease
rrs_tenryriskContinuousReynolds Risk Score
ascvd_tenyrriskContinuousAmerican College of Cardiology—American Heart Association—Atherosclerotic Cardiovascular Disease

References

  1. Conrad, N.; Judge, A.; Tran, J.; Mohseni, H.; Hedgecott, D.; Crespillo, A.P.; Allison, M.; Hemingway, H.; Cleland, J.G.; McMurray, J.J.V.; et al. Temporal trends and patterns in heart failure incidence: A population-based study of 4 million individuals. Lancet 2018, 391, 572–580. [Google Scholar] [CrossRef] [Green Version]
  2. Roger, V.L. Epidemiology of heart failure. Circ. Res. 2013, 113, 646–659. [Google Scholar] [CrossRef] [PubMed]
  3. Bahrami, H.; Kronmal, R.; Bluemke, D.A.; Olson, J.; Shea, S.; Liu, K.; Burke, G.L.; Lima, J.A. Differences in the incidence of congestive heart failure by ethnicity: The multi-ethnic study of atherosclerosis. Arch. Intern. Med. 2008, 168, 2138–2145. [Google Scholar] [CrossRef] [PubMed]
  4. Bibbins-Domingo, K.; Pletcher, M.J.; Lin, F.; Vittinghoff, E.; Gardin, J.M.; Arynchyn, A.; Lewis, C.E.; Williams, O.D.; Hulley, S.B. Racial differences in incident heart failure among young adults. N. Engl. J. Med. 2009, 360, 1179–1190. [Google Scholar] [CrossRef] [Green Version]
  5. Langley, P. Elements of Machine Learning; Morgan Kaufmann: Burlington, MA, USA, 1996. [Google Scholar]
  6. Mitchell, T.M. Machine Learning; McGraw-Hill: New York, NY, USA, 1997. [Google Scholar]
  7. Leinweber, D.J. Stupid data miner tricks: Overfitting the S&P 500. J. Invest. 2007, 16, 15–22. [Google Scholar]
  8. Rodriguez, J.D.; Perez, A.; Lozano, J.A. Sensitivity analysis of k-fold cross validation in prediction error estimation. IEEE Trans. Pattern Anal. Mach. Intell. 2009, 32, 569–575. [Google Scholar] [CrossRef] [PubMed]
  9. Ho, T.K. Random decision forests. In Proceedings of the 3rd International Conference on Document Analysis and Recognition, Montreal, QC, Canada, 14–16 August 1995; pp. 278–282. [Google Scholar]
  10. Breiman, L.J. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
  11. Polikar, R. Ensemble based systems in decision making. IEEE Circuits Syst. Mag. 2006, 6, 21–45. [Google Scholar] [CrossRef]
  12. Breiman, L.; Friedman, J.; Stone, C.J.; Olshen, R.A. Classification and Regression Trees; CRC press: Boca Raton, FL, USA, 1984. [Google Scholar]
  13. Chen, T.; Guestrin, C. Xgboost: A scalable tree boosting system. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 785–794. [Google Scholar]
  14. Casanova, R.; Saldana, S.; Simpson, S.L.; Lacy, M.E.; Subauste, A.R.; Blackshear, C.; Wagenknecht, L.; Bertoni, A.G. Prediction of incident diabetes in the Jackson Heart Study using high-dimensional machine learning. PLoS ONE 2016, 11, e0163942. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Cover, T.; Hart, P. Nearest neighbor pattern classification. IEEE Trans. Inf. Theory 1967, 13, 21–27. [Google Scholar] [CrossRef]
  16. Jonsson, P.; Wohlin, C. An evaluation of k-nearest neighbour imputation using likert data. In Proceedings of the 10th International Symposium on Software Metrics, Chicago, IL, USA, 22 November 2004; pp. 108–118. [Google Scholar]
  17. Stekhoven, D.J.; Bühlmann, P. MissForest—non-parametric missing value imputation for mixed-type data. Bioinformatics 2012, 28, 112–118. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  18. Hosmer, D.W.; Lemeshow, S. Applied Logistic Regression; John Wiley & Sons: Hoboken, NJ, USA, 2000. [Google Scholar]
  19. Tibshirani, R. Regression shrinkage and selection via the lasso. J. R. Stat. Soc. Ser. B Methodol. 1996, 58, 267–288. [Google Scholar] [CrossRef]
  20. Boser, B.E.; Guyon, I.M.; Vapnik, V.N. A training algorithm for optimal margin classifiers. In Proceedings of the Fifth Annual Workshop on Computational Learning Theory (COLT’92), Pittsburgh, PA, USA, 27–29 July 1992; pp. 144–152. [Google Scholar]
  21. Bauters, C.; Lamblin, N.; Mc Fadden, E.P.; Van Belle, E.; Millaire, A.; De Groote, P. Influence of diabetes mellitus on heart failure risk and outcome. Cardiovasc. Diabetol. 2003, 2, 1. [Google Scholar] [CrossRef] [Green Version]
  22. Damman, K.; Van Veldhuisen, D.J.; Navis, G.; Vaidya, V.S.; Smilde, T.D.; Westenbrink, B.D.; Bonventre, J.V.; Voors, A.A.; Hillege, H.L. Tubular damage in chronic systolic heart failure is associated with reduced survival independent of glomerular filtration rate. Heart 2010, 96, 1297–1302. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  23. Metra, M.; Cotter, G.; Gheorghiade, M.; Dei Cas, L.; Voors, A.A. The role of the kidney in heart failure. Eur. Heart J. 2012, 33, 2135–2142. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. Echouffo-Tcheugui, J.B.; Greene, S.J.; Papadimitriou, L.; Zannad, F.; Yancy, C.W.; Gheorghiade, M.; Butler, J. Population risk prediction models for incident heart failure: A systematic review. Circ. Heart Fail. 2015, 8, 438–447. [Google Scholar] [CrossRef] [Green Version]
  25. Spinar, J.; Jarkovsky, J.; Spinarova, L.; Mebazaa, A.; Gayat, E.; Vitovec, J.; Linhart, A.; Widimsky, P.; Miklik, R.; Zeman, K.; et al. AHEAD score--Long-term risk classification in acute heart failure. Int. J. Cardiol. 2016, 202, 21–26. [Google Scholar] [CrossRef]
  26. Chen, Y.-J.; Sung, S.-H.; Cheng, H.-M.; Huang, W.-M.; Wu, C.-L.; Huang, C.-J.; Hsu, P.-F.; Yeh, J.-S.; Guo, C.-Y.; Yu, W.-C.; et al. Performance of AHEAD Score in an Asian Cohort of Acute Heart Failure With Either Preserved or Reduced Left Ventricular Systolic Function. J. Am. Heart Assoc. 2017, 6, e004297. [Google Scholar] [CrossRef]
  27. Guo, C.Y.; Chan, C.H.; Chou, Y.C.; Sung, S.H.; Cheng, H.M. A Statistical Predictive Model Consistent Within a 5-Year Follow-up Period for Patients with Acute Heart Failure. J. Chin. Med Assoc. 2020, in press. [Google Scholar] [CrossRef]
  28. Alotaibi, F.S. Implementation of Machine Learning Model to Predict Heart Failure Disease. Int. J. Adv. Comput. Sci. Appl. 2019, 10. [Google Scholar] [CrossRef]
  29. Swedberg, K. Heart failure subtypes: Pathophysiology and definitions. Diabetes Res. Clin. Pract. 2021, 13, 108815. [Google Scholar] [CrossRef]
  30. Gvozdanović, Z.; Farčić, N.; Šimić, H.; Buljanović, V.; Gvozdanović, L.; Katalinić, S.; Pačarić, S.; Gvozdanović, D.; Dujmić, Ž.; Miškić, B.; et al. The Impact of Education, COVID-19 and Risk Factors on the Quality of Life in Patients with Type 2 Diabetes. Int. J. Environ. Res. Public Health 2021, 27, 2332, PMCID:PMC7956830. [Google Scholar] [CrossRef] [PubMed]
  31. Segar, M.W.; Vaduganathan, M.; Patel, K.V.; McGuire, D.K.; Butler, J.; Fonarow, G.C.; Basit, M.; Kannan, V.; Grodin, J.L.; Everett, B. Machine Learning to Predict the Risk of Incident Heart Failure Hospitalization Among Patients With Diabetes: The WATCH-DM Risk Score. Diabetes Care 2019, 42, 2298–2306. [Google Scholar] [CrossRef]
  32. Wu, J.; Roy, J.; Stewart, W.F. Prediction modeling using EHR data: Challenges, strategies, and a comparison of machine learning approaches. Med Care 2010, 48 (Suppl. 6), S106–S113. [Google Scholar] [CrossRef] [PubMed]
  33. Fawcett, T. An introduction to ROC analysis. Pattern Recognit. Lett. 2006, 27, 861–874. [Google Scholar] [CrossRef]
  34. Youden, W.J. Index for rating diagnostic tests. Cancer 1950, 3, 32–35. [Google Scholar] [CrossRef]
  35. Rosner, B. Fundamentals of Biostatistics; Nelson Education: Boston, MA, USA, 2015. [Google Scholar]
  36. Hanley, J.A.; McNeil, B.J. The meaning and use of the area under a receiver operating characteristic (ROC) curve. Radiology 1982, 143, 29–36. [Google Scholar] [CrossRef] [Green Version]
  37. Mandrekar, J.N. Receiver operating characteristic curve in diagnostic test assessment. J. Thorac. Oncol. 2010, 5, 1315–1316. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  38. Chawla, N.V.; Bowyer, K.W.; Hall, L.O.; Kegelmeyer, W.P. SMOTE: Synthetic minority over-sampling technique. J. Artif. Intell. Res. 2002, 16, 321–357. [Google Scholar] [CrossRef]
Figure 1. Scenarios considered.
Figure 1. Scenarios considered.
Ijerph 18 04943 g001
Figure 2. Study flow chart.
Figure 2. Study flow chart.
Ijerph 18 04943 g002
Figure 3. The training process of XGBoost.
Figure 3. The training process of XGBoost.
Ijerph 18 04943 g003
Figure 4. AUC for complete cases.
Figure 4. AUC for complete cases.
Ijerph 18 04943 g004
Figure 5. AUC with simple imputation.
Figure 5. AUC with simple imputation.
Ijerph 18 04943 g005
Figure 6. AUC with KNN imputation.
Figure 6. AUC with KNN imputation.
Ijerph 18 04943 g006
Figure 7. AUC with RF imputation.
Figure 7. AUC with RF imputation.
Ijerph 18 04943 g007
Table 1. Descriptive statistics of the study population.
Table 1. Descriptive statistics of the study population.
Baseline
Characteristic
Total Population
n = 3327
Non-HF
n = 3081
(92.6%)
HF
n = 246
(7.4%)
Age54.96 (12.59)54.24 (12.37)63.91 (11.98)
BMI31.82 (7.2)31.72 (7.18)33.02 (7.31)
Waist100.83 (16.03)100.4 (15.93)106.9 (16.14)
High School Graduate2761 (83.26%)2607 (84.92%)154 (62.60%)
Gender
Male1228 (36.91%)1132 (36.74%)96 (39.02%)
Female2099 (63.09%)1949 (63.26%)150 (60.98%)
Current Smoker406 (12.31%)374 (12.25%)32 (13.11%)
Hypertension (HTN)1845 (55.47%)1644 (53.38%)201 (81.71%)
Diabetes Mellitus (DM)710 (21.5%)593 (19.39%)117 (47.95%)
Table 2. The selected parameters of the final HF prediction model for XGBoost using complete cases (feature importance is in the parenthesis).
Table 2. The selected parameters of the final HF prediction model for XGBoost using complete cases (feature importance is in the parenthesis).
<1%<3%<5%<10%<20%<30%<40%
age
(0.0923)
age
(0.0213)
DMmeds
(0.0492)
DMmeds
(0.0222)
dmMeds
(0.0178)
CVDHx
(0.0267)
frs_chdtenyrrisk
(0.0507)
DMmeds
(0.0647)
RepolarAntLat
(0.0189)
age
(0.0290)
Diabetes
(0.0192)
MIHx
(0.0174)
ascvd_tenyrrisk
(0.0260)
ALDOSTERONE (0.0367)
Diabetes
(0.0431)
DMmeds
(0.0187)
BP3cat
(0.0270)
CVDHx
(0.0189)
EF
(0.0170)
rrs_tenyrrisk (0.0220)eGFRmdrd (0.0352)
eGFRckdepi
(0.0315)
Diabetes
(0.0180)
HTN
(0.0267)
bpjnc7_3
(0.0187)
HbA1cIFCC (0.0150)numbnessEver
(0.0203)
occupation (0.0331)
MIecg
(0.0305)
CVDHx
(0.0174)
sbp
(0.0237)
CHDHx
(0.0176)
HbA1c
(0.0148)
nutrition3cat (0.0194)abi
(0.0317)
RepolarAntLat
(0.0297)
eGFRmdrd
(0.0173)
eGFRckdepi
(0.0233)
eGFRckdepi
(0.0174)
strokeHx
(0.0141)
FEV1PP
(0.0193)
sbp
(0.0297)
antiArythMeds
(0.028)
eGFRckdepi
(0.0173)
CVDHx
(0.0188)
FPG
(0.0170)
Diabetes
(0.0141)
totchol
(0.0186)
calBlkMeds (0.0292)
RepolarAnt
(0.0272)
statinMeds
(0.0168)
QTcFrid
(0.0187)
age
(0.0170)
visionLossEver
(0.0140)
asthma
(0.0180)
FVC
(0.0289)
statinMeds
(0.0268)
edu3cat
(0.0163)
BPmeds
(0.0184)
sbp
(0.0164)
statinMeds (0.0129)BPmeds
(0.0177)
eGFRckdepi (0.0253)
CVDHx
(0.0262)
CardiacProcHx
(0.0162)
waist
(0.01831)
waist
(0.0160)
frs_chdtenyrrisk
(0.0129)
HTN
(0.0177)
SCrCC
(0.0242)
Note: The above abbreviations are available in Table A1.
Table 3. The selected parameters of the final HF prediction model for XGBoost using simple imputations.
Table 3. The selected parameters of the final HF prediction model for XGBoost using simple imputations.
<1%<3%<5%<10%<20%<30%<40%
Diabetes
(0.0455)
DMmeds
(0.0346)
DMmeds
(0.0588)
DMmeds
(0.0233)
DMmeds
(0.0630)
DMmeds
(0.0210)
DMmeds
(0.0337)
DMmeds
(0.0439)
age
(0.0273)
Diabetes
(0.0327)
age
(0.0221)
age
(0.0383)
age
(0.0196)
DialysisEver (0.0252)
age
(0.0408)
CVDHx
(0.0265)
DialysisEver (0.0243)eGFRckdepi (0.0185)Diabetes
(0.0297)
Diabetes
(0.0158)
CVDHx
(0.0207)
HTN
(0.0336)
eGFRckdepi (0.0263)MIAntLat
(0.0191)
Diabetes
(0.0183)
DialysisEver (0.0190)BPmeds
(0.0150)
age
(0.0151)
CVDHx
(0.0277)
HTN
(0.0260)
age
(0.0189)
FEV1
(0.0163)
ConductionDefect (0.0183)CVDHx
(0.0147)
Afib
(0.0142)
HSgrad
(0.0273)
HSgrad
(0.0246)
HSgrad
(0.0157)
CVDHx
(0.0158)
sex
(0.0180)
age
(0.0145)
MIHx
(0.0136)
BPmeds
(0.0272)
Diabetes
(0.0237)
Afib
(0.0148)
FVC
(0.0156)
occupation (0.0159)ConductionDefect (0.0141)eGFRckdepi (0.0129)
eGFRckdepi
(0.0238)
eGFRmdrd (0.0225)edu3cat
(0.0138)
CHDHx
(0.0152)
MIant
(0.0143)
FEV1
(0.0141)
SystLVdia
(0.0128)
RepolarAntLat
(0.0229)
MIHx
(0.0212)
CVDHx
(0.0135)
eGFRmdrd (0.0151)CVDHx
(0.0129)
eGFRckdepi (0.0139)EF
(0.0117)
ecgHR
(0.0206)
sbp
(0.0191)
EF
(0.0130)
HbA1cIFCC (0.0148)idealHealthSMK (0.0125)CHDHx
(0.0126)
ConductionDefect (0.0115)
Table 4. The selected parameters of the final HF prediction model for XGBoost using KNN imputations.
Table 4. The selected parameters of the final HF prediction model for XGBoost using KNN imputations.
<1%<3%<5%<10%<20%<30%<40%
Diabetes
(0.0435)
age
(0.0320)
DMmeds
(0.0848)
DMmeds (0.0261)DMmeds (0.0443)DMmeds (0.0145)DMmeds (0.0318)
DMmeds
(0.0409)
DMmeds
(0.0266)
age
(0.0302)
age
(0.0177)
age
(0.0420)
age
(0.0142)
DialysisEver (0.0294)
age
(0.0386)
MIHx
(0.0234)
MIant
(0.0245)
CVDHx
(0.0175)
CVDHx
(0.0321)
eGFRmdrd (0.0138)age
(0.0234)
HTN
(0.0299)
Diabetes
(0.0213)
CVDHx
(0.0241)
eGFRckdepi (0.0173)EF
(0.0188)
Diabetes
(0.0130)
Diabetes
(0.0145)
CVDHx
(0.0277)
CVDHx
(0.0209)
Diabetes
(0.0236)
Diabetes
(0.0167)
eGFRckdepi (0.0182)SCrCC
(0.0122)
Afib
(0.0143)
BPmeds
(0.0274)
eGFRckdepi (0.0205)HSgrad
(0.0206)
eGFRmdrd (0.0153)FEV1
(0.0179)
eGFRckdepi (0.0118)CVDHx
(0.0128)
HSgrad
(0.0264)
HSgrad
(0.0176)
ConductionDefect (0.0201)FEV1
(0.0153)
ConductionDefect (0.0174)statinMeds (0.0115)eGFRckdepi (0.0126)
eGFRckdepi (0.0222)HTN
(0.0172)
antiArythMedsSelf
(0.0152)
CHDHx
(0.0152)
MIHx
(0.0172)
CVDHx
(0.0112)
MIHx
(0.0120)
RepolarAntLat (0.0209)antiArythMeds (0.0171)CHDHx
(0.1431)
edu3cat
(0.0142)
MajorScarAnt (0.0172)everSmoker (0.0111)calBlkMeds (0.0116)
ecgHR
(0.0196)
eGFRmdrd (0.0164)AntiArythMeds (0.1374)DialysisEver (0.0139)eGFRmdrd (0.0170)rrs_tenyrrisk (0.0108)FEV1
(0.0113)
Table 5. The selected parameters of the final HF prediction model for XGBoost using MissForest imputations.
Table 5. The selected parameters of the final HF prediction model for XGBoost using MissForest imputations.
<1%<3%<5%<10%<20%<30%<40%
Diabetes
(0.0455)
antiArythMeds
(0.0339)
dmMeds
(0.0340)
DMmeds (0.0263)DMmeds
(0.0443)
DMmeds
(0.0163)
DMmeds
(0.0290)
DMmeds
(0.0363)
DMmeds
(0.0332)
age
(0.0279)
age
(0.0251)
DialysisEver (0.0323)Diabetes
(0.0161)
ascvd_tenyrrisk (0.0255)
age
(0.0344)
age
(0.0301)
Diabetes
(0.0271)
Diabetes
(0.0234)
MIAntLat
(0.0241)
rrs_tenyrrisk (0.0148)age
(0.0218)
HTN
(0.0336)
eGFRckdepi
(0.0241)
eGFRckdepi (0.0269)CVDHx
(0.0218)
Diabetes
(0.0208)
age
(0.0132)
eGFRckdepi (0.0210)
CVDHx
(0.0318)
HTN
(0.0232)
CVDHx
(0.0235)
CHDHx
(0.0194)
age
(0.0194)
ascvd_tenyrrisk (0.0130)rrs_tenyrrisk (0.0191)
HSgrad
(0.0274)
SCrIDMS
(0.0220)
eGFRmdrd (0.0212)eGFRmdrd (0.0192)Afib
(0.0184)
MIant
(0.0127)
frs_cvdtenyrrisk (0.0179)
eGFRckdepi (0.0243)MIHx
(0.0207)
HSgrad
(0.0198)
eGFRckdepi (0.0162)calBlkMeds (0.0154)eGFRckdepi (0.0125)MIHx
(0.0162)
CHDHx
(0.0241)
CVDHx
(0.0198)
SCrIDMS
(0.0187)
HSgrad
(0.0162)
CVDHx
(0.0152)
CVDHx
(0.0118)
LEPTIN
(0.0147)
RepolarAntLat (0.0238)eGFRmdrd (0.0197)BPMeds
(0.0170)
FEV1
(0.0149)
eGFRckdepi (0.0149)FEV1
(0.0106)
calBlkMeds (0.0135)
QTcBaz
(0.0221)
Diabetes
(0.0195)
HbA1c
(0.0158)
SCrIDMS
(0.0148)
EF
(0.0140)
CHDHx
(0.0104)
CardiacProcHx (0.0127)
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Guo, C.-Y.; Wu, M.-Y.; Cheng, H.-M. The Comprehensive Machine Learning Analytics for Heart Failure. Int. J. Environ. Res. Public Health 2021, 18, 4943. https://doi.org/10.3390/ijerph18094943

AMA Style

Guo C-Y, Wu M-Y, Cheng H-M. The Comprehensive Machine Learning Analytics for Heart Failure. International Journal of Environmental Research and Public Health. 2021; 18(9):4943. https://doi.org/10.3390/ijerph18094943

Chicago/Turabian Style

Guo, Chao-Yu, Min-Yang Wu, and Hao-Min Cheng. 2021. "The Comprehensive Machine Learning Analytics for Heart Failure" International Journal of Environmental Research and Public Health 18, no. 9: 4943. https://doi.org/10.3390/ijerph18094943

APA Style

Guo, C. -Y., Wu, M. -Y., & Cheng, H. -M. (2021). The Comprehensive Machine Learning Analytics for Heart Failure. International Journal of Environmental Research and Public Health, 18(9), 4943. https://doi.org/10.3390/ijerph18094943

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop