Next Article in Journal
Generating Daily Soil Moisture at 16 m Spatial Resolution Using a Spatiotemporal Fusion Model and Modified Perpendicular Drought Index
Next Article in Special Issue
Machine Learning Models for Weight-Bearing Activity Type Recognition Based on Accelerometry in Postmenopausal Women
Previous Article in Journal
The Time-of-Arrival Offset Estimation in Neural Network Atomic Denoising in Wireless Location
Previous Article in Special Issue
Data-Driven Machine-Learning Methods for Diabetes Risk Prediction
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Machine Learning Methods for Hypercholesterolemia Long-Term Risk Prediction

Department of Computer Engineering and Informatics, University of Patras, 26504 Patras, Greece
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(14), 5365; https://doi.org/10.3390/s22145365
Submission received: 19 June 2022 / Revised: 12 July 2022 / Accepted: 16 July 2022 / Published: 18 July 2022
(This article belongs to the Special Issue Machine Learning for Biomedical Sensing and Healthcare)

Abstract

:
Cholesterol is a waxy substance found in blood lipids. Its role in the human body is helpful in the process of producing new cells as long as it is at a healthy level. When cholesterol exceeds the permissible limits, it works the opposite, causing serious heart health problems. When a person has high cholesterol (hypercholesterolemia), the blood vessels are blocked by fats, and thus, circulation through the arteries becomes difficult. The heart does not receive the oxygen it needs, and the risk of heart attack increases. Nowadays, machine learning (ML) has gained special interest from physicians, medical centers and healthcare providers due to its key capabilities in health-related issues, such as risk prediction, prognosis, treatment and management of various conditions. In this article, a supervised ML methodology is outlined whose main objective is to create risk prediction tools with high efficiency for hypercholesterolemia occurrence. Specifically, a data understanding analysis is conducted to explore the features association and importance to hypercholesterolemia. These factors are utilized to train and test several ML models to find the most efficient for our purpose. For the evaluation of the ML models, precision, recall, accuracy, F-measure, and AUC metrics have been taken into consideration. The derived results highlighted Soft Voting with Rotation and Random Forest trees as base models, which achieved better performance in comparison to the other models with an AUC of 94.5%, precision of 92%, recall of 91.8%, F-measure of 91.7% and an accuracy equal to 91.75%.

1. Introduction

Cholesterol is a form of fat and a key component of cells. It plays a very important role in health as it participates in the synthesis of hormones, in the production of vitamin D and in the digestion and assimilation of fats. The molecules that result from the binding of cholesterol to proteins are called lipoproteins and are categorized into “bad” LDL cholesterol and “good” HDL cholesterol. LDL cholesterol is responsible for transporting cholesterol molecules from the liver to tissues and organs, while HDL cholesterol transports cholesterol molecules from tissues back to the liver [1,2].
Cholesterol is calculated in milligrams (mg) of cholesterol per deciliter (dL) of blood. An effect below 200 mg/dL (5.2 mmol/L) is ideal. A level somewhere between 200 and 239 mg/dL (5.2–6.2 mmol/L) is marginally below the high-risk number. A value above 240 mg/dL (6.3 mmol/L) is the high-risk limit. Total cholesterol results from the sum of HDL and LDL values [3].
In HDL, a value below 40 mg/dL (1 mmol/L) for men and 50 mg/dL (1.3 mmol /L) for women is low. This increases the risk of cardiovascular problems. A normal HDL cholesterol level is between 40 and 49 mg/dL (1–1.3 mmol/L) for men. For women, it is between 50 and 59 mg/dL (1.3–1.5 mmol/L). When this level is higher than 60 mg/dL (1.6 mmol/L), there is increased defense against coronary heart disease [4].
In LDL cholesterol, a value below 100 mg/dL (2.6 mmol/L) is ideal. The value between 100 and 129 mg/dL (2.6 and 3.3 mmol/L) is close to ideal, while between 130 and 159 mg/dL (3.4 and 4.1 mmol/L) it is marginally increased. LDL between 160 and 189 mg/dL (4.1 and 4.9 mmol/L) is considered high. When it exceeds 190 mg/dL (4.9 mmol/L) it is very high. It is recommended that LDL cholesterol be below 70 mg/dL (1.8 mmol/L) [4].
Although cholesterol is essential for the human body, high levels in the blood are associated with vascular damage and cardiovascular disease. Sometimes, our body produces more cholesterol than it needs, and this excess circulates in the blood. High blood cholesterol levels can cause blood vessels to clot and increase the risk of atherosclerotic plaque, coronary heart disease, angina, heart attack, peripheral arterial disease and stroke [5,6,7].
It should be noted that a high level of cholesterol is estimated to cause 2.6 million deaths (4.5% of total) and 29.7 million disability-adjusted life years (DALYS), or 2% of total DALYS. Raised total cholesterol is a major cause of disease burden in both the developed and developing world [8].
High cholesterol levels (hypercholesterolemia) may be due to lifestyle, genes (heredity) or, secondarily, to health conditions such as kidney disease. Factors responsible for the increase in cholesterol are poor diet, lack of physical activity, smoking, certain medications as well as pathological conditions such as diabetes, obesity, chronic kidney disease, HIV, hypothyroidism and polycystic ovary syndrome [9,10].
Diet plays an important role in improving cholesterol levels. High blood cholesterol levels are significantly reduced with lower consumption of fatty foods (meat, dairy, cold cuts), increased fibre intake (fruits, legumes) and frequent consumption of fatty fish rich in omega-3 fatty acids (sardines, mackerel). In addition, daily physical exercise will have beneficial effects in lowering cholesterol [11].
Information and communication technologies (ICTs), and especially the fields of artificial intelligence (AI) and machine learning (ML), are moving in this direction. ML techniques now play an important role in the early diagnosis of various diseases, such as diabetes (as classification [12] or regression task for continuous glucose prediction [13,14]), hypertension [15], COPD [16], COVID-19 [17], CVDs [18], stroke [19], CKD [20], ALF [21], hepatitis [22], sleep disorders [23], cancer [24], etc.
Especially, in this study, ML models are explored to estimate the long-term risk of hypercholesterolemia occurrence with the aid of various risk factors. The main contribution of this article is a comparative evaluation of several supervised learning classifiers to find the one with the highest sensitivity and separability, which means that it is the most appropriate to correctly identify those at high risk. An essential aspect of the data mining process is data pre-processing in the context of which data cleaning, features selection and class balancing were applied. Several performance metrics are utilized to evaluate the classifiers’ performance, such as precision, recall, F-measure, accuracy and AUC. Performance analysis showed that data quality has a significant impact on the training of efficient models. Finally, the quantitative analysis demonstrated that the soft voting is the most competent model, and thus it constitutes the main suggestion of this study.
The rest of the paper is organized as follows. Section 2 describes the relevant works with the subject under consideration. In Section 3, a dataset description and analysis of the methodology followed is made. In addition, in Section 4, we discuss the acquired research results. Finally, conclusions and future directions are outlined in Section 5.

2. Related Work

The estimation of individual risk for the development of a chronic condition has gained high popularity in the medical field. Therefore, in predictive analytics and, especially machine learning, numerous studies have been conducted to estimate personal risk using various data related to socioeconomic features (age group, gender and race), behavioural data and, recently, clinical risk factors. In this direction, we will present some recent studies’ outcomes that use machine learning techniques to predict hypercholesterolemia.
Familial hypercholesterolemia (FH) is a dominant genetic condition with an increased risk of coronary artery disease (in untreated cases) [25]. Machine learning-based strategies can lead to the effective identification of high-risk patients to enhance FH management.
First, the authors in [26] selected three machine learning algorithms including a classification tree, a gradient boosting machine and a neural network to predict the presence of FH in two different cohorts. The evaluation was based on the area under the ROC curve. The findings have shown the superiority of machine learning models against the clinical Dutch Lipid Score in predicting carriers of FH-causative mutations.
In [27], a Random Forest classifier was developed to identify potential FH patients using electronic health record (EHR) data. The model was trained on 197 known patients and 6590 without FH, achieving a positive predictive value (PPV) of 0.88 and a sensitivity of 0.75 on a hold-out test set. The accuracy of the classifier’s predictions was further evaluated by a chart review of patients at risk of FH not included in the original dataset. The classifier correctly categorized 84% of patients at the highest probability threshold. Finally, the same classifier was validated on an external dataset from the Geisinger Healthcare System and achieved a PPV of 0.85.
Moreover, in [28], the authors developed a model for predicting hypercholesterolemia using a comprehensive set of body fat mass variables based on machine learning techniques, in addition to studying the correlation between body fat mass and hypercholesterolemia. They obtained the area under the receiver operating characteristic curve value of 0.739 and the Matthews correlation coefficient value of 0.36 in the model using the correlation-based feature subset selection and the Naive Bayes algorithm.
A machine learning approach for the prediction of cholesterol levels via regression using non-invasive and easy-to-collect data (clinical and anthropometric) is presented in [29]. In addition, clustering analysis is carried out to identify different groups of patients sharing some characteristics and give valuable information to clinical experts for diagnosis or prognosis.
Moreover, the authors in [30] aimed to compare the performance of various machine learning models to predict the prevalence of hypercholesterolemia associated with exposure to lead, mercury and cadmium. Five machine learning models, such as Logistic Regression, K-Nearest Neighbors, Decision Trees, Random Forest, and Support Vector Machines were constructed, and their predictive performance were compared. Finally, the Support Vector Machine model was the most accurate, and the logistic regression model had the highest area under the ROC curve of 0.718 (95% CI: 0.688–0.748).
In [31], the authors developed a high accuracy (97.45%) convolutional neural network-based Android application that determines cholesterol levels in a person’s body by capturing the image of the iris. A user with high cholesterol levels has a white–greyish circle on the outer circle of the iris.
Finally, the authors in [32] constructed a dataset based on the ELSA database, aiming at the prognosis of high cholesterol (hypercholesterolemia), targeting the elderly office workers. Naive Bayes, Support Vector Machines, Artificial Neural Network using two hidden layers, 5-Nearest Neighbors, Rotation Forest, Decision Trees, Logistic Model Trees and Random Forest were applied on the constructed dataset using a 10-fold cross-validation experimentation setup. The best overall performance was obtained with the Logistic Model Trees model, which performed best both concerning accuracy and recall metrics.

3. Materials and Methods

This section describes the dataset under consideration and the methodology adopted for determining the risk of being diagnosed with hypercholesterolemia.

3.1. Dataset Description

Our experimental results were based on the dataset of research work [32], which is derived from the English Longitudinal Study of Aging (ELSA) [33]. The initial features set included 106 variables, with 61 being nominal and 45 numerical attributes, and all the participants were over 50 years old. From the features list, we excluded socioeconomic data, including the type of employment, education, income, residence type, marital status, the number of children and insurance type. Moreover, after data cleaning and feature selection, the final list of features was reduced to 13, the number of participants to 350, and all the attributes (13 as input to ML models and 1 for target class) are described as follows:
  • Age (years) [34]: This feature refers to the age of a person who is over 50 years old. It is numerical data.
  • Gender [34]: This feature refers to a person’s gender. The number of men is 172 (49.15%), while the number of women is 178 (50.85%) It is nominal data.
  • BMI (Kg/m2) [35]: This feature captures the participant’s body mass index. It is numerical data.
  • Waist (cm) [36]: It is the measurement taken around the abdomen at the level of the umbilicus. It is numerical data.
  • SBP (mmHg) [37]: This feature captures the participant’s systolic blood pressure. It is numerical data.
  • DBP (mmHg) [37]: This feature captures the participant’s diastolic blood pressure. It is numerical data.
  • Hypertension [38]: This feature refers to whether a participant is hypertensive or not. The percentage of participants who have hypertension is 58.9%. It is nominal data.
  • HDL (mg/dL) [2]: This feature captures the participant’s high-density lipoprotein. It is numerical data.
  • LDL (mg/dL) [2]: This feature captures the participant’s low-density lipoprotein. It is numerical data.
  • TotChol (mg/dL) [2]: This feature captures the participant’s total cholesterol. It is numerical data.
  • Physical Activity [39]: This feature captures the participant’s physical activity and has 4 categories (high 2.6%, medium 11.2%, low 55.4% and very low 30.8%). It is nominal data.
  • Alcohol Consumption [40]: This feature refers to whether this participant consumes alcohol or not. The percentage of participants who consume alcohol more than normal is 44.1%. It is nominal data.
  • Diabetes [41]: This feature refers to whether the participant has been diagnosed with diabetes or not. The percentage of participants who suffer from diabetes is 20.6%. It is nominal data.
  • Hypercholesterolemia: This feature stands for whether the participant has been diagnosed with hypercholesterolemia. The percentage of participants who have been diagnosed with hypercholesterolemia is 44.6%. It is nominal data.

3.2. Hypercholesterolemia Risk Prediction

Supervised machine learning models have become an important asset for clinicians and healthcare providers as they allow them to evaluate the long-term risk of a condition occurrence based on several risk factors. More specifically, here, our purpose is to formulate a binary classification problem with target class c = “HyperChol” (hypercholesterolemia occurrence) or c = “Non-HyperChol” (non-occurrence of the hypercholesterolemia) and design models which will achieve high recall and area under curve (AUC) to ensure that instances with hypercholesterolemia can be accurately classified.
Assuming an instance with an unknown class label (HyperChol, Non-HyperChol), the trained ML models will predict its class based on the features’ values and thus the risk of occurring hypercholesterolemia in the long term.
The proposed methodology includes some specific steps, namely, data preprocessing, feature ranking, classification models training and performance evaluation.

3.2.1. Data Preprocessing

Data quality is a prerequisite for the development of efficient models suitable for the correct identification of healthy and with hypercholesterolemia instances. Hence, to ensure data validity, several preprocessing steps are usually applied. Data quality is ensured via the application of data cleaning methods, selecting the most appropriate for the data under consideration, such as excluding unnecessary or duplicate values, avoiding typos, handling missing values, data imputation, etc. [42]. As for the current dataset, we selected to remove instances whose feature values are missing and not valid (namely, out of the normal limits).
Moreover, the skewed class distribution constitutes a factor that can create ML models of poor performance. For this purpose, the imbalanced distribution of participants among the HyperChol and Non-HyperChol classes was tackled by employing SMOTE [43]. SMOTE uses a 5-NN classifier to create synthetic data on a minority class, i.e., HyperChol, which is oversampled such that the instances in two classes are equally distributed (i.e., 50%–50%). In Table 1, we present the minimum, maximum, mean and standard deviation of the numerical features in the balanced data.

3.2.2. Features Ranking

We employed four ranking methods to understand the importance of a feature in the target class. First, we applied the Pearson correlation coefficient [44] to evaluate the strength of association between all features and especially the worth of a feature in predicting the hypercholesterolemia class. Figure 1 demonstrates the outcomes of the correlation analysis. In the correlation matrix, we observe that the highest linear correlation of 0.92 is captured between total cholesterol levels and HDL, and total cholesterol levels and LDL. The next but still high linear relationship of rank 0.80 is noted between waist and BMI features, a high association of 0.75 is shown between hypertension and systolic blood pressure, while hypertension has a low association of 0.18 with the diastolic blood pressure. A moderate positive association of 0.3 is shown between age and systolic blood pressure. However, diastolic blood pressure has a moderate negative relationship with age. Finally, the target class records a moderate association of 0.68 and 0.62 with the total cholesterol and HDL, respectively, while a low relationship seems to exist with the other features.
Then, we applied the Information Gain method (InfoGain) [45] which evaluates the worth of a feature by measuring the information gain with respect to the class, according to the formula
I n f o G a i n ( c , y i j ) = H ( c ) H ( c y i j ) , j 1 , 2 , , n
where H ( c ) and H ( c y i j ) are the entropy of the hypercholesterolemia class and the conditional entropy of the class given the feature j  y i j of an instance i. From (1), we see that this measure captures the difference in entropy before and after the split of a feature set based on a specific y i j . Alternatively, it shows the uncertainty reduction after splitting the set on a feature. The best feature for splitting is the one with the highest information gain.
Furthermore, we employed the Gain Ratio (GR) method [46] which is calculated as G R ( y i j ) = H ( c ) H ( c y i j ) H ( y i j ) , where H ( y i j ) = p y i j l o g 2 ( p y i j ) is the entropy of an instance with feature y i j (with p y i j denoting the probability of selecting feature y i j ), H ( c ) = p c l o g 2 ( p c ) is the entropy of class c (with p c be the probability of selecting an instance in class c) and H ( c y i j ) being the conditional entropy of feature y i j given class c. Gain ratio indicates the relevance of a feature and selects the ones that maximize gain ratio based on the probability of each feature value.
Finally, the Random Forest classifier was selected to measure the importance of the features. Random Forest creates a forest of trees and per tree measures the discrimination ability of a potential feature to create the optimal split, namely the one that separates the instances of the two classes, using the Gini impurity.
In Table 2, we summarize features’ importance in the balanced dataset concerning the hypercholesterolemia class. All considered methods show that TotChol and HDL features are of the highest importance for the prediction of hypercholesterolemia. In addition, we observe that DiasBP is the next most important for the three out of four methods. Moreover, InfoGain and Gain ratio, due to their relationship, assign the features in the same order except for physical activity and hypertension. Since all features are among the risk factors that are utilized by the clinicians for the diagnosis and management of hypercholesterolemia, the models’ training and validation will be based on all of them. Finally, given that ML models can be retrained with more data, their importance will be re-investigated.

3.3. Data Exploration

In this subsection, after data preprocessing and features ranking, we will analyze the association between features and the HyperChol class.
Initially, in Figure 2, we present the participants’ distribution according to their age group and their gender. Most of them who have HyperChol belong to the age group 60–64, where men’s and women’s percentages are approximately similar. In addition, from this figure, we observe that HyperChol prevails in women who are between 50 and 59 and men who are older than 65 years old.
In Figure 3, we present the participants’ distribution in terms of the BMI and waist categories. To define the BMI categories, we used the following rules [13]:
(1)
B M I < 18.5 : underweight
(2)
18.5 B M I < 25 : healthy
(3)
25 B M I < 30 : overweight
(4)
B M I 30.0 : obesity
(a)
Class I: 30 B M I < 35
(b)
Class II: 35 B M I < 40
(c)
Class III: B M I 40 (severe obesity).
Cut-off points for the waist size were considered 88 cm for women and 102 cm for men [47]. Based on these points, the labels F88 and M102 capture women and men with waist circumference higher than 88 and 102 cm, respectively. The label “normal” refers to men and women with waist sizes lower than 88 and 102 cm, correspondingly. From this figure, we see that most of the participants with HyperChol are distributed in healthy and overweight and obese I categories of BMI. Moreover, when HyperChol and overweight classes coexist, the women with waists higher than the cut-off point are much more than men. In addition, some instances have HyperChol and are overweight, but they have normal waist size.
Furthermore, in Figure 4, we capture the coexistence of hypertension and diabetes diseases in relation to HyperChol. We see that 25% of the participants have HyperChol, are hypertensive and have not been diagnosed with diabetes. A small percentage of 6% fulfills all criteria; 16% of the participants who have HyperChol do not suffer from hypertension and diabetes.
Figure 5 and Figure 6 show the association of HyperChol with participants’ habits in terms of alcohol consumption and physical activity. It is shown that the HyperChol participants are roughly the same distributed among the two possible values of the alcohol consumption feature. Finally, as for Figure 6, it should be noted that most of the HyperChol participants are very low or low physically active.

3.4. Machine Learning Models

Several models have been selected to evaluate their prediction performance. More specifically, Naive Bayes, Support Vector Machine, Artificial Neural Network, K-NN, Rotation Forests, Decision Trees, Logistic Model Trees, Random Forest, and Ensemble Learning (Stacking and Soft Voting) classification methods will be presented. We assume that each instance i in the dataset is represented by a features vector y i = y i 1 , y i 2 , y i 3 , , y i n T , where n is the number of the features.

3.4.1. Naive Bayes

Naive Bayes(NB) classifies an instance y i at that class c for which P c y i 1 , , y i n ) is maximized (under the assumption that the features are highly independent [48]). The conditional probability is defined as P c y i 1 , , y i n ) = P y i 1 , , y i n | c P c P y i 1 , , y i n , where P y i 1 , , y i n | c = j = 1 n P y i j | c is the features probability given class and P y i 1 , , y i n , P ( c ) are the prior probability of features and class, respectively. The estimated class is derived by maximizing P c j = 1 n P y i j | c , where c { H y p e r C h o l , N o n H y p e r C h o l } .

3.4.2. K-Nearest Neighbors

K-Nearest Neighbors (K-NN) is a non-parametric lazy learning classifier which measures the distance (i.e., via Euclidean, Manhattan methods) between the test instance and every other instance in the training dataset [49]. Then it determines the K instances that are closest to the test instance which are finally categorized into the class that most of its K neighbors stem from.

3.4.3. Logistic Regression

Logistic Regression (LR) [50] is a supervised classifier for binary and multinomial tasks. It uses a logistic or a sigmoid function to model the dependent output variable. The model output is dichotomous in nature, i.e., with two possible classes, in which p captures the probability of an instance to belong in the HyperChol class; thus, 1 p is the probability of an instance belonging to the Non-HyperChol class. The relationship of log-odds with base b and model parameters β i is written as:
l o g b p 1 p = β 0 + β 1 y i 1 + + β n y i n

3.4.4. Rotation Forest

The Rotation Forest (RotF) [51] applies a rotation transformation matrix to the training before the training of each decision tree to increase the diversity of individual decision trees. A feature reduction technique creates a new feature set for every classifier in the ensemble. It randomly splits the feature set into subsets and applies principal component analysis (PCA) to every created subset separately. Then, a new feature set is acquired by combining principle components of each subset. In this study, the base classifier for the RotF is a J48 [52] decision tree.

3.4.5. Artificial Neural Network

Multilayer Perceptron (MLP) is the simplest fully connected feed-forward Neural Network. It consists of input and output layers and at least one hidden layer. Its neurons are trained by employing back-propagation learning which allows for classification into multiple labels. The MLP is able to learn non-linear models and execute online learning. Finally, it can use any arbitrary activation function [53].

3.4.6. Support Vector Machine

Support Vector Machine (SVM) [54] finds the hyperplane that can optimally separate instances into two classes. The most characteristic kernel functions are linear, polynomial, radial basis and quadratic. An instance x i can be optimally classified based on function:
f ( y ) = S g n i = 1 M α i c i K ( y i , y ) + b 0 α i C , α i c i = 0 , α i 0 , i = 1 , 2 , , M
where M is the size of training instances, y i , c i are the training instance feature vector and its class label, respectively, b is a bias, c i { 1 , 1 } , K ( y i , y ) is the kernel function which corresponds the input vectors into an expanded feature space and f ( y ) { 1 , + 1 } is the kernelized binary classifier’s predicted class for the unlabeled instance y .

3.4.7. Decision Tree

From the available Decision Trees, we considered the Reduced Error Pruning Tree (RepTree) [55]. It is a simple and fast decision learner which builds a decision/regression tree using information gain as an impurity measure and prunes it using reduced-error pruning. RepTree is even more accurate when dealing with a large volume of data.

3.4.8. Logistic Model Tree

A Logistic Model Tree (LMT) [56] consists of a standard decision tree structure with logistic regression functions f ( y i ) = β 0 + j = 1 n ( β i y i j ) at the leaves. LMTree constructs the tree growing process using the LogitBoost algorithm and the tree pruning is performed using Classification And Regression Tree (CART).

3.4.9. Random Forest

Random Forest (RF) is a bootstrapping technique based on a decision tree with high-performance outcomes, in both regression and classification tasks. It considers the Information Gain or Gini index to find the optimal subset of features, trains multiple decision trees and then classifies an instance by applying majority voting on the results of multiple Decision Trees [57].

3.4.10. Ensemble Learning

Ensemble Learning is a machine learning method that combines the outcomes of several single classifiers called base models. Voting and Stacking are two common approaches which are utilized to acquire more accurate predictions than the single models’. Concerning Voting, there are two types, Soft and Hard. Soft Voting, which is exploited in this study, averages the corresponding probabilities of the considered base classifiers and assigns a test instance to the class with the highest probability [58]. On the other hand, Stacking uses the predicted class labels of the base models as input features to train a meta-classifier which undertakes to find the class label [59]. Figure 7 illustrates the two schemes as they will be evaluated in the experiments.

3.5. Evaluation Metrics

To assess the ML models’ performance, accuracy, precision, recall, F-measure, and AUC metrics were considered [60].
Precision (or positive predicted value) shows the ratio of positive instances in relation to true and false positive instances. We also considered recall which captures the true positive rate or a model’s sensitivity to identify the participants who actually had HyperChol and correctly considered as positive, concerning all positive participants. Precision is a measure of quality, while recall is a measure of quantity. F-measure is the harmonic mean of precision and recall and allows the evaluation of a model using a single score. Moreover, we computed accuracy, which shows a model’s ability to correctly identify both HyperChol (positive) and Non-HyperChol (negative) instances. Taking into account the confusion matrix, TP, TN, FP and FN denote the number of true positive, true negative, false positive and false negative instances. Based on these quantities, the performance metrics are written as follows:
P r e c i s i o n = T P T P + F P , R e c a l l = T P T P + F N ,
F M e a s u r e = 2 × P r e c i s i o n · R e c a l l P r e c i s i o n + R e c a l l , A c c u r a c y = T N + T P T N + T P + F N + F P
Another important metric which will be taken into consideration for the models’ evaluation is area under curve (AUC) which takes values in the range [ 0 , 1 ] . AUC is a measure of separability. The ML models’ performance in distinguishing HyperChol from Non-HyperChol instances is captured by AUC. If AUC attains one, it means that the models have the perfect discrimination ability of the two classes distributions.

4. Results and Discussion

4.1. Experiments Setup

In this section, the ML models performance is evaluated in the WEKA 3.8.6 environment [61]. WEKA is a free JAVA-based data mining tool created and distributed under the GNU General Public License. It provides a library of various models for data preprocessing, classification, clustering, forecasting, visualization, etc. The computing system in which the experiments were conducted has the following characteristics: Intel(R) Core(TM) i7-9750H CPU @ 2.60 GHz 2.59 GHz 16 GB Memory, Windows 10 Home, 64-bit Operating System, x64-based processor. For our experiments, 10-fold cross-validation was applied to measure the models’ efficiency in the balanced dataset of 388 instances. In Table 3, the settings of the considered ML models are shown.

4.2. Evaluation

To fully evaluate the effectiveness of models, we should examine both precision and recall since the improvement of precision typically reduces recall and vice versa. However, if the classes’ distribution is uniform, these metrics may achieve the same outcomes.
In Table 4, we show the performance of various models, which were obtained after the application of data cleaning and class balancing. Selecting the two best-performing single classifiers, we combined them under two schemes, Soft Voting and Stacking. All models’ accuracy is higher than 86% except for the 3-NN model, which still achieved an acceptable accuracy of 70.62%. In addition, precision, AUC, recall and F-measure demonstrate the same promising outcomes as the accuracy.
Soft Voting (SoV) is the model with consistently high efficiency in all metrics. Focusing on AUC, which aggregates the classification performance of a model, the SoV model is able to distinguish between HyperChol class and Non-HyperChol class with a chance of 94.5%. In addition, we see that the combination of rotation with the random forest improved the performance of individual models and especially achieved a higher upgrade in the case of RF than in the RotF. Stacking is the second model with excellent separation performance, although its AUC is 8%, 6% and 2% lower than the ones of SoV, RF and RotF, respectively. SoV’s performance superiority lies in the fact that the base-models have been configured to predict probabilities instead of class labels. In the case of Stacking, class labels may add higher uncertainty to the predictions which are provided in the meta-model to learn how to find the best combination of them.
In Table 5, we capture recall and accuracy metrics comparing only single classifiers before [32] and after data cleaning and further restricting features number. A significant performance improvement is observed revealing the role of data quality, class balancing and dimensionality in the classifiers’ performance.
In addition, we observe that our proposed models are superior to the work [32] in terms of recall and accuracy of at least 10%. The only case where we can see similar performance is in the recall of the 5-NN. It should be mentioned that the 3-NN (Table 4) showed better accuracy and recall than the 5-NN with a percentage gap of about 3%.
In conclusion, the performance of Stacking outperforms the prediction performance of each individual model and is closer to the best single model RotF. Soft Voting, which is a probabilistic scheme, achieves higher performance than Stacking since in the meta-level of the latter, the predicted classes are combined using a logistic regression classifier to achieve an output that is a simple linear combination of the predictions of the sub-models. In either case, ensemble schemes indicated promising efficiency compared to the rest of the classifiers.

5. Conclusions

In this study, we exploited supervised learning to develop models for the identification of individuals at risk for hypercholesterolemia manifestation based on several risk factors. Healthcare professionals and clinical experts can benefit from such models to prevent the severe consequences of hypercholesterolemia, such as cardiovascular disease. Data exploration through risk factors analysis can help identify associations among the features and HyperChol. A critical aspect of ML models is that they allow medical experts to regularly reassess the associated risk and give proper guidelines and interventions for its management and treatment or prevent its occurrence.
Performance analysis revealed that data preprocessing is an important step for the design of efficient and accurate models for hypercholesterolemia occurrence. The experimental results showed that Soft Voting having as base classifiers the Random and Rotation Forest prevailed with an AUC of 94.5%, precision of 92%, recall of 91.8%, F-measure of 91.7% and an accuracy equal to 91.75%. Hence, it constitutes a candidate HyperChol risk prediction model.
The future purpose of this study is to extend the ML framework via the employment of deep learning methods by applying the Long Short-Term-Memory (LSTM) algorithm and Convolutional Neural Networks (CNN) in the same data comparing the results in terms of accuracy.

Author Contributions

E.D. and M.T. conceived of the idea, designed and performed the experiments, analyzed the results, drafted the initial manuscript and revised the final manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Steinberg, D.; Parthasarathy, S.; Carew, T.E.; Khoo, J.C.; Witztum, J.L. Beyond cholesterol. N. Engl. J. Med. 1989, 320, 915–924. [Google Scholar] [PubMed]
  2. Vincent, M.J.; Allen, B.; Palacios, O.M.; Haber, L.T.; Maki, K.C. Meta-regression analysis of the effects of dietary cholesterol intake on LDL and HDL cholesterol. Am. J. Clin. Nutr. 2019, 109, 7–16. [Google Scholar] [CrossRef] [PubMed]
  3. Ford, E.S.; Mokdad, A.H.; Giles, W.H.; Mensah, G.A. Serum total cholesterol concentrations and awareness, treatment, and control of hypercholesterolemia among US adults: Findings from the National Health and Nutrition Examination Survey, 1999 to 2000. Circulation 2003, 107, 2185–2189. [Google Scholar] [CrossRef] [PubMed]
  4. Larson, D.E. Mayo Clinic Family Health Book; W. Morrow: New York, NY, USA, 1990. [Google Scholar]
  5. Gidding, S.S.; Allen, N.B. Cholesterol and atherosclerotic cardiovascular disease: A lifelong problem. J. Am. Heart Assoc. 2019, 8, e012924. [Google Scholar] [CrossRef]
  6. Mohler, E.R., III; Hiatt, W.R.; Creager, M.A. Cholesterol reduction with atorvastatin improves walking distance in patients with peripheral arterial disease. Circulation 2003, 108, 1481–1486. [Google Scholar] [CrossRef] [Green Version]
  7. Tirschwell, D.; Smith, N.; Heckbert, S.; Lemaitre, R.; Longstreth, W.; Psaty, B. Association of cholesterol with stroke risk varies in stroke subtypes and patient subgroups. Neurology 2004, 63, 1868–1875. [Google Scholar] [CrossRef]
  8. Raised Cholesterol. Available online: https://www.who.int/data/gho/indicator-metadata-registry/imr-details/3236 (accessed on 19 June 2022).
  9. Kanter, M.M.; Kris-Etherton, P.M.; Fernandez, M.L.; Vickers, K.C.; Katz, D.L. Exploring the factors that affect blood cholesterol and heart disease risk: Is dietary cholesterol as bad for you as history leads us to believe? Adv. Nutr. 2012, 3, 711–717. [Google Scholar] [CrossRef] [Green Version]
  10. Hu, P.; Dharmayat, K.I.; Stevens, C.A.; Sharabiani, M.T.; Jones, R.S.; Watts, G.F.; Genest, J.; Ray, K.K.; Vallejo-Vaz, A.J. Prevalence of familial hypercholesterolemia among the general population and patients with atherosclerotic cardiovascular disease: A systematic review and meta-analysis. Circulation 2020, 141, 1742–1759. [Google Scholar] [CrossRef]
  11. Bouillon, K.; Singh-Manoux, A.; Jokela, M.; Shipley, M.J.; Batty, G.D.; Brunner, E.J.; Sabia, S.; Tabák, A.G.; Akbaraly, T.; Ferrie, J.E.; et al. Decline in low-density lipoprotein cholesterol concentration: Lipid-lowering drugs, diet, or physical activity? Evidence from the Whitehall II study. Heart 2011, 97, 923–930. [Google Scholar] [CrossRef]
  12. Fazakis, N.; Kocsis, O.; Dritsas, E.; Alexiou, S.; Fakotakis, N.; Moustakas, K. Machine learning tools for long-term type 2 diabetes risk prediction. IEEE Access 2021, 9, 103737–103757. [Google Scholar] [CrossRef]
  13. Alexiou, S.; Dritsas, E.; Kocsis, O.; Moustakas, K.; Fakotakis, N. An approach for Personalized Continuous Glucose Prediction with Regression Trees. In Proceedings of the 2021 IEEE 6th South-East Europe Design Automation, Computer Engineering, Computer Networks and Social Media Conference (SEEDA-CECNSM), Preveza, Greece, 24–26 September 2021; pp. 1–6. [Google Scholar]
  14. Dritsas, E.; Alexiou, S.; Konstantoulas, I.; Moustakas, K. Short-term Glucose Prediction based on Oral Glucose Tolerance Test Values. In Proceedings of the International Joint Conference on Biomedical Engineering Systems and Technologies—HEALTHINF, Vienna, Austria, 9–11 February 2022; Volume 5, pp. 249–255. [Google Scholar]
  15. Dritsas, E.; Fazakis, N.; Kocsis, O.; Fakotakis, N.; Moustakas, K. Long-Term Hypertension Risk Prediction with ML Techniques in ELSA Database. In International Conference on Learning and Intelligent Optimization; Springer: Berlin/Heidelberg, Germany, 2021; pp. 113–120. [Google Scholar]
  16. Moll, M.; Qiao, D.; Regan, E.A.; Hunninghake, G.M.; Make, B.J.; Tal-Singer, R.; McGeachie, M.J.; Castaldi, P.J.; Estepar, R.S.J.; Washko, G.R.; et al. Machine learning and prediction of all-cause mortality in COPD. Chest 2020, 158, 952–964. [Google Scholar] [CrossRef] [PubMed]
  17. Zoabi, Y.; Deri-Rozov, S.; Shomron, N. Machine learning-based prediction of COVID-19 diagnosis based on symptoms. NPJ Digit. Med. 2021, 4, 1–5. [Google Scholar] [CrossRef] [PubMed]
  18. Dritsas., E.; Alexiou., S.; Moustakas., K. Cardiovascular Disease Risk Prediction with Supervised Machine Learning Techniques. In Proceedings of the 8th International Conference on Information and Communication Technologies for Ageing Well and e-Health—ICT4AWE, INSTICC, Prague, Czech Republic, 22–24 April 2022; SciTePress: Setúbal, Portugal, 2022; pp. 315–321. [Google Scholar]
  19. Dritsas, E.; Trigka, M. Stroke risk prediction with machine learning techniques. Sensors 2022, 22, 4670. [Google Scholar] [CrossRef] [PubMed]
  20. Wang, W.; Chakraborty, G.; Chakraborty, B. Predicting the risk of chronic kidney disease (ckd) using machine learning algorithm. Appl. Sci. 2020, 11, 202. [Google Scholar] [CrossRef]
  21. Speiser, J.L.; Karvellas, C.J.; Wolf, B.J.; Chung, D.; Koch, D.G.; Durkalski, V.L. Predicting daily outcomes in acetaminophen-induced acute liver failure patients with machine learning techniques. Comput. Methods Programs Biomed. 2019, 175, 111–120. [Google Scholar] [CrossRef] [PubMed]
  22. Yarasuri, V.K.; Indukuri, G.K.; Nair, A.K. Prediction of hepatitis disease using machine learning technique. In Proceedings of the 2019 IEEE Third International conference on I-SMAC (IoT in Social, Mobile, Analytics and Cloud)(I-SMAC), Palladam, India, 12–14 December 2019; pp. 265–269. [Google Scholar]
  23. Konstantoulas, I.; Kocsis, O.; Dritsas, E.; Fakotakis, N.; Moustakas, K. Sleep Quality Monitoring with Human Assisted Corrections. In Proceedings of the International Joint Conference on Computational Intelligence (IJCCI), SCIPTRESS, Valletta, Malta, 25–27 October 2021; pp. 435–444. [Google Scholar]
  24. Saba, T. Recent advancement in cancer detection using machine learning: Systematic survey of decades, comparisons and challenges. J. Infect. Public Health 2020, 13, 1274–1289. [Google Scholar] [CrossRef]
  25. Brautbar, A.; Leary, E.; Rasmussen, K.; Wilson, D.P.; Steiner, R.D.; Virani, S. Genetics of familial hypercholesterolemia. Curr. Atheroscler. Rep. 2015, 17, 1–17. [Google Scholar] [CrossRef]
  26. Pina, A.; Helgadottir, S.; Mancina, R.M.; Pavanello, C.; Pirazzi, C.; Montalcini, T.; Henriques, R.; Calabresi, L.; Wiklund, O.; Macedo, M.P.; et al. Virtual genetic diagnosis for familial hypercholesterolemia powered by machine learning. Eur. J. Prev. Cardiol. 2020, 27, 1639–1646. [Google Scholar] [CrossRef]
  27. Banda, J.M.; Sarraju, A.; Abbasi, F.; Parizo, J.; Pariani, M.; Ison, H.; Briskin, E.; Wand, H.; Dubois, S.; Jung, K.; et al. Finding missed cases of familial hypercholesterolemia in health systems using machine learning. npj Digit. Med. 2019, 2, 1–8. [Google Scholar] [CrossRef] [Green Version]
  28. Lee, B.J. Prediction model of hypercholesterolemia using body fat mass based on machine learning. J. Converg. Cult. Technol. 2019, 5, 413–420. [Google Scholar]
  29. García-d’Urso, N.; Climent-Pérez, P.; Sánchez-SanSegundo, M.; Zaragoza-Martí, A.; Fuster-Guillo, A.; Azorin-Lopez, J. A non-invasive approach for total cholesterol level prediction using machine learning. IEEE Access 2022, 10, 58566–58577. [Google Scholar] [CrossRef]
  30. Park, H.; Kim, K. Comparisons among machine learning models for the prediction of hypercholestrolemia associated with exposure to lead, mercury, and cadmium. Int. J. Environ. Res. Public Health 2019, 16, 2666. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  31. Banowati, C.; Novianty, A.; Setianingsih, C. Cholesterol level detection based on iris recognition using convolutional neural network method. In Proceedings of the 2019 IEEE Conference on Sustainable Utilization and Development in Engineering and Technologies (CSUDET), Penang, Malaysia, 7–9 November 2019; pp. 116–121. [Google Scholar]
  32. Fazakis, N.; Dritsas, E.; Kocsis, O.; Fakotakis, N.; Moustakas, K. Long-Term Cholesterol Risk Prediction with Machine Learning Techniques in ELSA Database. In Proceedings of the 13th International Joint Conference on Computational Intelligence (IJCCI), SCIPTRESS, Online, 25–27 October 2021; pp. 445–450. [Google Scholar]
  33. ENGLISH LONGITUDINAL STUDY OF AGEING. Available online: https://www.elsa-project.ac.uk/ (accessed on 19 June 2022).
  34. Yi, S.W.; Yi, J.J.; Ohrr, H. Total cholesterol and all-cause mortality by sex and age: A prospective cohort study among 12.8 million adults. Sci. Rep. 2019, 9, 1–10. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  35. Jiang, M.; Zou, Y.; Xin, Q.; Cai, Y.; Wang, Y.; Qin, X.; Ma, D. Dose–response relationship between body mass index and risks of all-cause mortality and disability among the elderly: A systematic review and meta-analysis. Clin. Nutr. 2019, 38, 1511–1523. [Google Scholar] [CrossRef] [PubMed]
  36. Lee, H.S.; Lee, J. Effects of combined exercise and low carbohydrate ketogenic diet interventions on waist circumference and triglycerides in overweight and obese individuals: A systematic review and meta-analysis. Int. J. Environ. Res. Public Health 2021, 18, 828. [Google Scholar] [CrossRef] [PubMed]
  37. Flint, A.C.; Conell, C.; Ren, X.; Banki, N.M.; Chan, S.L.; Rao, V.A.; Melles, R.B.; Bhatt, D.L. Effect of systolic and diastolic blood pressure on cardiovascular outcomes. N. Engl. J. Med. 2019, 381, 243–251. [Google Scholar] [CrossRef]
  38. Khirfan, G.; Tejwani, V.; Wang, X.; Li, M.; DiDonato, J.; Dweik, R.A.; Smedira, N.; Heresi, G.A. Plasma levels of high density lipoprotein cholesterol and outcomes in chronic thromboembolic pulmonary hypertension. PLoS ONE 2018, 13, e0197700. [Google Scholar]
  39. Barone Gibbs, B.; Hivert, M.F.; Jerome, G.J.; Kraus, W.E.; Rosenkranz, S.K.; Schorr, E.N.; Spartano, N.L.; Lobelo, F.; on Lifestyle, A.H.A.C.; on Cardiovascular, C.H.C.; et al. Physical Activity as a Critical Component of First-Line Treatment for Elevated Blood Pressure or Cholesterol: Who, What, and How?: A Scientific Statement From the American Heart Association. Hypertension 2021, 78, e26–e37. [Google Scholar] [CrossRef]
  40. Li, B.; Lei, S.S.; Su, J.; Cai, X.M.; Xu, H.; He, X.; Chen, Y.H.; Lu, H.X.; Li, H.; Qian, L.Q.; et al. Alcohol induces more severe fatty liver disease by influencing cholesterol metabolism. Evid.-Based Complement. Altern. Med. 2019, 2019, 7095684. [Google Scholar] [CrossRef]
  41. Gao, F.; Cui, C.Y. Dietary cholesterol intake and risk of gestational diabetes mellitus: A meta-analysis of observational studies. J. Am. Nutr. Assoc. 2022, 41, 107–115. [Google Scholar] [CrossRef]
  42. Ilyas, I.F.; Chu, X. Data Cleaning; Morgan & Claypool: San Rafael, CA, USA, 2019. [Google Scholar]
  43. Maldonado, S.; López, J.; Vairetti, C. An alternative SMOTE oversampling strategy for high-dimensional datasets. Appl. Soft Comput. 2019, 76, 380–389. [Google Scholar] [CrossRef]
  44. Pavithra, V.; Jayalakshmi, V. Hybrid feature selection technique for prediction of cardiovascular diseases. Mater. Today Proc. 2021, in press. [CrossRef]
  45. Tangirala, S. Evaluating the impact of GINI index and information gain on classification using decision tree classifier algorithm. Int. J. Adv. Comput. Sci. Appl. 2020, 11, 612–619. [Google Scholar] [CrossRef]
  46. Gnanambal, S.; Thangaraj, M.; Meenatchi, V.; Gayathri, V. Classification algorithms with attribute selection: An evaluation study using WEKA. Int. J. Adv. Netw. Appl. 2018, 9, 3640–3644. [Google Scholar]
  47. Silveira, E.A.; Pagotto, V.; Barbosa, L.S.; Oliveira, C.d.; Pena, G.d.G.; Velasquez-Melendez, G. Accuracy of BMI and waist circumference cut-off points to predict obesity in older adults. Ciênc. Saúde Coletiva 2020, 25, 1073–1082. [Google Scholar] [CrossRef]
  48. Krishnan, S.; Geetha, S. Prediction of Heart Disease Using Machine Learning Algorithms. In Proceedings of the 2019 IEEE 1st International Conference on Innovations in Information and Communication Technology (ICIICT), Chennai, India, 25–26 April 2019; pp. 1–5. [Google Scholar]
  49. Cunningham, P.; Delany, S.J. k-Nearest neighbour classifiers-A Tutorial. ACM Comput. Surv. (CSUR) 2021, 54, 1–25. [Google Scholar] [CrossRef]
  50. Nusinovici, S.; Tham, Y.C.; Yan, M.Y.C.; Ting, D.S.W.; Li, J.; Sabanayagam, C.; Wong, T.Y.; Cheng, C.Y. Logistic regression was as good as machine learning for predicting major chronic diseases. J. Clin. Epidemiol. 2020, 122, 56–69. [Google Scholar] [CrossRef]
  51. Ani, R.; Jose, J.; Wilson, M.; Deepa, O. Modified rotation forest ensemble classifier for medical diagnosis in decision support systems. In Progress in Advanced Computing and Intelligent Engineering; Springer: Berlin/Heidelberg, Germany, 2018; pp. 137–146. [Google Scholar]
  52. Bhargava, N.; Sharma, G.; Bhargava, R.; Mathuria, M. Decision tree analysis on j48 algorithm for data mining. Proc. Int. J. Adv. Res. Comput. Sci. Softw. Eng. 2013, 3. [Google Scholar]
  53. Masih, N.; Naz, H.; Ahuja, S. Multilayer perceptron based deep neural network for early detection of coronary heart disease. Health Technol. 2021, 11, 127–138. [Google Scholar] [CrossRef]
  54. Yang, Y.; Li, J.; Yang, Y. The research of the fast SVM classifier method. In Proceedings of the 2015 IEEE 12th International Computer Conference on Wavelet Active Media Technology and Information Processing (ICCWAMTIP), Chengdu, China, 18–20 December 2015; pp. 121–124. [Google Scholar]
  55. Dinesh, K.G.; Arumugaraj, K.; Santhosh, K.D.; Mareeswari, V. Prediction of cardiovascular disease using machine learning algorithms. In Proceedings of the 2018 IEEE International Conference on Current Trends towards Converging Technologies (ICCTCT), Coimbatore, India, 1–3 March 2018; pp. 1–7. [Google Scholar]
  56. Truong, X.L.; Mitamura, M.; Kono, Y.; Raghavan, V.; Yonezawa, G.; Truong, X.Q.; Do, T.H.; Tien Bui, D.; Lee, S. Enhancing prediction performance of landslide susceptibility model using hybrid machine learning approach of bagging ensemble and logistic model tree. Appl. Sci. 2018, 8, 1046. [Google Scholar] [CrossRef] [Green Version]
  57. Li, R.; Shen, S.; Zhang, X.; Li, R.; Wang, S.; Zhou, B.; Wang, Z. Cardiovascular disease risk prediction based on random forest. In Proceedings of the International Conference on Healthcare Science and Engineering, Guilin, China, 10–12 September 2018; Springer: Berlin/Heidelberg, Germany, 2018; pp. 31–43. [Google Scholar]
  58. Kumari, S.; Kumar, D.; Mittal, M. An ensemble approach for classification and prediction of diabetes mellitus using soft voting classifier. Int. J. Cogn. Comput. Eng. 2021, 2, 40–46. [Google Scholar] [CrossRef]
  59. Pavlyshenko, B. Using stacking approaches for machine learning models. In Proceedings of the 2018 IEEE Second International Conference on Data Stream Mining & Processing (DSMP), Lviv, Ukraine, 21–25 August 2018; pp. 255–258. [Google Scholar]
  60. Hossin, M.; Sulaiman, M.N. A review on evaluation metrics for data classification evaluations. Int. J. Data Min. Knowl. Manag. Process. 2015, 5, 1. [Google Scholar]
  61. Waikato Environment for Knowledge Analysis. Available online: https://www.weka.io/ (accessed on 12 July 2022).
Figure 1. Pearson correlation analysis.
Figure 1. Pearson correlation analysis.
Sensors 22 05365 g001
Figure 2. Participants’ distribution per age group and gender type in the balanced dataset.
Figure 2. Participants’ distribution per age group and gender type in the balanced dataset.
Sensors 22 05365 g002
Figure 3. Participants’ distribution in terms of BMI and waist categories in the balanced dataset.
Figure 3. Participants’ distribution in terms of BMI and waist categories in the balanced dataset.
Sensors 22 05365 g003
Figure 4. Participants’ distribution for both diabetes and hypertension in the balanced dataset.
Figure 4. Participants’ distribution for both diabetes and hypertension in the balanced dataset.
Sensors 22 05365 g004
Figure 5. Participants’ distribution in terms of alcohol consumption in the balanced dataset.
Figure 5. Participants’ distribution in terms of alcohol consumption in the balanced dataset.
Sensors 22 05365 g005
Figure 6. Participants’ distribution in terms of physical activity in the balanced dataset.
Figure 6. Participants’ distribution in terms of physical activity in the balanced dataset.
Sensors 22 05365 g006
Figure 7. Ensemble Learners: Soft Voting and Stacking.
Figure 7. Ensemble Learners: Soft Voting and Stacking.
Sensors 22 05365 g007
Table 1. Statistical Description of the Numerical Features in the Balanced Dataset.
Table 1. Statistical Description of the Numerical Features in the Balanced Dataset.
MinMaxMean ± stdv
Age508566.4 ± 9.5
BMI18.353.128.61 ± 5.02
Waist70148.6101.76 ± 13.18
SysBP90201136.6 ± 20.5
DiasBP1310870.27 ± 12.22
HDL1911450.94 ± 16.42
LDL51328157.6 ± 40.1
TotalChol75360208.49 ± 39.69
Table 2. Features’ order of importance in the balanced data.
Table 2. Features’ order of importance in the balanced data.
FeaturePearson
Rank
FeatureGain
Ratio
FeatureInfoGain
Ratio
FeatureRandom
Forest
(AUPRC)
TotChol0.6777TotChol0.3061TotChol0.5633TotChol0.3790
LDL0.6152LDL0.2171LDL0.3963LDL0.3165
HDL0.1366DiasBP0.1142DiasBP0.0283DiasBP0.0788
DiasBP0.1148Gender0.0085Gender0.0085Age0.0512
BMI0.1106Alcohol
Consumption
0.0079Alcohol
Consumption
0.0079BMI0.0262
Gender0.1038Hypertension0.0034Physical Activity0.0043Alcohol
Consumption
0.0242
Alcohol
Consumption
0.1042Physical Activity0.0029Hypertension0.0034HDL0.0182
Age0.0711Diabetes0.0027Diabetes0.0019SysBP0.0154
Hypertension0.0681SysBP0SysBP0Waist0.0151
Physical
Activity
0.0586HDL0HDL0Gender0.0145
Diabetes0.0520BMI0BMI0Hypertension0.0124
SysBP0.0502Waist0Waist0Diabetes0.0000
Waist0.0192Age0Age0Physical
Activity
−0.0021
Table 3. Machine Learning Models’ Settings.
Table 3. Machine Learning Models’ Settings.
ModelParameters
NBuseKernerEstimator = false
LRridge = 10 8 , useConjugateGradientDescent = false
LMTLR modesl at leaves
errorOnProbabilities = false, fastRegression = false,
numInstances = 15, useAIC = false
DTnoPruning: false, MinVarianceProp = 0.001
numfolds = 3
RotF (using J48)confidence_factor: 0.25, unpruned: false
minimum_instances per_leaf_node default binary split: false
RFmax_depth = 0, numIterations = 100 numFeatures = 0
ANNhidden layers: ‘a’, learning rate: 0.3 momentum factor 0.2,
training time 500
SVMkernel type: linear
K-NNK = 3, 5
Search Algorithm: LinearNNSearch with Euclidean
StackingBase Models:RF, RotF Meta-model:LR
Soft VotingBase Models:RF, RotF Average Probabilities
Table 4. Performance Evaluation of ML Models.
Table 4. Performance Evaluation of ML Models.
AccuracyPrecisionRecallF-MeasureAUC
NB87.37%0.8770.8740.8730.931
SVM88.40%0.8840.8840.8840.884
LR87.63%0.8760.8760.8760.927
ANN82.73%0.8280.8270.8270.912
3-NN70.62%0.7070.7060.7060.758
RotF90.98%0.9110.9100.9100.939
LMT86.85%0.8690.8690.8690.928
RF89.69%0.9000.8970.8970.943
DT88.92%0.8920.8890.8890.902
Stacking91.24%0.9150.9120.9120.937
Soft Voting91.75%0.9200.9180.9170.945
Table 5. Performance Comparison of ML Models.
Table 5. Performance Comparison of ML Models.
RecallAccuracy
Proposed models[32]Proposed models[32]
NB87.40%68.90%87.37%62.69%
SVM88.40%72.70%88.40%59.51%
ANN82.70%66.70%82.73%61.42%
5-NN67.30%67.70%67.27%56.56%
RotF91%69.60%90.98%61.86%
DT88.90%72.20%88.92%61.39%
LMT86.90%73.50%86.85%62.99%
RF89.70%68.80%89.69%61.36%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Dritsas, E.; Trigka, M. Machine Learning Methods for Hypercholesterolemia Long-Term Risk Prediction. Sensors 2022, 22, 5365. https://doi.org/10.3390/s22145365

AMA Style

Dritsas E, Trigka M. Machine Learning Methods for Hypercholesterolemia Long-Term Risk Prediction. Sensors. 2022; 22(14):5365. https://doi.org/10.3390/s22145365

Chicago/Turabian Style

Dritsas, Elias, and Maria Trigka. 2022. "Machine Learning Methods for Hypercholesterolemia Long-Term Risk Prediction" Sensors 22, no. 14: 5365. https://doi.org/10.3390/s22145365

APA Style

Dritsas, E., & Trigka, M. (2022). Machine Learning Methods for Hypercholesterolemia Long-Term Risk Prediction. Sensors, 22(14), 5365. https://doi.org/10.3390/s22145365

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop