Next Article in Journal
A Systematic Database Approach to Identify Companion Diagnostic Testing in Clinical Trials under the New In Vitro Diagnostic Medical Devices Regulation
Next Article in Special Issue
Improving the Accuracy of Continuous Blood Glucose Measurement Using Personalized Calibration and Machine Learning
Previous Article in Journal
The Vertebral Artery: A Systematic Review and a Meta-Analysis of the Current Literature
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Classification of Diabetes Using Feature Selection and Hybrid Al-Biruni Earth Radius and Dipper Throated Optimization

by
Amel Ali Alhussan
1,
Abdelaziz A. Abdelhamid
2,3,*,
S. K. Towfek
4,5,
Abdelhameed Ibrahim
6,
Marwa M. Eid
7,
Doaa Sami Khafaga
1,* and
Mohamed S. Saraya
6
1
Department of Computer Sciences, College of Computer and Information Sciences, Princess Nourah bint Abdulrahman University, P.O. Box 84428, Riyadh 11671, Saudi Arabia
2
Department of Computer Science, College of Computing and Information Technology, Shaqra University, Shaqra 11961, Saudi Arabia
3
Department of Computer Science, Faculty of Computer and Information Sciences, Ain Shams University, Cairo 11566, Egypt
4
Computer Science and Intelligent Systems Research Center, Blacksburg, VA 24060, USA
5
Department of Communications and Electronics, Delta Higher Institute of Engineering and Technology, Mansoura 35111, Egypt
6
Computer Engineering and Control Systems Department, Faculty of Engineering, Mansoura University, Mansoura 35516, Egypt
7
Faculty of Artificial Intelligence, Delta University for Science and Technology, Mansoura 11152, Egypt
*
Authors to whom correspondence should be addressed.
Diagnostics 2023, 13(12), 2038; https://doi.org/10.3390/diagnostics13122038
Submission received: 30 April 2023 / Revised: 3 June 2023 / Accepted: 6 June 2023 / Published: 12 June 2023
(This article belongs to the Special Issue Machine Learning Models in Diagnosis and Treatment of Diabetes)

Abstract

:
Introduction: In public health, machine learning algorithms have been used to predict or diagnose chronic epidemiological disorders such as diabetes mellitus, which has reached epidemic proportions due to its widespread occurrence around the world. Diabetes is just one of several diseases for which machine learning techniques can be used in the diagnosis, prognosis, and assessment procedures. Methodology: In this paper, we propose a new approach for boosting the classification of diabetes based on a new metaheuristic optimization algorithm. The proposed approach proposes a new feature selection algorithm based on a dynamic Al-Biruni earth radius and dipper-throated optimization algorithm (DBERDTO). The selected features are then classified using a random forest classifier with its parameters optimized using the proposed DBERDTO. Results: The proposed methodology is evaluated and compared with recent optimization methods and machine learning models to prove its efficiency and superiority. The overall accuracy of diabetes classification achieved by the proposed approach is 98.6%. On the other hand, statistical tests have been conducted to assess the significance and the statistical difference of the proposed approach based on the analysis of variance (ANOVA) and Wilcoxon signed-rank tests. Conclusions: The results of these tests confirmed the superiority of the proposed approach compared to the other classification and optimization methods.

1. Introduction

Hyperglycemia due to abnormalities in insulin secretion, insulin action, or both characterize the metabolic condition known as diabetes mellitus (DM). Long-term damage, dysfunction, and failure of multiple organs, including the heart, eyes, kidneys, blood vessels, and nerves, can be attributed to the chronic hyperglycemia that is associated with DM [1]. Diabetes mellitus (DM) is divided into three subtypes based on its etiology and clinical presentation: type 1 diabetes (T1DM), type 2 diabetes (T2DM), and gestational diabetes. In type 1 diabetes, beta cells in the pancreas are destroyed, typically as a result of a cellular-mediated autoimmune response, leading to an utter lack of insulin. Type 2 diabetes is brought on by insulin resistance and a little insulin shortage. Gestational diabetes, often known as pregnancy-onset diabetes, is characterized by glucose intolerance of varied degrees [2]. The vast majority of diabetes cases are type 2 diabetes. Although type 2 diabetes is more common in people over 40, it can also affect those of any age. The symptoms of this type may not manifest for years, and many individuals are diagnosed by chance when they seek care for unrelated issues. Patients with T2DM are not insulin-dependent, but they may require insulin therapy to manage hyperglycemia if it cannot be achieved via diet alone or with oral hypoglycemic medications [3].
The causes of type 2 diabetes are varied and convoluted. Many factors increase or decrease the likelihood of contracting the disease, although not all of them are direct causes. These factors may be hereditary, demographic (such as age), or behavioral (food, smoking, obesity, and lack of exercise). Behavioral risk factors are sometimes referred to as “modifiable” risk factors [4] since they can be altered or improved upon. More than 460 million individuals worldwide were diagnosed with diabetes in 2019, and that number is anticipated to rise to 578 million by 2030 and 700 million by 2045, as reported by the International Diabetes Federation (IDF) [5]. The prevalence of type 2 diabetes mellitus (T2DM) continues to rise, making it one of the most concerning NCDs.
To predict the future incidence and global prevalence of diabetes, numerous studies over the past two decades have used various data sets and analytical methodologies [6,7]. Accurate forecasts of the future burden of diabetes are vital for health policy planning and establishing the costs of managing the condition [8,9]. Disease prediction and diagnosis for pandemic chronic diseases such as diabetes are two areas where machine learning algorithms have lately seen extensive application in public health. Machine learning techniques examine historical patterns in data to foretell what will happen in the future. Algorithms in machine learning provide data modeling, analysis, and visualization mention [10,11]. Several machine learning techniques have been utilized in diabetes modeling studies, including support vector machines (SVMs), artificial neural networks (ANNs), k-nearest neighbor (KNN), and decision trees (DTs) [12,13]. Despite the importance of studying diabetes prevalence trends and predicting future burdens using risk factors in specific populations, little work has been done to adopt machine learning classification methods.
Feature selection is essential in analyzing healthcare datasets, such as diabetes-related ones. The goal is to find the small subset of features that makes the most significant impact on a task of classification or prediction. The problem is high-dimensional datasets, which frequently have extraneous or redundant features that can cause overfitting and lower classification accuracy. Researchers have looked into using metaheuristic optimization techniques for feature selection in diabetic datasets to address this problem. On the other hand, metaheuristic optimization algorithms provide effective means of exploring and ultimately settling on subsets of features that can be used to improve machine learning performance. These algorithms take their cues from real-world occurrences or problem-solving techniques, and they’re built to efficiently probe the range of possible solutions. These algorithms aim to make diabetes categorization models more accurate, less dimensional, and easier to read [14,15].
Feature selection is challenging in machine learning due to several inherent difficulties. One of the primary challenges is dealing with high-dimensional datasets that contain a large number of features. High dimensionality often leads to increased computational complexity, reduced model interpretability, and the risk of overfitting. Additionally, datasets may include irrelevant or redundant features that can adversely affect the performance of machine-learning models. Metaheuristic optimization methods offer promising solutions to address these challenges. By leveraging these the exploration and exploitation capabilities of these algorithms, researchers can enhance the feature selection accuracy, reduce dimensionality, and optimize the performance of machine learning models for tasks such as diabetes classification [16,17].
In this paper, a new optimization algorithm is proposed for feature selection and optimization of the parameters of the random forest classifier. The proposed optimization algorithm is a dynamic hybrid of the Al-Biruni Earth Radius and Dipper-Throated Optimization algorithm and is denoted by (DBERDTO). The main advantage of this algorithm is the improved exploration and exploitation of the search space while performing the optimization process. This advantage appears in the promising classification results, which are superior to those of the other competing methods. The following is a summary of the novelty of this work:
  • A new binary optimization algorithm (bDBERDTO) is proposed for feature selection to select the most significant set of features that can improve the classification results;
  • A new optimization algorithm (DBERDTO) is proposed to optimize the parameters of the random forest classifier to boost the classification accuracy;
  • A comparison with the other feature selection methods is performed to prove the superiority of the proposed feature selection method;
  • A comparison with the other classification models is performed to show the effectiveness of the proposed optimization method and the proposed approach to optimizing the parameters of the random forest classifier;
  • A statistical analysis is performed to show the stability and statistical difference between the proposed approach and the other competing approaches;
  • The ANOVA and Wilcoxon signed rank tests are performed, and the results are analyzed to show the effectiveness of the proposed methodology;
  • The proposed approach is evaluated in terms of four diabetes datasets available on Kaggle to prove its effectiveness and generalization.
The paper proceeds as follows. There is a literature review in Section 2. Section 3 discusses the proposed methodology. The experimental approach is analyzed in Section 4. In Section 5, the outcomes are concluded and the future perspectives are highlighted.

2. Literature Review

Machine learning algorithms have found extensive application in public health, particularly in disease prediction and diagnosis for chronic epidemic conditions such as diabetes. Various machine learning methods, such as support vector machines, artificial neural networks, k-nearest neighbors, and decision tree models, have all been used in diabetes models that have been published. Various applications have found success with these models, including the early detection of diabetes and the modeling of its consequences. To compare the suggested classifiers’ accuracy with existing ones, this section includes the commonly used machine learning techniques and their respective accuracy rates. Table 1 provides a summary of all the research that is mentioned in this section. In a 2019 study, Ref. [18] compared the performance of various classification models for diabetes prediction using a variety of machine learning techniques (including SVM, C4.5 decision tree, naive Bayes, and KNN) and evaluation metrics (accuracy, recall, and precision). Medical Centre Chittagong (MCC) in Bangladesh served as the source for the diabetes data used in the study. There are 200 patients in the dataset, and they all have unique features, including age, sex, weight, blood pressure, and other potential health issues. Study findings showed that the C4.5 decision tree model performed the best, with an accuracy of 73%. The authors of Ref. [19], another study published in 2018, set out to assess how well categorization algorithms might foretell cases of diabetes. A total of 768 samples were used in this analysis, all of which came from the PIMA Indian data repository. Seventy percent (n = 583) of the data was used for training, whereas thirty percent (n = 231) was used for testing. Logistic regression (LR), k-nearest neighbor (KNN), support vector machine (SVM), gradient boost, decision tree (DT), boosted DT, boosted MLP, random forest (RF), and Gaussian naive Bayes (GNB) were the eight machine learning models tested in this work. According to the findings, LR’s accuracy was the highest at 79.54%.
The authors in Ref. [20] built an ANN model with varying numbers of hidden network neurons (from 5 to 50). Female diabetes was predicted using information from the National Institute of Diabetes and Digestive and Kidney Diseases. Using the Pima Indian Diabetes dataset for validation and the assessment measures of accuracy and mean squared error (MSE), it was shown that the ANN model trained with 8 features had a 93% accuracy. Using AdaBoost, k-NN, decision trees, XG Boost, naive Bayes, and random forest, the authors of Ref. [21] conducted a study in 2020 to diagnose and forecast the onset of diabetes. A total of 768 female patients were used to train the models, 268 of whom had diabetes (positive) and 500 did not (negative). This analysis considered eight factors: blood sugar, insulin, pregnancy, blood pressure, triceps strength, body mass index, family history, and age. Feature selection, data normalization, outlier rejection, mean substitution for missing values, and k-fold cross-validation (five-fold cross-validation) were all part of the data preparation process [22]. An ensemble method was also employed to improve performance with numerous classifiers further. The predictive power of ensemble methods can be increased by combining the results from multiple models. When combined, AdaBoost and XGBoost produced the highest-quality results. The area under the curve (AUC) was used as the effectiveness measure. An AUC of 0.95 was achieved, making their study the most accurate.
In Ref. [23], the authors from several institutions used the PIDD dataset to create a number of machine-learning models to identify whether or not a given female patient has diabetes. The mean substitution was used to deal with the missing data, and a standardization procedure was used to rescale all attributes. Four different types of support vector machines (linear and polynomial) and one type of random forest (RF) were used to build the models. According to their research findings, the RF model had the best accuracy rate. A classification methodology for early diabetes detection using machine learning techniques is offered in a 2020 study [24,25]. This study aimed to produce results consistent with clinical outcomes using prominent traits. The information for this study came from a survey given to patients at the Diabetes Hospital in Sylhet, Bangladesh. A total of 520 examples with 16 variables are included in this dataset, all of which reflect symptoms associated with diabetes. Both positive (risk of diabetes) and negative (no risk of diabetes) diagnoses were determined using the two class features of the authors. We trained a multilayer perceptron (MLP), a radial basis function network (RBF), and a random forest (RF) to see which of these three classifiers was most effective in accurately predicting diabetes. Compared to other models, it was found that the RBF model was the most effective [26].
Some current studies have addressed the issue of using machine learning techniques to build prediction models for diabetic complications, in addition to the studies that predicted or diagnosed diabetes. The model established in 2020 by the authors in Ref. [27] to forecast hyperlipidemia, coronary heart disease, kidney disease, and eye disease as potential outcomes for people with diabetes is one such example. This research made use of a dataset consisting of 455. The dataset went through some selection and cleaning procedures that cut down on the number of records included in the model’s construction. The model was built using the iterative decision tree (ID3) algorithm. An accuracy of 92.35% was obtained using a 10-fold cross-validation procedure to assess the effectiveness of the suggested model. Especially when training on unbalanced data, the high accuracy score attained in this study is insufficient to evaluate the model’s performance. The key reason is that during training, the model can disregard a minority class and still produce accurate predictions for the majority class. To better anticipate the onset of retinopathy, neuropathy, and nephropathy in T2DM patients, the authors in a study conducted in 2018 [28] built various classification models such as LR, NB, SVMs, and random forest. The authors made their projections based on three different time frames: three, five, and seven years after the initial diabetic hospitalization. The dataset used to train the suggested models was compiled over a decade by researchers at the Istituto Clinico Scientifico Maugeri (ICSM), Hospital of Pavia, Italy. There are a total of 943 records here. They include topics such as gender, age, body mass index (BMI), time since diagnosis, hypertension, glycated hemoglobin (HbA1c), and smoking status. When dealing with missing data, the random forest method was used, and when dealing with uneven class sizes, oversampling the smaller group helped. The data collected showed that LR had the greatest accuracy score (77.7%).

3. The Proposed Methodology

In this section, the proposed methodology is presented and explained. The methodology starts with data preprocessing, feature selection, and finally, diabetes classification. The key contribution of the proposed methodology is the proposed optimization algorithm. This optimization algorithm selects and optimizes the classifier parameters used in feature classification. The steps of the proposed methodology are shown in Figure 1, and these steps are detailed in the following sections.

3.1. Data Preprocessing

Data preprocessing is an essential step for cleaning, transforming, and preparing data to be used in modeling. One of the initial preprocessing steps is handling missing data. We can use mean, median, or mode imputation to fill in missing data values. Mathematically, we use the following equation to replace missing data with the mean value:
Mean = i = 1 n x i n
Another preprocessing step is identifying and handling outliers. Mathematical functions such as the log and square root functions can be used to transform the data. This data preprocessing is critical to improving the accuracy of the models. The following equation demonstrates the log transformation of the data:
L o g T r a n s f o r m = log ( x )
Normalization is also a crucial preprocessing step to ensure that all features are on the same scale. Min-max scaling is a common technique used to scale the data between 0 and 1, while the Z-score normalization scales the data using each feature’s mean and standard deviation. The following equation shows the min-max scaling of data:
M i n M a x S c a l i n g = x x m i n x m a x x m i n

3.2. Metaheuristic Optimization

Feature selection and the optimization of machine learning model parameters have benefited greatly from the rise in the popularity of metaheuristic optimization in recent years. This method shines when applied to situations with many moving parts or when extensive computational resources are needed to investigate all conceivable outcomes. Metaheuristic optimization algorithms excel in these situations because they can quickly and efficiently scour a large solution space for a desirable outcome. Non-differentiable, discontinuous, or multimodal objective functions are no problem for metaheuristic optimization. These functions frequently come up in feature selection and model parameter optimization tasks, where the goal is to discover the optimal set of features and model parameters to either reduce the error or maximize the accuracy of a machine learning model. In many fields, including medicine, business, and engineering, metaheuristic optimization techniques, including genetic algorithms, particle swarm optimization, and simulated annealing, have greatly affected feature selection and model parameter optimization [29,30]. The workings of nature or social systems, such as natural selection or swarm behavior, inspire these adaptive algorithms. Metaheuristic optimization is a robust method for feature selection and optimization of machine learning model parameters that can effectively deal with complex, non-differentiable, or multimodal objective functions. Discovering the optimal feature and parameter combinations that maximize accuracy for a given dataset can enhance performance and facilitate more informed decision-making across various domains [31,32,33,34,35].

3.3. Al-Biruni Earth Radius Optimization Algorithm

The Al-Biruni Earth Radius (BER) is an optimization technique that can enhance the search efficiency by dividing individuals in the search space into two groups that focus on exploration and exploitation [36]. This technique involves agents dynamically shifting the composition of subgroups to balance exploratory and exploitative pursuits. The exploration team, which constitutes 70% of the individuals, utilizes mathematical methods to look for promising new territory nearby. This is achieved through an iterative process of exploring alternatives until an optimal fitness level is achieved. Meanwhile, the exploitation team, consisting of 30% of individuals, focuses on exploiting the discovered optimal regions. The number of agents in both groups has increased to improve their global average fitness. To employ the BER optimization algorithm, each individual in the population is treated as a vector S representing the optimization parameter or features d in the search space of size S d R . The fitness function F measures an individual’s success up to a given threshold. The optimization stages aim to probe populations and discover the S* value that maximizes fitness. The BER technique can be applied by specifying the fitness function, population size, dimension, and minimum and maximum acceptable solution sizes. Optimization algorithms aim to find the optimal solution within defined limits, and the BER technique can aid in achieving this goal. The BER technique has proven useful in optimizing machine learning models by improving search efficiency through a balance of exploration and exploitation. By dividing individuals into two groups and dynamically adjusting their composition, the BER technique can efficiently explore the search space and find the optimal combination of parameters and features. Furthermore, it can be easily employed by specifying the necessary parameters, making it a practical solution for optimizing complex problems.
The exploration process involves searching the search space for promising regions that can lead to finding the optimal solution. The lone explorer in the group looks for new locations to explore near their current location to move closer to the perfect solution. However, the effectiveness of exploration must be evaluated by exploring a variety of local possibilities and selecting the best ones. The BER technique utilizes the equations given below to achieve this goal:
P ( t + 1 ) = P ( t ) + D ( 2 r 2 1 ) , D = r 1 ( P ( t ) 1 )
The solution vector at iteration t is denoted by P ( t ) , and the diameter of the circle within which the search agent searches for interesting regions is represented by D. The range of x in the coefficient vectors r 1 and r 2 is from 0 to 180, while the value of h is a scalar randomly chosen between [ 0 , 2 ] . The coefficients r 1 and r 2 can be obtained by solving the equation r = h c o s ( x ) 1 c o s ( x ) .
The team tasked with seizing opportunities must constantly improve existing methods. At the end of each cycle, the BER rewards those who have put the most effort toward achieving the highest fitness levels. The BER employs two distinct methods to achieve its exploitation goal, which we will discuss in detail. By using the equation provided below, we can take steps toward finding the best solution and move closer to the solution.
P ( t + 1 ) = r 2 ( P ( t ) + D ) , D = r 3 ( L ( t ) P ( t ) )
At each iteration t, P ( t ) is the solution vector, L ( t ) is the best solution vector, and D is the distance vector. r 3 is a random vector generated using the formula r 1 s = h c o s ( x ) 1 c o s ( x ) . This formula governs the movement towards exploring the space around the best solution, which is the most promising of all possible solutions (leader). This encourages the exploration of solutions close to the ideal.
Furthermore, the BER uses another equation to optimize its search. The optimal solution is denoted by P ( t ) , and the following equation guides the implementation of the optimal P :
P ( t + 1 ) = r ( P * ( t ) + k ) , k = 1 + 2 × t 2 M a x i t e r 2
In this equation, k is a scalar factor that gradually increases with time, and  M a x i t e r is the maximum number of iterations. The best fitness value is compared between P ( t + 1 ) and P ( t + 1 ) to choose the optimal P * implementation. If there is no improvement in fitness during the previous two iterations, the following equation is used to update the solution:
P ( t + 1 ) = k z 2 h c o s ( x ) 1 c o s ( x )
In this equation, z is a random number in the [ 0 , 1 ] range. By constantly improving and updating its methods, the BER can effectively seize opportunities and optimize its search for the best solution.

3.4. The Dipper Throated Optimization Algorithm

The Dipper Throated Optimization (DTO) algorithm makes an innovative assumption that there are two groups of birds: the first group comprises swimming birds, and the second includes flying birds. These two groups cooperate in foraging food, mapped onto exploration and exploitation groups to find the best solution. The birds in these groups have positions and velocities that can be illustrated using matrices. The position matrix, P, contains the positions of the birds in each dimension, whereas the velocity matrix, V, includes the velocities of the birds in each dimension. Each bird’s fitness is measured by a fitness function, f, defined using the position matrix. During fitness evaluation, the mother bird has the highest fitness score, and the best solution is referred to as P b e s t . Common birds play the role of followers and are represented by P n d , while the best solution in the search space is identified as P * . The first approach of the DTO algorithm to track the swimming bird relies on the following equations to update the location and velocity of the birds in the population:
X = P b e s t ( i ) K 1 . | K 2 . P b e s t ( i ) P ( i ) |
Y = V ( i + 1 ) + P ( i )
P ( i + 1 ) = X if R < 0.5 Y otherwise ,
V ( i + 1 ) = K 3 V ( i ) + K 4 r 1 ( P b e s t ( i ) P ( i ) ) + K 5 r 2 ( P * P ( i ) )
Here, i denotes the current iteration index, while i + 1 represents the next iteration index. Equation X determines the change in the position of the bird from the best bird’s position. The Y equation updates the bird’s velocity based on its current velocity and position. In the P ( i + 1 ) equation, the bird’s position in the next iteration is updated based on the value of R. If R is less than 0.5 , the bird’s position is updated using equation X. Otherwise, it is updated using equation Y. Finally, the  V ( i + 1 ) equation updates the bird’s velocity by considering its current velocity, the distance between the best bird and the bird’s current position, and the distance between the P * and the bird’s current position. The constants K 1 to K 5 are coefficients that determine the impact of each factor on the bird’s position and velocity.

3.5. The Proposed Feature Selection Algorithm

Selecting the most pertinent features that contribute most to the classification accuracy is an important part of the diabetes classification process. Feature selection improves classification accuracy by decreasing the dimensionality of the, discarding non-essential or redundant features. This simplifies the model, enabling it to run more quickly and efficiently in real-time settings. Feature selection is the process of choosing which features are the most useful by evaluating them against a set of criteria, such as their correlation with the target variable or their ability to distinguish across classes. Feature selection can be accomplished in a number of ways, such as the filter, wrapper, or embedding approaches. To determine the importance of each feature outside of the context of the classification model, filter approaches use statistical tests or correlation coefficients. In contrast, wrapper techniques iteratively add and remove features based on the classification model’s assessment of their relative relevance. The feature weights of an embedded approach are learned directly from the data, and the method combines feature selection with the training process of the classification model. Feature selection is crucial in diabetes classification because it helps make classification models more accurate and interpretable. Multiple studies have demonstrated that by employing feature selection approaches, the dimensionality of the dataset can be drastically reduced without sacrificing accuracy in classification.
The accuracy and efficiency of classification models can be enhanced by feature selection, making it an essential strategy in diabetes classification. The scientists discovered that feature selection strategies increased classification performance compared to employing all the features, with mutual information-based feature selection yielding the best results. The scientists also stated that feature selection assisted in determining the most important traits associated with GDM, which could lead to improved diagnostic tools. Feature selection can assist in reducing the complexity of a dataset, making it more manageable for analysis and interpretation by highlighting the most important elements. In addition, feature selection can help zero in on the most informative features of the disease, which could lead to better, more interpretable diagnostic tools; for high-dimensional and complicated datasets as those used in diabetes classification, metaheuristic optimization algorithms have emerged as strong tools for feature selection. These algorithms take cues from natural occurrences or human behavior to get near-optimal results through efficient search space exploration by avoiding local optima. Feature selection for diabetes classification often uses metaheuristic algorithms, including genetic algorithms, particle swarm optimization, ant colony optimization, simulated annealing, and artificial bee colony. A fitness function, which might be based on classification accuracy, information gain, or some other criterion, can be used by these algorithms to search for an outstanding collection of features. Metaheuristic optimization can improve the speed and overcome the limits of other feature selection methods, such as wrapper, filter, and embedding approaches. When applied to feature selection for diabetes classification, metaheuristic optimization has been shown to be effective and can lead to the development of more accurate and resilient models.
When deciding whether or not a given feature is important, feature selection issues have a small search space consisting only of the binary values 0 and 1. To better accommodate the feature selection procedure, we present a binary version of the DBERDTO method, which transforms the continuous values produced by the original algorithm into binary [0, 1] values. The  S i g m o i d function is given by the following equation, which is used to execute the conversion to binary values.
S ( t + 1 ) = 1 if S i g m o i d ( S b e s t ) 0.5 0 o t h e r w i s e , S i g m o i d ( S B e s t ) = 1 1 + e 10 ( S B e s t 0.5 )
where S b e s t is the optimal solution for step t in the iterative process. Scaling the continuous values found in Algorithm 1 to the discrete range [0, 1] is the primary focus of the sigmoid function.
Algorithm 1: The proposed binary DBERDTO algorithm
1:
Initialize the configuration, parameters, and population of DBERDTO.
2:
Binarize the retrieved solutions using the sigmoid function
3:
Evaluate the objective function
4:
Find the best solutions.
5:
Train k-NN and calculate error
6:
while  t M a x i t e r  do
7:
   Execute the DBERDTO algorithm
8:
   Binarize the solutions using the sigmoid function
9:
   Evaluate the objective function
10:
   Update agents’ positions
11:
end while
12:
Return best binary solution

3.6. Objective Function

Using the proposed optimization approach, the following equation can be used to evaluate the received solution quality.
F n = α E r r o r ( P ) + β | S | | A |
where P stands for some set of inputs to the model. The significance of the chosen features in the population is reflected by the values of α [0, 1], β = 1 α . The number of selected features, denoted by | S | , is smaller than the total number of features in the dataset, denoted by | A | . The optimal strategy is the one that uses the fewest features to make the most accurate classifications.

3.7. Optimizing the Hyperparameters of the Random Forest Classifier

The number of trees in the forest, the maximum depth of each tree, the minimum number of samples required to split a node, the minimum number of samples required to be at a leaf node, and the number of features to consider when looking for the best split, and the criterion used for splitting are all hyperparameters of the random forest classifier that can be optimized with metaheuristic optimization techniques. The number of trees is a critical hyperparameter since it controls how many individual decision trees are in the forest. A larger number of trees can improve the accurary of the model, but it will take longer to compute. The possible splits in each decision tree are constrained by its maximum depth, which might reduce the likelihood of overfitting. Each decision tree can be made as simple or complex as desired by adjusting the thresholds for when a node should be split and when a node should be considered a leaf. One of the most crucial hyperparameters in finding the optimal split is the number of features to evaluate. The random forest classifier uses a default setting that considers the square root of the total number of features. Both the Gini impurity and the entropy can be used as the criterion for splitting, and this hyperparameter can be tuned to boost the efficiency of the model. Finding the best settings for these hyperparameters helps boost the random forest classifier’s overall efficiency and precision, and metaheuristic optimization methods can help one achieve just that. These methods can be useful for locating optimal solutions that may not be obtainable using more conventional approaches since they search across a wide range of possible values for the relevant hyperparameters. These parameters are classified using the proposed optimization algorithm presented in Algorithm 2.
Algorithm 2: The proposed DBERDTO optimization algorithm
1:
Initialize DBERDTO population P i ( i = 1 , 2 , , d ) with size d, iterations M a x i t e r , fitness function F n , t = 1 , n 1 , n 2 , a, r 1 , r 2 , r 3 , r 4 , r 5
2:
Calculate fitness function F n for each P i
3:
Find best solution as P *
4:
while  t M a x i t e r  do
5:
   if t%2 == 0 then
6:
     for ( i = 1 : i n 1 ) do
7:
        Update  r 1 = h c o s ( x ) 1 c o s ( x )
8:
        Calculate  D = r 1 ( P ( t ) 1 )
9:
        Update  P ( t + 1 ) = P ( t ) + D ( 2 r 2 1 )
10:
     end for
11:
     for ( i = 1 : i n 2 ) do
12:
        Calculate  D = r 3 ( L ( t ) P ( t ) )
13:
        Update positions of best solution as P ( t + 1 ) = r 2 ( P ( t ) + D )
14:
        Calculate  k = 1 + 2   ×   t 2 M a x i t e r 2
15:
        Investigate area around best solution as P ( t + 1 ) = r 1 ( P * ( t ) + k )
16:
        Compare  P ( t + 1 ) and P ( t + 1 ) to select best solution P *
17:
        if best fitness is not changed for last two iterations then
18:
          Mutate solution as P ( t + 1 ) = k z 2 h c o s ( x ) 1 c o s ( x )
19:
        end if
20:
     end for
21:
   else
22:
     for ( i = 1 : i n 2 ) do
23:
        if ( r 3 < 0.5 ) then
24:
          Update the individuals’ position as: P ( t + 1 ) = P b e s t ( t ) K 1 . | K 2 . P b e s t ( t ) P ( t ) |
25:
        else
26:
          Update the individuals’ velocity as: V ( t + 1 ) = K 3 V ( t ) + K 4 r 4 ( P b e s t ( t ) P ( t ) ) + K 5 r 5 ( P * P ( t ) )
27:
          Update the individuals’ position as: P ( t + 1 ) = P ( t ) + V ( t + 1 )
28:
        end if
29:
     end for
30:
   end if
31:
   Update the fitness function F n for each P ( t )
32:
   Update BER and DTO parameters at t = t + 1
33:
end while
34:
Return P *

3.8. The Classification Process Using Random Forest Classifier

The random forest classifier, following Algorithm 3, applies the bagging approach to each tree in the ensemble. Trees are fitted to this new, random sample instead of the training sample. One variable that can be learned automatically with the help of out-of-bag errors is the total number of trees used in the ensemble.
Algorithm 3: The classification of diabetes using random forest classification
1:
Initialize the selected feature vectors with n-samples and d-dimension
2:
while  b N ( b a g g i n g n u m b e r s )  do
3:
   Get a representative bootstrap ( X b , Y b )
4:
   Develop a random forest tree T b using ( X b , Y b )
5:
   while tree node size n m i n  do
6:
     Get random variables of size m
7:
     Select the most promising variables or split the m variables
8:
     Break the main node in the tree into two smaller nodes
9:
   end while
10:
   Result the trees ensemble as { T b } 1 N
11:
end while
12:
Return the posterior P ( x ) = V o t i n g P k ( x ) 1 N , where the classification of the kth RF is denoted by P k ( x )

4. Experimental Results

In this section, we use the numerical figures, confusion matrix, and charts that resulted from using the proposed architecture to classify breast cancer. The experiments were run using the improved CBIS-DDSM public dataset. The cross-validation value is set at 5, and the training/testing split is 70:30. Several optimization techniques, such as the whale optimization algorithm (WOA) [37], genetic algorithm (GA) [38], particle swarm optimization (PSO) [39], grey wolf optimization (GWO) [40], Al-Biruni earth radius (BER) [36] optimization, and the proposed advanced BER algorithm, have been used to optimize the parameters of the CNN. Several trials, including a deep feature classification on the original dataset and a deep feature classification using the improved CNN, are used to arrive at the final results.

4.1. The Diabetes Dataset

The Diabetes Dataset, denoted by D1 and available on Kaggle [41], contains data collected from 768 female patients of Pima Indian heritage residing in Arizona, USA. The dataset includes eight features or variables, including age, number of pregnancies, glucose level, insulin level, blood pressure, body mass index (BMI), diabetes pedigree function, and an outcome variable indicating whether or not the patient has diabetes. Age is a continuous variable representing the age of the patient in years. The number of pregnancies is an integer variable indicating the number of times the patient has been pregnant. Glucose level is a continuous variable representing the 2-h plasma glucose concentration in the patient’s blood. Insulin level is a continuous variable representing the serum insulin level in the patient’s blood. Blood pressure is a continuous variable representing the patient’s diastolic blood pressure in mm Hg. BMI is a continuous variable representing the body mass index of the patient. Diabetes pedigree function is a continuous variable representing the diabetes pedigree function for the patient, which provides information about the patient’s genetic predisposition for diabetes. Finally, the outcome variable is a binary variable indicating whether or not the patient has diabetes, with 1 representing the presence of diabetes and 0 representing the absence of diabetes. The Diabetes Dataset is a valuable resource for researchers and healthcare professionals interested in studying the risk factors associated with diabetes and developing better ways to prevent, manage, and treat the disease. By analyzing the relationships between these features and the outcome variable, researchers can identify important risk factors for diabetes and develop effective interventions and treatments to improve patient outcomes. The correlation among the features of the dataset is depicted in Figure 2, and the histogram of each feature vector is shown in the plots of Figure 3.

4.2. Feature Selection Evaluation Criteria

Table 2 provides the metrics by which the results have been evaluated. The performance of the suggested feature selection approach is measured against the metrics detailed in this table. The predicted values are denoted by V n ^ , while the observed values are shown by V n . The best solution at iteration j is represented by S j * , and the size of the best solution vector is denoted by s i z e ( S j * ) . M indicates the number of iterations of the proposed and other competing optimizers. A total of N points were used for the evaluation.

4.3. Classification Evaluation Criteria

The effectiveness of the proposed methods is measured using the benchmark metrics presented in Table 3. These metrics evaluate how well the proposed optimized CNN performs as a classification method. In the table, M represents the total number of iterations through an optimizer, g j * represents the optimal solution for iteration j, and s i z e ( g j * ) indicates the total length of the optimal solution vector. The number of data points in the test set, is denoted by N; the corresponding label, denoted by C i , is determined by the classifier used. The number of features, denoted by D, and the class label, L i , are two distinct quantities. True positive, true negative, false positive, and false negative abbreviations are T P , T N , F P , and F N , respectively.

4.4. Configuration Parameters

The configuration parameters of the employed optimization algorithms and the adopted machine learning models are presented in Table 4 and Table 5, respectively. In addition, the conducted experiments are operated based on 30 runs for the optimization algorithms, with 500 iterations in each run.

4.5. Results of Feature Selection

Using seven alternative metaheuristic optimization algorithms (bDBERDTO, bBER, bDTO, bPSO, bWAO, bGWO, and bFA), the authors report their findings about feature selection from diabetic features. The findings for each algorithm are summarized by their average select size. The average error results show the average error rate of each method, and they show that bDBERDTO has the lowest average error rate of 0.460, implying that the features selected by bDBERDTO can classify diabetes occurrences with a high degree of accuracy.
According to the results presented in Table 6, bWAO has the largest average select size (0.776), indicating that it has selected the most features on average. This suggests that bWAO is less efficient at reducing the dimensionality of the dataset, which may result in overfitting. The average fitness results show how each algorithm generally performs. Based on the results, bDBERDTO can be considered a powerful feature selection algorithm for diabetes classification, with an average fitness value of 0.523, which is significantly higher than the other algorithms. The highest fitness values show the greatest results for each algorithm. Based on the findings, bDBERDTO’s feature selection yields superior classification results to the other algorithms, with a best fitness value of 0.425. The lowest fitness values show the worst possible outcomes for each algorithm. Finally, the standard deviation fitness results show the variation in fitness values obtained by each algorithm, revealing that bFA has the worst fitness value of 0.606, indicating that bFA could not select a good subset of features that can perform well in diabetes classification. This suggests that the features chosen by bDBERDTO are more stable and can lead to consistent performance in diabetes classification, as evidenced by the lowest standard deviation of fitness values produced by bDBERDTO (0.346). As a whole, these findings are very suggestive of bDBERDTO’s potential as a feature selection method for diabetes classification.
The analysis of variance (ANOVA) is a statistical technique used to compare the means of two or more groups to establish statistical significance. The analysis was performed on a dataset of 69 samples from 7 groups, as presented in Table 7. With an F-value of 136.1 and a p-value of less than 0.0001, the results demonstrate that the Treatment factor significantly affected the data. This indicates a statistically significant change in the response variable due to the treatment and a difference in means between the groups. The error variance is estimated using the residual findings, which show the variation within each group. For this situation, 63 degrees of freedom (DF) were available, and the residual mean square was 0.00179. This suggests a tiny amount of diversity inside each group, but the variation across groups is much more significant. The results in the total column reflect the sum of all possible differences in the data, both between and within groups. The total number of degrees of freedom was 69, and the sum of squares (SS) was 0.02502. The analysis of variance test results indicates that the differences between the treatment groups are statistically significant and not coincidental. The relationship between therapies and the response variable can be better understood, and judgments regarding which treatment is most effective can be made with the help of this data.
A non-parametric alternative is the Wilcoxon signed-rank test to compare two related samples. For each of the seven feature selection techniques (bDBERDTO, bBER, bDTO, bPSO, bWAO, bGWO, and bFA), the test is used to compare the theoretical median values (assumed to be zero) with the actual median values presented in Table 8. Ten examples of each technique are used in the test. Each observation is given a rank, with higher ranks going to observations above the median and lower ranks going to observations below the median. These ranks are the “sum of signed ranks” (W). In this situation, the median values are continuously greater than the theoretical median of zero, as the sum of signed rankings is 55 across all 7 approaches. The total rankings for all “positive” and “negative” observations are summed to get the “sum of positive ranks” and “sum of negative ranks”, respectively. In this situation, all methods have median values greater than the theoretical median of zero, as seen by the total of the positive ranks being 55. No actual median values are below the theoretical median, as the sum of negative ranks is 0 for all techniques. Suppose that the null hypothesis (which here states that the observed median values are not significantly different from the theoretical median of zero) is correct. In that case, the “p-value” (two-tailed) is the probability of obtaining a test statistic as extreme as or more extreme than the observed test statistic. For all seven approaches, the P value is less than 0.002, suggesting sufficient evidence to reject the null hypothesis and infer that the observed median values differ substantially from the theoretical median. If the p-value was calculated exactly or estimated, as the information in the “Exact or estimate?” field. Here, the calculated P values are accurate. If you want to know if your results are statistically significant at the 0.05 level, enter that number into the “Significant (alpha = 0.05)?” column. Given that the P value is less than 0.05, these findings are statistically significant at the 5% level of confidence. Finally, the discrepancy between the observed median values and the ideal zero median is displayed in the “Discrepancy” section. In this category, the numbers represent the reported medians for each technique. The results show that, similar to every other feature selection approach evaluated, the suggested method (bDBERDTO) deviates significantly from the expected zero medians.
In Figure 4, the proposed method, bDBERDTO, has the lowest average error in a plot comparing the average errors achieved by other feature selection techniques. The plot also included the other methods bFA, bGWO, bWOA, bPSO, bDTO, and bBER, but none of them could match the performance of bDBERDTO. The plot shows that the proposed method is quite good at picking out the most important features in a dataset, which improves the overall performance of the system. It’s important to remember that even the other methods managed to attain low average mistakes, showing that they are not without merit. However, the recommended strategy emerges as the undisputed victor in this comparison. The ramifications of these findings for professionals in domains that use feature selection techniques to boost model performance are substantial. Compared to popular feature selection approaches, bDBERDTO is expected to produce even better outcomes.
In Figure 5, the residual, homoscedasticity, quartile-quartile (QQ), and heatmap plots are used to examine the ANOVA outcomes of the suggested feature selection technique. Residual and homoscedasticity plots are used to verify that the errors have the same variance and that the data is normally distributed. QQ plots are used to evaluate the normality of the residuals, while heatmap plots are used to see the connections between the chosen features. Providing that the assumptions are met, the residual plot, which displays the discrepancy between the observed and anticipated values, should exhibit no discernible pattern. The homoscedasticity plot’s residuals should be randomly dispersed around the horizontal line to ensure that the error variance is the same across all predictor variables levels. A straight line in the QQ plot indicates properly distributed residuals. We can gain insight into potential multicollinearity issues by displaying highly correlated features in a heatmap. These plots, when combined, give a thorough picture of the ANOVA results of the proposed feature selection approach, allowing practitioners to evaluate the assumptions and see any problems that may need fixing before interpreting the results.
Different classification algorithms were evaluated on a diabetes dataset consisting of patient medical records and other parameters used to determine diabetes risk, and the findings are presented in Table 9. Some performance indicators are p-values, F-scores, precision, sensitivity, and specificity. The p-value measures how likely it is to find a test statistic that is at least as outlandish as the one found in the dataset. The F-score examines how well true positive and false positive rates are balanced. F is the harmonic mean of precision and recall. The accuracy of a prediction system is measured by how many of those forecasts come true. An accurate positive rate is known as sensitivity, and an accurate negative rate is known as specificity. With an accuracy of 0.813, a sensitivity of 0.859, a specificity of 0.741, an F-score of 0.768, and a p-value of 0.840, random forest (RF) outperformed the other classification algorithms. Logistic regression (LR) came in second, scoring an F-score of 0.760 and a p-value of 0.800 with an accuracy of 0.787, sensitivity of 0.870, and specificity of 0.655. The support vector machine (SVM) results were similarly quite good: an F-score of 0.838, a p-value of 0.761, a sensitivity of 0.935, and a specificity of 0.534. Regarding accuracy and F-scores, k-nearest neighbors (KNN) and decision tree (DT) were less effective than RF, LR, and SVM. Stochastic Gradient Descent (SGD) performed the worst in accuracy, sensitivity, and F-score compared to other classifiers. These findings indicate that RF, LR, and SVM are superior to KNN and DT when predicting diabetes from the provided information.
After applying feature selection, the classification outcomes for a diabetes dataset are presented in Table 10. p-values, F-scores, degrees of precision, and specificity are all reported. Stochastic Gradient Descent (SGD), Decision Trees (DT), K-Nearest Neighbors (KNN), Gaussian Naive Bayes (GNB), Support Vector Machines (SVM), Logistic Regression (LR), and Random Forest (RF) were all employed as classification methods. A small p-value (0.588) indicates that the features used to train the SGD classifier are not very predictive of the outcome. Moderate F-score and accuracy (0.657) are accompanied by high sensitivity (0.909) and specificity (0.432) but low F-score (0.714) and accuracy (0.657). The p-value for the DT classifier is somewhat higher (0.682) than the p-value for SGD, but neither is statistically significant. Compared to SGD, the F-score and accuracy are better (0.779 and 0.73), although the sensitivity and specificity are the same (100). The KNN classifier’s greatest p-value (0.769) indicates that the selected features are more important in predicting the outcome than any other classifiers. High levels of sensitivity (0.943) and specificity (0.545) are accompanied by a high F-score (0.847) and high levels of accuracy (0.791). The GNB classifier has a moderately low p-value (0.806) and a high F-score (0.87) and accuracy (0.819). In contrast to the moderate specificity (0.6), the sensitivity is relatively high (0.943). With a p-value of 0.857, SVM is the second most accurate classifier, suggesting that the features used to make the prediction are crucial. Very high sensitivity (0.909) and moderate specificity (0.762) are accompanied by a high F-score (0.882) and high accuracy (0.852). High p-value (0.851), high F-score (0.889), and high accuracy (0.853) are all features of the LR classifier. Although the specificity is at 0.72, the sensitivity is very high at 0.93. Last but not least, the RF classifier excels in all three metrics studied here: F-score (0.909), accuracy (0.885), and sensitivity (0.909). The sensitivity (0.842) and the specificity (0.909) are pretty good. In conclusion, the KNN and RF classifiers achieved the highest F-scores, accuracy, sensitivity, and specificity following feature selection. It is important to highlight that the classifier selection is problem- and data-specific and that additional investigation may be required to identify the optimal model.
The analysis presented in Table 11 provides multiple statistical indicators of diabetes categorization model performance. There is valuable insight to be gleaned from each metric concerning the precision and consistency of the model’s predictions. According to the first metric, “Number of values”, there were ten occurrences across all seven categories. Although their precise meanings aren’t specified, we can infer that they correspond to different aspects of the model’s performance or the data used to train and evaluate it. Both the minimum and maximum values for a given category are indicated by the respective “Minimum” and “Maximum” indicators. For instance, the first group ranges from 0.986 to 0.992 for its values. Outliers can be found, and the range of values for each category is determined with these metrics. In order to determine how far any number can go, we can use the “Range” metric. There was a wide disparity between the highest and lowest values, as seen by the range of 0.030 in the fifth group. Data on the distribution of values within each category is made available through the “Percentile” measurements. For example, the “Median” value divides the data in half, while the “25% Percentile” reflects the number below which 25% of the data falls. The “75% Percentile” indicates the figure below which 75% of the data falls. These metrics can help show where the data is most and least concentrated and whether or not it is biased. Similar information is provided by the “10% Percentile” and “90% Percentile” measures, but for the 10th and 90th percentiles of the data, instead of the median and interquartile ranges. Information about the results’ reliability is provided by using the “Actual confidence level”, “Lower confidence limit”, and “Upper confidence limit” metrics.
These estimates are based on a confidence interval, a range of numbers thought to contain the actual value of the population parameter under study. With a confidence level of 97.85%, the actual number is likely inside the estimated margin of error. Indicators of central tendency, dispersion, and variability in data are provided by the “Mean”, “Std. Deviation”, “Std. Error of Mean”, and “Coefficient of Variation” measurements. The mean is the value that is most often encountered, whereas the standard deviation and standard error of the mean tell us about the variation in the data and the precision with which we may estimate the mean. The coefficient of variation is a measure of variability independent of measurement units and can be used to assess similarities and differences in data distribution across distinct groups. If the data is highly skewed, “Geometric Mean” measure can be used as an alternate central tendency measure. The “Geometric SD Factor” measures the dispersion of the data. At the same time, the “Lower 95% CI of geo. means” and “Upper 95% CI of geo. mean” give the minimum and maximum values of the 95% confidence interval for the geometric mean, respectively. The “Harmonic Mean” measures offer a different kind of central tendency metric that might be helpful when working with severely skewed data. Lower and upper limits of the 95% confidence interval for the harmonic mean are provided under the headings “Lower 95% CI of harm. mean” and “Upper 95% CI of harm. mean”, respectively. In cases where the data is highly skewed, the “Quadratic Mean” measurements provide yet another alternative central tendency measure. The lower and upper bounds of the 95% confidence interval for the quadratic mean are provided by the “Lower 95% CI of the quad. mean” and “Upper 95% CI of the quad. mean” headings, respectively.
One way to compare and contrast many groups or treatments is via an analysis of variance (ANOVA) test, which is presented in Table 12. The primary sections of the ANOVA table are labeled “Treatment”, “Residual”, and “Total”, respectively. The term “treatment” describes the differences in means between the various groups. The term “residual” is used to describe the ambiguous variation or inaccuracy that exists among the groups. What we mean by “total” here is the sum of the squares of all the observations or the entire amount of variability. The following is a breakdown of the metrics in each component of the ANOVA table:
Treatment = ( 0.02327 and 6 and 0.003878 and F (6, 63) = 211.8 and p < 0.0001):
  • SS (sum of squares) = 0.02327, representing the treatment groups’ variability;
  • DF (degrees of freedom) = 6, representing the number of groups being compared minus one;
  • MS (mean square) = 0.003878, representing the treatment group variance;
  • F (DFn, DFd) = 211.8, representing the F-statistic or the variance ratio between the treatment groups to the variance within the treatment groups;
  • p-value < 0.0001 represents the probability of observing such an extreme F-statistic or more powerful under the null hypothesis that no difference exists between the treatment groups.
Residual = ( 0.00115 and 63 and 0.000018):
  • SS = 0.00115 represents the unexplained variability or error within the groups;
  • DF = 63, representing the total number of observations minus the number of groups being compared;
  • MS = 0.000018, representing the treatment group variance.
Total = ( 0.02442 and 69):
  • SS = 0.02442 represents the total variability or sum of squares of all the observations;
  • DF = 69, representing the total number of observations minus one.
With an F-statistic of 211.8 and a p-value of 0.0001, the variance analysis shows a statistically significant difference between the treatment groups. The ANOVA table’s Treatment column indicates more variation between treatment groups than within them. In contrast, the Residual column suggests some variation or error within the groups that cannot be accounted for by the other two columns.
The Wilcoxon signed-rank test is a useful non-parametric option when comparing two groups with some features. The evaluation in this scenario aims to assess the relative merits of several machine learning approaches as presented in Table 13. If there is no difference between the two approaches, then the theoretical median is a vector of zeros. The empirical median equals the mean of the performance gaps between the various approaches and the theoretical median. Values equal the total number of test observations. In this scenario, there are 10 data points for each approach under consideration. Differences between the true and theoretical median can be quantified by adding up the signed rankings (W). When the rank is higher than zero, the real median exceeds the theoretical median; when the rank is lower than zero, the actual median falls short of the theoretical median. Considering that 55 is the sum of positive and negative ranks, we may conclude that all deviations from the theoretical median are positive. A test’s statistical significance is shown by its p-value. Evidence exists to reject the null hypothesis if the p-value is smaller than the significance level (alpha). All of the p-values shown are smaller than 0.05, indicating that there is indeed a statistically significant difference between the two approaches. The p-values are determined precisely, rather than being approximated, thanks to the accurate nature of the test. Positive signed ranks and a smaller difference between the actual and theoretical medians show that the proposed approach (DBERDTO + RF) outperforms the other methods by a wide margin.
The plot shown in Figure 6 comparing the classification accuracy attained by several strategies for diabetes case classification reveals that the proposed DBERDTO + RF strategy delivers the maximum accuracy. In addition to BER + RF, DTO + RF, PSO + RF, WOA + RF, GWO + RF, and FA + RF, we plotted these and other categorization approaches to see how they stacked up. The scatter plot shows that the proposed technique is very good at correctly categorizing instances of diabetes. Classification accuracy was still rather good for the remaining approaches, indicating that they had merit in their own right. Although the offered solution is not the only viable option, it is the most advantageous in this scenario. As correct diabetes case classification is crucial for both diagnosis and treatment, these findings have substantial implications for medical professionals. To get even better results than with other standard classification approaches, users can try the DBERDTO + RF strategy. The high accuracy of the proposed method has the potential to improve the diagnosis and treatment of diabetes, which in turn will enhance the lives of those who suffer from the disease.
As shown in Figure 7, the proposed DBERDTO + RF strategy achieves the highest accuracy compared to other methods in a histogram plot, comparing the classification accuracy attained by various approaches in classifying diabetes cases. Each method’s classification accuracy is displayed as a histogram, with the height of each bar indicating the frequency with which that accuracy number was obtained. The histogram shows that the most frequent accuracy values are concentrated at the maximum accuracy attained by the DBERDTO + RF method, confirming the superiority of the suggested method. This visualization helps practitioners select the most efficient way for their specific use case and further demonstrates the proposed strategy’s superiority in accurately identifying diabetes patients. On the other hand, the plots shown in Figure 8 show the significance of the proposed approach in classifying diabetes cases compared to the other methods.

4.6. Discussion

The Dynamic Al-Biruni Earth Radius and Dipper Throated Optimization (DBERDTO) algorithm is used in the proposed method for diabetes classification, and it shows promise in both feature selection and classification. The method outperforms competing modern optimization techniques by using the binary version of DBERDTO for feature selection and the continuous version for optimizing the parameters of the random forest classifier. Feature selection plays a crucial role in selecting the most informative features for a classification task. When applied to feature selection, the binary version of DBERDTO shows superior results, demonstrating its efficacy in discovering important features for diabetes categorization. DBERDTO effectively searches the feature space and picks a subset of features that maximizes the classification performance by utilizing the algorithm’s distinctive exploration and exploitation methodologies. DBERDTO’s flexibility in balancing worldwide exploration with local exploitation is one of its greatest strengths. The algorithm can successfully navigate the feature space since it considers the variety of features and their local correlations. Therefore, the proposed method benefits from a strong feature subset that captures the crucial discriminatory data about diabetes. To maximize the performance of a random forest classifier, one should fine-tune its settings. This method successfully tunes the parameters of a random forest using a continuous version of DBERDTO. DBERDTO optimizes the random forest parameters to improve the classification performance by adjusting its search behavior dynamically throughout the optimization process. Random forests are an ensemble classifier combining the predictive power of several decision trees. Selecting the correct hyperparameters, such as the number of trees, maximum depth, and split criterion, is crucial to the performance of a random forest. The proposed method employs DBERDTO, automatically tweaking these settings to produce optimal classification outcomes.
Compared to other modern optimization techniques, the proposed method based on DBERDTO performs better in feature selection and classification tasks for diabetes classification. There are several important reasons why this strategy produces better results. The first distinguishing feature of DBERDTO is its capacity to efficiently search the solution space and settle on good solutions by modifying its search behavior. This flexibility allows the algorithm to locate optimal feature subsets and parameter combinations despite potential optimization problems. The proposed method provides a complete optimization framework because it combines the binary and continuous forms of DBERDTO. The strategy takes advantage of the strengths of both the binary and continuous versions by using the former for feature selection and the latter for optimizing parameters. The method enhances computational efficiency and generalization performance by picking the most informative features within the dataset. DBERDTO’s feature subset selection collects the most important discriminatory information, increasing classification accuracy. The efficiency of the proposed method is further demonstrated by the superiority of the classification results achieved using the method. DBERDTO-optimized random forest parameters improve the accuracy, sensitivity, and specificity of diabetes detection classification. Accurate diabetes classification is essential for better patient treatment; these enhanced performance indicators are a key part of that. The DBERDTO algorithm, used in the proposed method for diabetes classification, shows substantial advantages over other contemporary optimization techniques. DBERDTO’s binary implementation allows for enhanced feature selection, accurately determining the most important features. The classification accuracy, sensitivity, and specificity are all improved in the continuous version by optimizing the random forest settings. The proposed approach is superior because of the DBERDTO algorithm’s novel features, such as its dynamic search behavior and integration of global exploration and local exploitation tactics. These features help it identify the best possible classification methods for diabetic patients.
The selected features obtained through the proposed feature selection method demonstrate their potential to improve the performance of the optimized random forest classifier on the adopted dataset. The results indicate that these selected features possess discriminative power and contribute significantly to the classification task. By focusing on relevant features, the classifier can make more accurate classifications and achieve higher performance metrics such as precision, recall, and accuracy. The implications of diabetes diagnostics are significant in terms of public health and individual well-being. Accurate and reliable diagnostic methods can aid in the early detection of diabetes, enabling timely intervention and management. With the selected features improving the performance of the classifier, it suggests that these features have strong associations with diabetes-related patterns or risk factors. This insight can help healthcare professionals and researchers better understand the underlying factors contributing to diabetes and develop targeted interventions. Moreover, the improved performance of the classifier implies the potential for creating efficient and reliable diagnostic tools for diabetes. By identifying the most informative features, future diagnostic models and algorithms can be developed that focus on these key factors. This can lead to the development of cost-effective, non-invasive, and accessible diagnostic methods, facilitating the early detection and proactive management of diabetes. However, it is essential to note that the discussion of the selected features and their potential performance improvements should be interpreted within the context of the adopted dataset and the applied feature selection methods. To validate these findings on larger and more diverse datasets to ensure the generalizability and robustness of the selected features in different populations and settings, three other datasets and the achieved results and findings are discussed in the Appendix A.

5. Conclusions and Future Perspectives

In this paper, we proposed a novel metaheuristic optimization-based method for improving diabetes classification. A novel feature selection algorithm is developed using a dynamic Al-Biruni earth radius and throated dipper optimization (DBERDTO) algorithm. Following feature selection, the proposed DBERDTO is used to optimize the parameters of a random forest classifier before applying it to the dataset. To demonstrate the efficacy and superiority of the suggested methodology, it is compared and evaluated against state-of-the-art optimization techniques and machine learning models. The proposed method can classify diabetes cases with an overall accuracy of 98.6%. Analysis of variance (ANOVA) and Wilcoxon signed-rank tests were performed, among others, to gauge the significance and difference of the proposed approach. The outcomes of the tests corroborated the predicted results. The proposed approach and its possible application to other medical datasets will be studied in greater depth in future research. In addition, there are a variety of data-balancing methods that can deal with outliers and will be considered in future work.

Author Contributions

Methodology, A.A.A. (Abdelaziz A. Abdelhamid), S.K.T., A.I. and M.M.E.; Software, A.A.A. (Abdelaziz A. Abdelhamid), S.K.T., A.I., M.M.E. and M.S.S.; Data curation, M.S.S.; Writing—original draft, A.A.A. (Abdelaziz A. Abdelhamid), S.K.T., A.I. and M.M.E.; Writing—review & editing, A.A.A. (Amel Ali Alhussan) and D.S.K.; Funding acquisition, A.A.A. (Amel Ali Alhussan) and D.S.K. All authors have read and agreed to the published version of the manuscript.

Funding

Princess Nourah bint Abdulrahman University Researchers Supporting Project number (PNURSP2023R 308), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Acknowledgments

Princess Nourah bint Abdulrahman University Researchers Supporting Project number (PNURSP2023R 308), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia.

Conflicts of Interest

The authors declare that they have no conflict of interest to report regarding the present study.

Appendix A

In this appendix, the proposed feature selection method is applied to three other datasets to prove its generalization and superiority in the classification of diabetes. The first dataset, denoted by D2, is originally from the National Institute of Diabetes and Digestive and Kidney Diseases [50]. The second dataset, denoted by D3, is collected from the Iraqi society, as the data were acquired from the laboratory of Medical City Hospital [51]. The third dataset, denoted by D4, is originally from the National Institute of Diabetes and Digestive and Kidney Diseases [52]. In the next sections, the proposed approach is evaluated in terms of the adopted criteria.

Appendix A.1. The Achieved Results Using Dataset D2

The proposed feature selection method bDBERDTO has been used for selecting the significant set of features in dataset D2 and the results are presented in Table A1. The results of the feature selection analysis using the bDBERDTO method on dataset D2, along with various other methods, including bBER, bDTO, bPSO, bWAO, bGWO, and bFA, reveal interesting insights. The bDBERDTO method demonstrates its effectiveness by achieving a notably low average error of 0.819, indicating improved accuracy in selecting relevant features. Moreover, bDBERDTO stands out by selecting a comparatively smaller subset of features, as evidenced by its average select size of 0.772. This suggests a more efficient and concise feature selection process. When considering the average fitness value, bDBERDTO attains a respectable score of 0.882, indicating strong overall performance. Notably, the best fitness value obtained by bDBERDTO is 0.784, signifying its ability to identify the most optimal subset of features within the dataset. Even in the worst-case scenarios, bDBERDTO maintains competitive performance with the worst fitness value of 0.882. The standard deviation of fitness, which is 0.704 for bDBERDTO, indicates consistent and stable results across different iterations. These results collectively showcase the bDBERDTO method’s efficacy in achieving low average error, selecting an optimal subset of features, and maintaining competitive fitness values.
Table A1. Evaluation of the results of the proposed feature selection method when applied to dataset D2.
Table A1. Evaluation of the results of the proposed feature selection method when applied to dataset D2.
D2bDBERDTObBERbDTObPSObWAObGWObFA
Avg. error0.8190.8360.8500.8700.8700.8560.868
Avg. Select size0.7720.9720.9140.9721.1350.8941.006
Avg. Fitness0.8820.8980.9100.8970.9040.9040.949
Best Fitness0.7840.8190.8130.8770.8690.8820.867
Worst Fitness0.8820.8850.9280.9450.9450.9580.965
Std. Fitness0.7040.7090.7110.7080.7110.7100.745
The classification results of the diabetes dataset D2, before applying feature selection, provide valuable insights into the performance of different classifiers as presented in Table A2. Among the classifiers evaluated, Random Forest demonstrates the highest positive predictive value (PPV) of 0.830, indicating its ability to identify positive instances correctly. Additionally, Random Forest exhibits a high negative predictive value (NPV) of 0.838, indicating its proficiency in accurately identifying negative instances. The F-score, which considers both precision and recall, is relatively high for Gaussian NB, reaching 0.686, suggesting its balanced performance. Accuracy, a measure of overall correctness, is highest for SVC, achieving a value of 0.770. Sensitivity, or true positive rate, is notably high for K-Neighbors, indicating its ability to classify positive instances correctly. On the other hand, specificity, or true negative rate, is highest for SGD, demonstrating its capability to classify negative instances accurately. Overall, these classification results provide a comprehensive overview of the performance of different classifiers on the diabetes dataset D2, serving as a baseline for evaluating the impact of feature selection techniques on their performance.
Table A2. Evaluation of the classifier results before applying feature selection in terms of dataset D2.
Table A2. Evaluation of the classifier results before applying feature selection in terms of dataset D2.
D2PPVNPVF-ScoreAccuracySensitivitySpecificity
SGD0.7200.4940.4350.5260.3750.766
Decision Tree0.7460.7210.5710.6710.6970.630
K-Neighbors0.7160.7530.6170.6840.7940.511
Gaussian NB0.8030.7980.6860.7570.7940.698
SVC0.7510.8280.8270.7700.9230.528
Logistic Regression0.7900.8220.7500.7760.8580.647
Random Forest0.8300.8380.7580.8030.8480.732
After applying feature selection to the diabetes dataset D2, the classification results demonstrate improvements in the performance of the evaluated classifiers. The classification results are shown in Table A3 and in Figure A1. SGD, which initially had a relatively low PPV, shows an increase to 0.596, indicating better identification of positive instances. Notably, Decision Tree substantially enhances its F-score, reaching 0.789, suggesting improved precision and recall balance. K-Neighbors demonstrate remarkable progress in multiple performance metrics, including an elevated PPV of 0.779, NPV of 0.868, and F-score of 0.859, indicating its ability to classify both positive and negative instances accurately. Gaussian NB showcases significant improvements in all metrics, with high PPV, NPV, and F-score values of 0.817, 0.868, and 0.881, respectively. SVC demonstrates enhanced performance in terms of PPV, NPV, and F-score, indicating improved accuracy and reliability. Logistic Regression exhibits a high F-score of 0.901, emphasizing its ability to balance precision and recall. Finally, Random Forest maintains its strong performance, achieving a high PPV of 0.921 and an F-score of 0.921, highlighting its effectiveness in accurately classifying instances. Overall, the feature selection process enhances the performance of the classifiers, resulting in improved accuracy and precision in classifying instances in the diabetes dataset D2.
Table A3. Evaluation of the classifier results after applying feature selection using the proposed bDBERDTO in terms of dataset D2.
Table A3. Evaluation of the classifier results after applying feature selection using the proposed bDBERDTO in terms of dataset D2.
D2PPVNPVF-ScoreAccuracySensitivitySpecificity
SGD0.5960.8530.7240.6660.9210.438
Decision Tree0.6910.8530.7890.7400.9210.540
K-Neighbors0.7790.8680.8590.8010.9560.553
Gaussian NB0.8170.8680.8810.8300.9560.608
SVC0.8680.8530.8940.8630.9210.772
Logistic Regression0.8620.8680.9010.8640.9420.729
Random Forest0.9210.8530.9210.8960.9210.853
Figure A1. The classification accuracy of diabetes in dataset D2 after feature selection using the proposed approach.
Figure A1. The classification accuracy of diabetes in dataset D2 after feature selection using the proposed approach.
Diagnostics 13 02038 g0a1
The ANOVA test was conducted on the classification results of the optimized Random Forest classifier for the diabetes dataset D2 and the results are presented in Table A4 and in Figure A2. The test evaluated the sources of variation between columns (treatments) and within columns (residuals) to determine their significance in explaining the overall variation. The treatment analysis showed a sum of squares (SS) of 0.027, degrees of freedom (DF) of 6, and mean square (MS) of 0.0046. The F-statistic (6, 63) = 100.3 indicated a highly significant difference between the treatment means. The corresponding p-value was found to be less than 0.0001, further confirming the statistical significance of the observed differences. The residual analysis, representing the unexplained variation within columns, yielded an SS of 0.003, DF of 63, and MS of 0.00005. The total variation accounted for a cumulative SS of 0.031 and a DF of 69. These results indicate that the optimized Random Forest classifier for the diabetes dataset D2 exhibits significant differences in performance across the evaluated treatments, highlighting the impact of feature selection on the classification outcomes.
Table A4. ANOVA test applied to the results of the optimized RF classifier in terms of the dataset D2.
Table A4. ANOVA test applied to the results of the optimized RF classifier in terms of the dataset D2.
D2SSDFMSF (DFn, DFd)p-Value
Treatment (between columns)0.02760.0046F (6, 63) = 100.3p < 0.0001
Residual (within columns)0.003630.00005
Total0.03169
Figure A2. Visualizing ANOVA test results applied to the optimized RF classifier outcomes.
Figure A2. Visualizing ANOVA test results applied to the optimized RF classifier outcomes.
Diagnostics 13 02038 g0a2
The Wilcoxon test was performed on the classification results of the optimized Random Forest classifier for the diabetes dataset D2, comparing different feature selection methods combined with Random Forest (RF). The results of this test are presented in Table A5. The test aimed to determine if there were significant differences in the performance of these combinations compared to the theoretical median of 0. The median values for DBERDTO + RF, BER + RF, DTO + RF, PSO + RF, WAO + RF, GWO + RF, and FA + RF were found to be 0.986, 0.941, 0.933, 0.933, 0.922, 0.913, and 0.926, respectively. The Wilcoxon test statistics showed that the sum of signed ranks for each combination was 55, indicating no significant deviations from the median. The sum of positive ranks and the sum of negative ranks were both 55, suggesting a balanced distribution. The resulting p-values for all combinations were calculated as 0.002, less than the significance level of 0.05. Therefore, it can be concluded that there is strong evidence to reject the null hypothesis and assert that these feature selection methods combined with Random Forest significantly outperformed the theoretical median of 0 in the classification of the diabetes dataset D2. In addition, the statistical results presented in Table A6 demonstrate an in-depth investigation of the statistical significance of the proposed methodology.
Table A5. Wilcoxon test applied to the results of the optimized RF classifier in terms of dataset D2.
Table A5. Wilcoxon test applied to the results of the optimized RF classifier in terms of dataset D2.
D2DBERDTO + RFBER + RFDTO + RFPSO + RFWAO + RFGWO + RFFA + RF
Theoretical median0000000
Median0.9860.9410.9330.9330.9220.9130.927
Number of values10101010101010
Sum of signed ranks55555555555555
Sum of +ve ranks55555555555555
Sum of −ve ranks0000000
p-value0.0020.0020.0020.0020.0020.0020.002
Discrepancy0.9860.9410.9330.9330.9220.9130.927
Table A6. Statistical analysis of the classification results generated by the optimized RF classifier applied to dataset D2.
Table A6. Statistical analysis of the classification results generated by the optimized RF classifier applied to dataset D2.
D2DBERDTO + RFBER + RFDTO + RFPSO + RFWAO + RFGWO + RFFA + RF
Number of values10.00010.00010.00010.00010.00010.00010.000
Minimum0.9660.9310.9330.9230.9220.9130.906
25% Percentile0.9750.9410.9330.9330.9220.9130.919
Median0.9860.9410.9330.9330.9220.9130.927
75% Percentile0.9860.9460.9360.9330.9300.9160.927
Maximum0.9860.9610.9460.9430.9320.9330.927
Range0.0200.0300.0120.0200.0100.0200.021
10% Percentile0.9660.9320.9330.9240.9220.9130.906
90% Percentile0.9860.9610.9450.9420.9320.9320.927
Actual confidence level97.85%97.85%97.85%97.85%97.85%97.85%97.85%
Lower confidence limit0.9720.9410.9330.9330.9220.9130.907
Upper confidence limit0.9860.9590.9430.9330.9300.9230.927
Mean0.9800.9440.9360.9330.9250.9160.922
Std. Deviation0.0070.0090.0050.0050.0040.0070.008
Std. Error of Mean0.0020.0030.0010.0010.0010.0020.003
Lower 95% CI of mean0.9750.9380.9320.9300.9220.9120.916
Upper 95% CI of mean0.9860.9510.9390.9370.9280.9210.928
Coefficient of variation0.7643%0.9626%0.5033%0.5051%0.4566%0.7366%0.9215%
Geometric mean0.9800.9440.9360.9330.9250.9160.922
Geometric SD factor1.0081.0101.0051.0051.0051.0071.009
Lower 95% CI of geo. mean0.9750.9380.9320.9300.9220.9120.916
Upper 95% CI of geo. mean0.9860.9500.9390.9370.9280.9210.928
Harmonic mean0.9800.9440.9360.9330.9250.9160.922
Lower 95% CI of harm. mean0.9750.9380.9320.9300.9220.9120.916
Upper 95% CI of harm. mean0.9860.9500.9390.9370.9280.9210.928
Quadratic mean0.9800.9440.9360.9330.9250.9160.922
Lower 95% CI of quad. mean0.9750.9380.9320.9300.9220.9110.916
Upper 95% CI of quad. mean0.9860.9510.9390.9370.9280.9210.928
Skewness−0.9911.1681.8320.0001.1022.277−1.702
Kurtosis−0.3921.0151.7774.500−0.8914.7651.228
Sum9.8029.4409.3559.3339.2459.1639.220

Appendix A.2. The Achieved Results Using Dataset D3

Similarly, the proposed feature selection method is applied to the features in the third dataset denoted by D3. The evaluation results of the selected features are presented in Table A7. The feature selection evaluation using the proposed bDBERDTO method for the diabetes dataset D3 provides valuable insights into its performance. The average error rates obtained using various feature selection algorithms range from 0.738 to 0.787, with bDBERDTO achieving an average error rate of 0.738. The average select size represents the number of features selected, ranging from 0.691 to 1.054, with bWAO having the largest average select size of 1.054. The average fitness scores, which measure the quality of the selected features, range from 0.801 to 0.868. The best fitness scores indicate the highest achieved fitness for each algorithm, ranging from 0.703 to 0.786. Conversely, the worst fitness scores range from 0.801 to 0.884. The standard deviation of fitness scores ranges from 0.623 to 0.664, indicating the degree of variability in the performance of the algorithms. Overall, the bDBERDTO method demonstrates competitive performance across various evaluation metrics, showcasing its potential as an effective feature selection technique for the diabetes dataset D3.
Table A7. Evaluation of the results of the proposed feature selection method when applied to dataset D3.
Table A7. Evaluation of the results of the proposed feature selection method when applied to dataset D3.
D3bDBERDTObBERbDTObPSObWAObGWObFA
Avg. error0.7380.7550.7690.7890.7890.7750.787
Avg. Select size0.6910.8910.8330.8911.0540.8140.925
Avg. Fitness0.8010.8170.8290.8160.8240.8230.868
Best Fitness0.7030.7380.7320.7960.7880.8010.786
Worst Fitness0.8010.8050.8470.8640.8640.8770.884
Std. Fitness0.6230.6280.6300.6280.6300.6290.664
The evaluation of classifier results for the dataset D3 prior to applying feature selection provides valuable insights into their performance as presented in Table A8. The SGD classifier achieved an accuracy of 0.480, indicating that it correctly classified approximately 48% of the instances. The Decision Tree classifier performed slightly better, with an accuracy of 0.612. The K-Neighbors classifier achieved an accuracy of 0.624, demonstrating its ability to classify instances with higher accuracy. The Gaussian NB classifier exhibited improved performance, with an accuracy of 0.690, indicating its effectiveness in handling the features of the dataset. The SVC classifier achieved an accuracy of 0.702, demonstrating its ability to classify instances with higher accuracy correctly. The Logistic Regression classifier achieved an accuracy of 0.708, while the Random Forest classifier achieved an accuracy of 0.732, showcasing their ability to classify instances more accurately than previous classifiers. Overall, the evaluation of classifier results for dataset D3 highlights the varying performance levels of different classifiers and provides valuable insights for further analysis and improvement.
Table A8. Evaluation of the classifier results before applying feature selection in terms of dataset D3.
Table A8. Evaluation of the classifier results before applying feature selection in terms of dataset D3.
D3PPVNPVF-ScoreAccuracySensitivitySpecificity
SGD0.6560.4500.3970.4800.3420.698
Decision Tree0.6800.6570.5200.6120.6360.574
K-Neighbors0.6530.6870.5620.6240.7240.465
Gaussian NB0.7320.7280.6250.6900.7240.636
SVC0.6850.7550.7540.7020.8410.481
Logistic Regression0.7200.7500.6840.7080.7830.590
Random Forest0.7560.7640.6910.7320.7730.667
The evaluation of classifier results after applying feature selection using the proposed bDBERDTO method on dataset D3 provides valuable insights into the performance of the classifiers as presented in Table A9 and in Figure A3. The SGD classifier achieved a precision of 0.529, indicating its ability to classify positive instances correctly, and a sensitivity of 0.818, suggesting its effectiveness in identifying true positive instances. However, its specificity was lower at 0.389, indicating a higher rate of false positives. The Decision Tree classifier showed improvements with a precision of 0.614, sensitivity of 0.818, and specificity of 0.480, indicating a more balanced performance. The K-Neighbors classifier demonstrated even better performance with a precision of 0.692, sensitivity of 0.849, and specificity of 0.491. The Gaussian NB classifier improved the results with a precision of 0.726, sensitivity of 0.849, and specificity of 0.540. The SVC classifier achieved high precision (0.771) and specificity (0.686), accurately classifying instances in both positive and negative classes. The Logistic Regression classifier achieved a precision of 0.766, sensitivity of 0.837, and specificity of 0.648, indicating its balanced performance. Finally, the Random Forest classifier demonstrated excellent performance across all metrics, with precision, sensitivity, and specificity of 0.889. Overall, the evaluation of classifier results after applying feature selection using the bDBERDTO method on dataset D3 showcases the improvement in classifier performance, particularly in terms of precision, sensitivity, and specificity, leading to more accurate and reliable classifications.
Table A9. Evaluation of the classifier results after applying feature selection using the proposed bDBERDTO in terms of dataset D3.
Table A9. Evaluation of the classifier results after applying feature selection using the proposed bDBERDTO in terms of dataset D3.
D3PPVNPVF-ScoreAccuracySensitivitySpecificity
SGD0.5290.7580.6430.5910.8180.389
Decision Tree0.6140.7580.7010.6570.8180.480
K-Neighbors0.6920.7710.7630.7120.8490.491
Gaussian NB0.7260.7710.7830.7370.8490.540
SVC0.7710.7580.7940.7670.8180.686
Logistic Regression0.7660.7710.8000.7680.8370.648
Random Forest0.8890.8230.8890.8650.8890.823
Figure A3. The classification accuracy of diabetes in dataset D3 after feature selection using the proposed approach.
Figure A3. The classification accuracy of diabetes in dataset D3 after feature selection using the proposed approach.
Diagnostics 13 02038 g0a3
The ANOVA test was conducted on the results of the classification using the optimized random forest classifier on dataset D3 as presented in Table A10 and in Figure A4. The test aimed to analyze the variation between the different treatments (classifier methods) and within the columns (residuals). The results showed that the treatment (between columns) accounted for significant variation, as indicated by the relatively high sum of squares (SS) value of 0.022. With 6 degrees of freedom (DF), the mean squares (MS) value was calculated as 0.0036. The F-value, obtained by dividing the treatment MS by the residual MS, was determined to be F (6, 63) = 17.07, indicating a significant difference between the treatment groups. The p-value associated with the F-test was less than 0.0001, further confirming the statistical significance of the results. In contrast, the residual (within columns) accounted for a lower variation, as indicated by the SS value of 0.013 and 63 degrees of freedom. The total sum of squares was determined to be 0.035, representing the overall variation in the dataset. These ANOVA test results provide evidence that the different classifier methods significantly impacted the classification results, underscoring the importance of selecting an optimized random forest classifier for accurate predictions on the D3 dataset.
Table A10. ANOVA test applied to the results of the optimized RF classifier in terms of dataset D3.
Table A10. ANOVA test applied to the results of the optimized RF classifier in terms of dataset D3.
D3SSDFMSF (DFn, DFd)p-Value
Treatment (between columns)0.02260.0036F (6, 63) = 17.07p < 0.0001
Residual (within columns)0.013630.0002
Total0.03569
Figure A4. Visualizing ANOVA test results when applied to the results of the optimized RF classifier.
Figure A4. Visualizing ANOVA test results when applied to the results of the optimized RF classifier.
Diagnostics 13 02038 g0a4
The Wilcoxon test was performed on the results of the optimized random forest classifier on the D3 dataset to assess the statistical significance of the differences between different feature selection methods as presented in Table A11. The test was conducted by comparing the performance of the feature selection methods, including DBERDTO + RF, BER + RF, DTO + RF, PSO + RF, WAO + RF, GWO + RF, and FA + RF. The theoretical median for all methods was set to 0. The actual medians were calculated as 0.9689, 0.9433, 0.9381, 0.9331, 0.9181, 0.9073, and 0.9027 for each method. The Wilcoxon test was applied to the dataset with 10 values, resulting in a sum of signed ranks of 55. All ranks were positive, with a sum of positive ranks also equal to 55. The sum of negative ranks was 0, indicating no negative differences. The p-value for the test was determined to be 0.002, suggesting a statistically significant difference between the feature selection methods. The discrepancy values matched the actual medians for each method. Overall, these Wilcoxon test results provide evidence of significant differences among the feature selection methods when combined with the optimized random forest classifier on the D3 dataset.
The statistical analysis of the results achieved by the optimized random forest classifier on the D3 dataset provides insights into the performance of various feature selection methods as presented in Table A12. The analysis reveals the minimum, median, range, and percentile values for each method, indicating variations in performance. Mean values, standard deviation, and confidence intervals are computed, offering a comprehensive understanding of the spread and variation of the results. Other measures such as coefficient of variation, geometric mean, harmonic mean, quadratic mean, skewness, kurtosis, and sum provide additional information about the data distribution and overall classifier performance. This analysis aids in informed decision-making regarding selecting the most effective feature selection approach.
Table A11. Wilcoxon test applied to the results of the optimized RF classifier in terms of dataset D3.
Table A11. Wilcoxon test applied to the results of the optimized RF classifier in terms of dataset D3.
D3DBERDTO + RFBER + RFDTO + RFPSO + RFWAO + RFGWO + RFFA + RF
Theoretical median0000000
Actual median0.9690.9430.9380.9330.9180.9070.903
Number of values10101010101010
Sum of signed ranks55555555555555
Sum of +ve ranks55555555555555
Sum of −ve ranks0000000
p-value0.0020.0020.0020.0020.0020.0020.002
Discrepancy0.9690.9430.9380.9330.9180.9070.903
Table A12. Statistical analysis of the classification results generated by the optimized RF classifier applied to dataset D3.
Table A12. Statistical analysis of the classification results generated by the optimized RF classifier applied to dataset D3.
D3DBERDTO + RFBER + RFDTO + RFPSO + RFWAO + RFGWO + RFFA + RF
Number of values10101010101010
Minimum0.95890.90330.91810.91810.90810.90730.9027
25% Percentile0.96890.94080.92560.91810.90810.90730.9027
Median0.96890.94330.93810.93310.91810.90730.9027
75% Percentile0.96890.94330.93810.93810.93810.90730.9265
Maximum0.96890.95330.95810.93810.93810.920.9357
Range0.0100.0500.0400.0200.0300.012670.033
10% Percentile0.95990.90630.91810.91810.90810.90730.9027
90% Percentile0.96890.95230.95610.93810.93810.91870.9347
Actual confidence level97.85%97.85%97.85%97.85%97.85%97.85%97.85%
Lower confidence limit0.96890.93330.91810.91810.90810.90730.9027
Upper confidence limit0.96890.94330.93810.93810.93810.90730.9265
Mean0.96790.93930.93510.92910.92210.90860.9107
Std. Deviation0.00320.01350.01160.00990.01430.004010.0132
Std. Error of Mean0.0010.00430.00370.00320.00450.00130.0042
Lower 95% CI of mean0.96560.92960.92680.9220.91190.90570.9013
Upper 95% CI of mean0.97020.9490.94340.93620.93230.91140.9202
Coefficient of variation0.3267%1.437%1.240%1.070%1.551%0.4411%1.453%
Geometric mean0.96790.93920.9350.92910.9220.90860.9106
Geometric SD factor1.0031.0151.0121.0111.0161.0041.015
Lower 95% CI of geo. mean0.96560.92950.92680.9220.91180.90570.9013
Upper 95% CI of geo. mean0.97020.94910.94340.93620.93230.91140.9201
Harmonic mean0.96790.93910.9350.9290.92190.90860.9105
Lower 95% CI of harm. mean0.96560.92930.92680.92190.91180.90570.9013
Upper 95% CI of harm. mean0.97020.94920.94330.93620.93220.91140.92
Quadratic mean0.96790.93940.93520.92910.92220.90860.9108
Lower 95% CI of quad. mean0.96560.92980.92680.9220.91190.90570.9012
Upper 95% CI of quad. mean0.97020.94890.94340.93620.93240.91150.9203
Skewness−3.162−2.4660.1924−0.23730.25093.1621.18
Kurtosis10.06.9741.092−2.30−2.16510.0−0.4931
Sum9.6799.3939.3519.2919.2219.0869.107

Appendix A.3. The Achieved Results Using Dataset D4

Finally, the proposed approach is evaluated in terms of a dataset addressed by several researchers in the literature [53,54,55,56]. This dataset is originally from the National Institute of Diabetes and Digestive and Kidney Diseases [52]. The proposed approach is used to select the best set of features to boost classification accuracy. The results are evaluated using the adopted metrics, and various statistical tests were conducted to study the statistical difference and significance of the proposed approach. Feature selection results in terms of dataset D4 are presented in Table A13. As presented in this table, the proposed bBERDTO achieves the best results when compared to the other optimization methods.
Table A13. Evaluation of the results of the proposed feature selection method when applied to dataset D4.
Table A13. Evaluation of the results of the proposed feature selection method when applied to dataset D4.
D4bDBERDTObBERbDTObPSObWAObGWObFA
Average error0.7180.7350.7490.7690.7690.7550.768
Average Select size0.6710.8710.8130.8711.0340.7940.905
Average Fitness0.7810.7980.8090.7960.8040.8040.848
Best Fitness0.6830.7180.7120.7760.7680.7810.767
Worst Fitness0.7820.7850.8270.8440.8440.8580.864
Std Fitness0.6040.6080.6100.6080.6100.6090.645
To determine the best classification model for this dataset, a set of experiments is conducted to assess seven machine learning models, and the results before feature selection and after feature selection are presented in Table A14 and Table A15, respectively. As presented in these tables, the Random Forest classifier is the best classifier, which achieves the best results before and after feature selection.
Table A14. Evaluation of the classifier results before applying feature selection in terms of dataset D4.
Table A14. Evaluation of the classifier results before applying feature selection in terms of dataset D4.
D4PPVNPVF-ScoreAccuracySensitivitySpecificity
SGD0.7130.4890.4310.5210.3720.759
Decision Tree0.7390.7140.5650.6650.6910.624
K-Neighbors0.7090.7460.6110.6780.7860.506
Gaussian NB0.7950.7910.6790.7500.7860.691
SVC0.7440.8200.8190.7630.9140.523
Logistic Regression0.7820.8150.7430.7690.8500.641
Random Forest0.8220.8310.7510.7950.8400.725
Table A15. Evaluation of the classifier results after applying feature selection in terms of dataset D4.
Table A15. Evaluation of the classifier results after applying feature selection in terms of dataset D4.
D4PPVNPVF-ScoreAccuracySensitivitySpecificity
SGD Classifier0.5760.8240.6990.6430.8900.423
Decision Tree Classifier0.6670.8240.7630.7150.8900.522
K-Neighbors Classifier0.7530.8390.8300.7740.9230.534
Gaussian NB0.7890.8390.8510.8020.9230.587
SVC0.8390.8240.8640.8340.8900.746
Logistic Regression0.8330.8390.8700.8350.9110.705
Random Forest Classifier0.8900.8240.8900.8660.8900.824
The accuracy of the optimized random forest classifier fed with the best set of features is depicted in the plot shown in Figure A5. In this plot, the average accuracy achieved by the proposed approach is 98.9%. This result is superior to that of the other competing approaches proposed by the authors in Refs. [53,54,55,56].
On the other hand, a set of statistical tests has been conducted to study the statistical difference and significance of the proposed approach. This set is composed of the ANOVA, Wilcoxon, and statistical analysis. The results of these tests are presented in Table A16, Table A17, and Table A18, respectively. The results presented in these tables confirm the expected outcomes of the proposed approach. In addition, the plots shown in Figure A6 depict the visual representation of the ANOVA test results in terms of the Residual, Homoscedasticity, QQ, and Heatmap plots. These plots emphasize the effectiveness and statistical difference of the proposed approach when compared to the other approaches based on different optimization methods.
Figure A5. The classification accuracy of diabetes in dataset D4 after feature selection using the proposed approach.
Figure A5. The classification accuracy of diabetes in dataset D4 after feature selection using the proposed approach.
Diagnostics 13 02038 g0a5
Figure A6. Visualizing ANOVA test results when applied to the results of the optimized RF classifier.
Figure A6. Visualizing ANOVA test results when applied to the results of the optimized RF classifier.
Diagnostics 13 02038 g0a6
Table A16. ANOVA test applied to the results of the optimized RF classifier in terms of dataset D4.
Table A16. ANOVA test applied to the results of the optimized RF classifier in terms of dataset D4.
D4SSDFMSF (DFn, DFd)p-Value
Treatment (between columns)0.04660.007603F (6, 63) = 223.9p < 0.0001
Residual (within columns)0.002630.00003395
Total0.04869
Table A17. Wilcoxon test applied to the results of the optimized RF classifier in terms of dataset D4.
Table A17. Wilcoxon test applied to the results of the optimized RF classifier in terms of dataset D4.
D4DBERDTO + RFBER + RFDTO + RFPSO + RFWAO + RFGWO + RFFA + RF
Theoretical median0000000
Actual median0.9900.9590.9430.9380.9270.9130.906
Number of values10101010101010
Sum of signed ranks (W)55555555555555
Sum of positive ranks55555555555555
Sum of negative ranks0000000
P value (two tailed)0.0020.0020.0020.0020.0020.0020.002
Exact or estimate?ExactExactExactExactExactExactExact
Discrepancy0.9900.9590.9430.9380.9270.9130.906
Table A18. Statistical analysis of the classification results generated by the optimized RF classifier applied to dataset D4.
Table A18. Statistical analysis of the classification results generated by the optimized RF classifier applied to dataset D4.
D4DBERDTO + RFBER + RFDTO + RFPSO + RFWAO + RFGWO + RFFA + RF
Number of values10101010101010
Minimum0.9800.9430.9330.9310.9100.9030.898
25% Percentile0.9900.9480.9430.9370.9250.9130.906
Median0.9900.9590.9430.9380.9270.9130.906
75% Percentile0.9900.9590.9430.9380.9270.9130.908
Maximum0.9900.9620.9530.9500.9390.9340.916
Range0.0100.0190.0210.0190.0290.0300.019
10% Percentile0.9810.9430.9340.9310.9110.9040.899
90% Percentile0.9900.9620.9530.9490.9380.9320.916
Actual confidence level97.85%97.85%97.85%97.85%97.85%97.85%97.85%
Lower confidence limit0.9900.9430.9430.9350.9220.9100.906
Upper confidence limit0.9900.9590.9450.9380.9270.9130.914
Mean0.9890.9550.9430.9380.9260.9140.907
Std. Deviation0.0030.0070.0050.0050.0070.0080.005
Std. Error of Mean0.0010.0020.0020.0010.0020.0020.002
Lower 95% CI of mean0.9870.9500.9390.9350.9210.9090.904
Upper 95% CI of mean0.9910.9600.9470.9410.9310.9200.911
Coefficient of variation0.0030.0070.0050.0050.0080.0080.005
Geometric mean0.9890.9550.9430.9380.9260.9140.907
Lower 95% CI of geo. mean0.9870.9500.9390.9350.9210.9090.904
Upper 95% CI of geo. mean0.9910.9600.9470.9410.9310.9200.911
Harmonic mean0.9890.9550.9430.9380.9260.9140.907
Lower 95% CI of harm. mean0.9870.9500.9390.9350.9210.9090.904
Upper 95% CI of harm. mean0.9910.9600.9470.9410.9310.9190.911
Lower 95% CI of quad. mean0.9870.9500.9390.9350.9210.9090.904
Upper 95% CI of quad. mean0.9910.9600.9470.9410.9310.9200.911
Skewness−3.162−1.1190.010671.605−0.77781.9920.2301
Kurtosis10.000−0.5454.0225.2733.9726.3101.699
Sum9.8879.5509.4299.3809.2579.1419.070

Appendix A.4. The Selected Best Sets of Features

Based on the proposed feature selection method, bDBERDTO, the best sets of features extracted from the four adopted datasets are presented in Table A19. These best sets of features are used as inputs to the optimized random forest classifier to achieve the best classification results.
Table A19. The selected set of features from the four datasets employed in this work.
Table A19. The selected set of features from the four datasets employed in this work.
DatasetThe Selected Best Set of Features
D1 [41]Number of times pregnant
Diastolic blood pressure
Body mass index
Diabetes pedigree function
Age
Class variable
D2 [50]Pregnancies
Glucose
Blood Pressure
Diabetes Pedigree Function
Age
Outcome
D3 [51]AGE
Urea
Cr
HbA1c
HDL
LDL
VLDL
CLASS
D4 [52]Pregnancies
Glucose
Blood Pressure
BMI
Diabetes Pedigree Function
Age
Outcome

References

  1. American Diabetes Association. Diagnosis and Classification of Diabetes Mellitus. Diabetes Care 2006, 29, s43–s48. [Google Scholar] [CrossRef]
  2. Jamison, D.T.; World Bank; Disease Control Priorities Project (Eds.) Disease Control Priorities in Developing Countries, 2nd ed.; OCLC: Ocm64588990; Oxford University Press: Oxford, UK; World Bank: New York, NY, USA; Washington, DC, USA, 2006. [Google Scholar]
  3. World Health Organization. Noncommunicable Diseases Country Profiles 2018; World Health Organization: Geneva, Switzerland, 2018. [Google Scholar]
  4. Ellulu, M.S.; Samouda, H. Clinical and biological risk factors associated with inflammation in patients with type 2 diabetes mellitus. BMC Endocr. Disord. 2022, 22, 16. [Google Scholar] [CrossRef]
  5. IDF Diabetes Atlas. Available online: https://diabetesatlas.org/ (accessed on 20 May 2023).
  6. Guariguata, L.; Whiting, D.R.; Hambleton, I.; Beagley, J.; Linnenkamp, U.; Shaw, J.E. Global estimates of diabetes prevalence for 2013 and projections for 2035. Diabetes Res. Clin. Pract. 2014, 103, 137–149. [Google Scholar] [CrossRef]
  7. NCD Risk Factor Collaboration (NCD-RisC). Worldwide trends in diabetes since 1980: A pooled analysis of 751 population-based studies with 4.4 million participants. Lancet 2016, 387, 1513–1530. [Google Scholar] [CrossRef] [Green Version]
  8. Weinstein, M.C.; Toy, E.L.; Sandberg, E.A.; Neumann, P.J.; Evans, J.S.; Kuntz, K.M.; Graham, J.D.; Hammitt, J.K. Modeling for Health Care and Other Policy Decisions: Uses, Roles, and Validity. Value Health 2001, 4, 348–361. [Google Scholar] [CrossRef] [Green Version]
  9. Eid, M.M.; El-Kenawy, E.S.M.; Khodadadi, N.; Mirjalili, S.; Khodadadi, E.; Abotaleb, M.; Alharbi, A.H.; Abdelhamid, A.A.; Ibrahim, A.; Amer, G.M.; et al. Meta-Heuristic Optimization of LSTM-Based Deep Network for Boosting the Prediction of Monkeypox Cases. Mathematics 2022, 10, 3845. [Google Scholar] [CrossRef]
  10. Mahmood, A.; Irfan, A.; Wang, J. Developing Efficient Small Molecule Acceptors with sp2-Hybridized Nitrogen at Different Positions by Density Functional Theory Calculations, Molecular Dynamics Simulations and Machine Learning. Chem.—A Eur. J. 2022, 28, e202103712. [Google Scholar] [CrossRef]
  11. Mahmood, A.; Irfan, A.; Wang, J.L. Machine Learning for Organic Photovoltaic Polymers: A Minireview. Chin. J. Polym. Sci. 2022, 40, 870–876. [Google Scholar] [CrossRef]
  12. Zou, Q.; Qu, K.; Luo, Y.; Yin, D.; Ju, Y.; Tang, H. Predicting Diabetes Mellitus With Machine Learning Techniques. Front. Genet. 2018, 9, 515. [Google Scholar] [CrossRef]
  13. Lai, H.; Huang, H.; Keshavjee, K.; Guergachi, A.; Gao, X. Predictive models for diabetes mellitus using machine learning techniques. BMC Endocr. Disord. 2019, 19, 101. [Google Scholar] [CrossRef] [Green Version]
  14. Kopitar, L.; Kocbek, P.; Cilar, L.; Sheikh, A.; Stiglic, G. Early detection of type 2 diabetes mellitus using machine learning-based prediction models. Sci. Rep. 2020, 10, 11981. [Google Scholar] [CrossRef]
  15. Refat, M.A.R.; Amin, M.A.; Kaushal, C.; Yeasmin, M.N.; Islam, M.K. A Comparative Analysis of Early Stage Diabetes Prediction using Machine Learning and Deep Learning Approach. In Proceedings of the 2021 6th International Conference on Signal Processing, Computing and Control (ISPCC), Solan, India, 7–9 October 2021; pp. 654–659. [Google Scholar] [CrossRef]
  16. Forbus, J.J.; Berleant, D. Discrete-Event Simulation in Healthcare Settings: A Review. Modelling 2022, 3, 27. [Google Scholar] [CrossRef]
  17. Weinstein, M.C.; O’Brien, B.; Hornberger, J.; Jackson, J.; Johannesson, M.; McCabe, C.; Luce, B.R. Principles of Good Practice for Decision Analytic Modeling in Health-Care Evaluation: Report of the ISPOR Task Force on Good Research Practices—Modeling Studies. Value Health 2003, 6, 9–17. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  18. Faruque, M.F.; Asaduzzaman; Sarker, I.H. Performance Analysis of Machine Learning Techniques to Predict Diabetes Mellitus. In Proceedings of the 2019 International Conference on Electrical, Computer and Communication Engineering (ECCE), Cox’s Bazar, Bangladesh, 7–9 February 2019; pp. 1–4. [Google Scholar] [CrossRef] [Green Version]
  19. Patil, R.; Tamane, S. A Comparative Analysis on the Evaluation of Classification Algorithms in the Prediction of Diabetes. Int. J. Electr. Comput. Eng. 2018, 8, 3966. [Google Scholar] [CrossRef] [Green Version]
  20. Bukhari, M.M.; Alkhamees, B.F.; Hussain, S.; Gumaei, A.; Assiri, A.; Ullah, S.S. An Improved Artificial Neural Network Model for Effective Diabetes Prediction. Complexity 2021, 2021, 5525271. [Google Scholar] [CrossRef]
  21. Hasan, M.K.; Alam, M.A.; Das, D.; Hossain, E.; Hasan, M. Diabetes Prediction Using Ensembling of Different Machine Learning Classifiers. IEEE Access 2020, 8, 76516–76531. [Google Scholar] [CrossRef]
  22. Abdelhamid, A.A.; Alotaibi, S.R. Robust Prediction of the Bandwidth of Metamaterial Antenna Using Deep Learning. Comput. Mater. Contin. 2022, 72, 2305–2321. [Google Scholar] [CrossRef]
  23. Abdulhadi, N.; Al-Mousa, A. Diabetes Detection Using Machine Learning Classification Methods. In Proceedings of the 2021 International Conference on Information Technology (ICIT), Amman, Jordan, 14–15 July 2021; pp. 350–354. [Google Scholar] [CrossRef]
  24. Almutairi, E.S.; Abbod, M.F. Machine Learning Methods for Diabetes Prevalence Classification in Saudi Arabia. Modelling 2023, 4, 4. [Google Scholar] [CrossRef]
  25. Khafaga, D.S.; Alhussan1, A.A.; El-kenawy, E.-S.M.; Ibrahim, A.; Elkhalik, S.H.A.; El-Mashad, S.Y.; Abdelhamid, A.A. Meta-heuristics for Feature Selection and Classification in Diagnostic Breast Cancer. Comput. Mater. Contin. 2022, 73, 749–765. [Google Scholar] [CrossRef]
  26. Khafaga, D.S.; Alhussan, A.A.; El-Kenawy, E.S.M.; Ibrahim, A.; Eid, M.M.; Abdelhamid, A.A. Solving Optimization Problems of Metamaterial and Double T-Shape Antennas Using Advanced Meta-Heuristics Algorithms. IEEE Access 2022, 10, 74449–74471. [Google Scholar] [CrossRef]
  27. Kantawong, K.; Tongphet, S.; Bhrommalee, P.; Rachata, N.; Pravesjit, S. The Methodology for Diabetes Complications Prediction Model. In Proceedings of the 2020 Joint International Conference on Digital Arts, Media and Technology with ECTI Northern Section Conference on Electrical, Electronics, Computer and Telecommunications Engineering (ECTI DAMT & NCON), Pattaya, Thailand, 11–14 March 2020; pp. 110–113. [Google Scholar] [CrossRef]
  28. Dagliati, A.; Marini, S.; Sacchi, L.; Cogni, G.; Teliti, M.; Tibollo, V.; De Cata, P.; Chiovato, L.; Bellazzi, R. Machine Learning Methods to Predict Diabetes Complications. J. Diabetes Sci. Technol. 2018, 12, 295–302. [Google Scholar] [CrossRef] [Green Version]
  29. Khafaga, D.S.; Alhussan1, A.A.; El-kenawy, E.-S.M.; Ibrahim, A.; Elkhalik, S.H.A.; El-Mashad, S.Y.; Abdelhamid, A.A. Improved Prediction of Metamaterial Antenna Bandwidth Using Adaptive Optimization of LSTM. Comput. Mater. Contin. 2022, 73, 865–881. [Google Scholar] [CrossRef]
  30. AlEisa, H.N.; El-kenawy, E.-S.M.; Alhussan1, A.A.; Saber, M.; Abdelhamid, A.A.; Khafaga, D.S. Transfer Learning for Chest X-rays Diagnosis Using Dipper Throated Algorithm. Comput. Mater. Contin. 2022, 73, 2371–2387. [Google Scholar] [CrossRef]
  31. Shams, M.Y. Hybrid Neural Networks in Generic Biometric System: A Survey. J. Artif. Intell. Metaheuristics 2022, 1, 20–26. [Google Scholar] [CrossRef]
  32. Saber, M. Removing Powerline Interference from EEG Signal using Optimized FIR Filters. J. Artif. Intell. Metaheuristics 2022, 1, 8–19. [Google Scholar] [CrossRef]
  33. Saber, M.; Abotaleb, M. Arrhythmia Modern Classification Techniques: A Review. J. Artif. Intell. Metaheuristics 2022, 1, 42–53. [Google Scholar] [CrossRef]
  34. Oubelaid, A.; Shams, M.Y.; Abotaleb, M. Energy Efficiency Modeling Using Whale Optimization Algorithm and Ensemble Model. J. Artif. Intell. Metaheuristics 2023, 2, 27–35. [Google Scholar] [CrossRef]
  35. Shazly, K.; Khodadadi, N. Credit Card Clients Classification Using Hybrid Guided wheel with Particle Swarm Optimized for Voting Ensemble. J. Artif. Intell. Metaheuristics 2023, 2, 46–54. [Google Scholar] [CrossRef]
  36. El-kenawy, E.S.M.; Abdelhamid, A.A.; Ibrahim, A.; Mirjalili, S.; Khodadad, N.; Duailij, M.A.A.; Alhussan, A.A.; Khafaga, D.S. Al-Biruni Earth Radius (BER) Metaheuristic Search Optimization Algorithm. Comput. Syst. Sci. Eng. 2023, 45, 1917–1934. [Google Scholar] [CrossRef]
  37. Mirjalili, S.; Lewis, A. The Whale Optimization Algorithm. Adv. Eng. Softw. 2016, 95, 51–67. [Google Scholar] [CrossRef]
  38. Immanuel, S.D.; Chakraborty, U.K. Genetic Algorithm: An Approach on Optimization. In Proceedings of the 2019 International Conference on Communication and Electronics Systems (ICCES), Coimbatore, India, 17–19 July 2019; pp. 701–708. [Google Scholar] [CrossRef]
  39. Awange, J.L.; Paláncz, B.; Lewis, R.H.; Völgyesi, L. Particle Swarm Optimization. In Mathematical Geosciences: Hybrid Symbolic-Numeric Methods; Awange, J.L., Paláncz, B., Lewis, R.H., Völgyesi, L., Eds.; Springer International Publishing: Cham, Switzerland, 2018; pp. 167–184. [Google Scholar] [CrossRef]
  40. Mirjalili, S.; Mirjalili, S.M.; Lewis, A. Grey Wolf Optimizer. Adv. Eng. Softw. 2014, 69, 46–61. [Google Scholar] [CrossRef] [Green Version]
  41. Akturk, M. Diabetes Dataset. 2020. Available online: https://www.kaggle.com/datasets/mathchi/diabetes-data-set (accessed on 20 April 2023).
  42. Takieldeen, A.E.; El-kenawy, E.-S.M.; Hadwan, M.; Zaki, R.M. Dipper Throated Optimization Algorithm for Unconstrained Function and Feature Selection. Comput. Mater. Contin. 2022, 72, 1465–1481. [Google Scholar] [CrossRef]
  43. Netrapalli, P. Stochastic Gradient Descent and Its Variants in Machine Learning. J. Indian Inst. Sci. 2019, 99, 201–213. [Google Scholar] [CrossRef]
  44. Charbuty, B.; Abdulazeez, A. Classification Based on Decision Tree Algorithm for Machine Learning. J. Appl. Sci. Technol. Trends 2021, 2, 20–28. [Google Scholar] [CrossRef]
  45. Dang, Y.; Jiang, N.; Hu, H.; Ji, Z.; Zhang, W. Image classification based on quantum K-Nearest-Neighbor algorithm. Quantum Inf. Process. 2018, 17, 239. [Google Scholar] [CrossRef]
  46. Berrar, D. Bayes’ Theorem and Naive Bayes Classifier. In Encyclopedia of Bioinformatics and Computational Biology; Ranganathan, S., Gribskov, M., Nakai, K., Schönbach, C., Eds.; Academic Press: Oxford, UK, 2019; pp. 403–412. [Google Scholar] [CrossRef]
  47. Anthony, G.; Greg, H.; Tshilidzi, M. Classification of Images Using Support Vector Machines. arXiv 2007, arXiv:0709.3967. [Google Scholar] [CrossRef]
  48. Zhou, M.; Samiappan, S.; Worch, E.; Ball, J.E. Hyperspectral Image Classification Using Fisher’s Linear Discriminant Analysis Feature Reduction with Gabor Filtering and CNN. In Proceedings of the IGARSS 2020—2020 IEEE International Geoscience and Remote Sensing Symposium, Waikoloa, HI, USA, 26 September–2 October 2020; pp. 493–496. [Google Scholar] [CrossRef]
  49. Mohapatra, N.; Shreya, K.; Chinmay, A. Optimization of the Random Forest Algorithm. In Advances in Data Science and Management, Proceedings of the ICDSM 2019, Changsha, China, 20–21 September 2019; Borah, S., Emilia Balas, V., Polkowski, Z., Eds.; Springer: Singapore, 2020; pp. 201–208. [Google Scholar]
  50. Dattatray, A. Diabetes Dataset. 2022. Available online: https://www.kaggle.com/datasets/akshaydattatraykhare/diabetes-dataset (accessed on 20 May 2023).
  51. Rashid, A. Diabetes Dataset. 2020. Available online: https://data.mendeley.com/datasets/wj9rwkp9c2/1 (accessed on 20 May 2023).
  52. Learning, U.M. Pima Indians Diabetes Database. 2016. Available online: https://www.kaggle.com/datasets/uciml/pima-indians-diabetes-database (accessed on 1 June 2023).
  53. Naz, H.; Ahuja, S. SMOTE-SMO-based expert system for type II diabetes detection using PIMA dataset. Int. J. Diabetes Dev. Ctries. 2022, 42, 245–253. [Google Scholar] [CrossRef]
  54. Chang, V.; Bailey, J.; Xu, Q.A.; Sun, Z. Pima Indians diabetes mellitus classification based on machine learning (ML) algorithms. Neural Comput. Appl. 2022. [Google Scholar] [CrossRef] [PubMed]
  55. Naz, H.; Nijhawan, R.; Ahuja, N.J. DT-DL Based Hybrid Approach for Early Detection of Diabetes Using PIMA Dataset. In Proceedings of the 2022 10th International Conference on Reliability, Infocom Technologies and Optimization (Trends and Future Directions) (ICRITO), Noida, India, 13–14 October 2022; pp. 1–6. [Google Scholar] [CrossRef]
  56. Gollapalli, M.; Alansari, A.; Alkhorasani, H.; Alsubaii, M.; Sakloua, R.; Alzahrani, R.; Al-Hariri, M.; Alfares, M.; AlKhafaji, D.; Al Argan, R.; et al. A novel stacking ensemble for detecting three types of diabetes mellitus using a Saudi Arabian dataset: Pre-diabetes, T1DM, and T2DM. Comput. Biol. Med. 2022, 147, 105757. [Google Scholar] [CrossRef] [PubMed]
Figure 1. The architecture of the proposed system.
Figure 1. The architecture of the proposed system.
Diagnostics 13 02038 g001
Figure 2. The correlation among the features of the diabetes dataset.
Figure 2. The correlation among the features of the diabetes dataset.
Diagnostics 13 02038 g002
Figure 3. The histograms of the features of the diabetes dataset.
Figure 3. The histograms of the features of the diabetes dataset.
Diagnostics 13 02038 g003
Figure 4. The average error of the proposed feature selection method results.
Figure 4. The average error of the proposed feature selection method results.
Diagnostics 13 02038 g004
Figure 5. Visualizing the ANOVA test applied to the proposed feature selection method results.
Figure 5. Visualizing the ANOVA test applied to the proposed feature selection method results.
Diagnostics 13 02038 g005
Figure 6. The accuracy achieved by the optimized RF classifier compared to other optimization methods.
Figure 6. The accuracy achieved by the optimized RF classifier compared to other optimization methods.
Diagnostics 13 02038 g006
Figure 7. The accuracy histogram achieved by the optimized RF classifier compared to other optimization methods.
Figure 7. The accuracy histogram achieved by the optimized RF classifier compared to other optimization methods.
Diagnostics 13 02038 g007
Figure 8. Visualizing ANOVA test results when applied to the results of the optimized RF classifier.
Figure 8. Visualizing ANOVA test results when applied to the results of the optimized RF classifier.
Diagnostics 13 02038 g008
Table 1. Review of the recent methods used in diabetes classification.
Table 1. Review of the recent methods used in diabetes classification.
PaperMethodFeaturesResultsLimitations
 [18]KNN, naive Bayes (NB), C4.5 decision tree, SVMUses relevant kernel functions to resolve various issues. When dealing with unstructured data, SVM excels. In addition, it can produce useful outcomes even when only partial data is available.The C4.5 decision tree model has the highest accuracy (73%) compared to the other classifiers studied.Using large datasets in the training process is time-consuming. Since their relative importance is not fixed, each input variable contributes in a unique way to the final result.
[20]Artificial neural network (ANN)The ANN is intuitive and quickly adapts to new situations. It is helpful for addressing issues where noisy input data is present.There was a 93% success rate for the ANN model.Knowing the required neurons and layers is difficult, and learning can be slow.
[21]Random forest, naive Bayes, XG Boost, decision trees, KNN and AdaBoostThe use of KNN is simple. It is optimal for datasets with several classes. Training is efficient, producing high-quality outcomes quickly.AdaBoost and XG Boost were the most effective models in an ensemble setting. Compared to other research, this one had the highest area under the curve (AUC) score, reaching 0.95.Memory limitation and a long time to find the nearest neighbors in large datasets.
[21]Random forest, SVM, logistic regression (LR), and linear discriminant analysis (LDA)Can be applied to two or multiple classes. The implementation is simple, and the classification is fast.The RF model performed best, with an accuracy score of 82%.Complex nonlinear data is beyond its capabilities. LDA distributions deviate significantly from Gaussianity. If the variance of the data is used as a discriminator instead of its mean, LDA will not work.
Table 2. Evaluation metrics used in assessing the proposed feature selection method.
Table 2. Evaluation metrics used in assessing the proposed feature selection method.
MetricValue
Mean 1 M i = 1 M S i *
Best Fitness m i n i = 1 M S i *
Average Error 1 M j = 1 M 1 N i = 1 N m s e ( V i ^ V i )
Worst Fitness m a x i = 1 M S i *
Average fitness size 1 M i = 1 M s i z e ( S i * )
Standard deviation 1 M 1 i = 1 M S i * M e a n 2
Table 3. Evaluation metrics used in assessing the optimized random forest classifier.
Table 3. Evaluation metrics used in assessing the optimized random forest classifier.
MetricValue
Accuracy T P + T N T P + T N + F P + F N
Sensitivity T P T P + F N
Specificity T N T N + F P
F1-Score T P T P + 0.5 ( F P + F N )
N-value T N T N + F N
P-value T P T P + F P
Table 4. Configuration parameters of the employed optimization algorithms.
Table 4. Configuration parameters of the employed optimization algorithms.
AlgorithmParameterValuesAlgorithmParameterValues
BER [36]Iterations80GWO [40]a2 to 0
Mutation probability0.5 Iterations80
Exploration percentage70 Wolves10
k (decreases from 2 to 0)1WOA [37]r[0, 1]
DTO [42]Iterations80 Iterations80
Number of runs30 Whales10
Exploration percentage70 a2 to 0
GA [38]Cross over0.9PSO [39]Acceleration constants[2, 2]
Mutation ratio0.1 Inertia W m a x , W m i n [0.6, 0.9]
MechanismRoulette wheel Particles10
Agents10 Iterations80
Iterations80
Table 5. Configuration parameters of the baseline classification models.
Table 5. Configuration parameters of the baseline classification models.
ModelParameterValue
SGD [43]losshinge
penaltyl2
DT [44]splitter‘best’
min_samples_split2
criterion‘gini’ to 0
min_samples_leaf1
KNN [45]n_neighbors5
weights‘uniform’
leaf_size30
p2
GNB [46]Likelihood of featuresGaussian
SVM [47]C1
kernel‘rbf’
penalty‘l2’
tol1.0 × 10 4
LR [48]solver‘svd’
tol1.0 × 10 4
shrinkage[0–1]
RF [49]max_depth2
random_state0
Table 6. Evaluation of the results of the proposed feature selection method.
Table 6. Evaluation of the results of the proposed feature selection method.
D1bDBERDTObBERbDTObPSObWAObGWObFA
Average error0.4600.4770.4910.5110.5110.4980.510
Average Select size0.4130.6130.5550.6130.7760.5360.648
Average Fitness0.5230.5400.5510.5380.5460.5460.590
Best Fitness0.4250.4600.4540.5180.5100.5240.509
Worst Fitness0.5240.5270.5690.5860.5860.6000.606
Std Fitness0.3460.3500.3520.3500.3520.3510.387
Table 7. Analysis of variance (ANOVA) of the feature selection results. In this table, SS denotes (sum of squares), DF (degrees of freedom), DFn denotes DF numerator and DFd denotes DF denominator. MS (mean square).
Table 7. Analysis of variance (ANOVA) of the feature selection results. In this table, SS denotes (sum of squares), DF (degrees of freedom), DFn denotes DF numerator and DFd denotes DF denominator. MS (mean square).
D1SSDFMSF (DFn, DFd)p-Value
Treatment0.0232360.003871F (6, 63) = 136.1p < 0.0001
Residual0.00179630.00002845
Total0.0250269
Table 8. Wilcoxon signed rank test of the proposed feature selection method.
Table 8. Wilcoxon signed rank test of the proposed feature selection method.
D1bDBERDTObBERbDTObPSObWAObGWObFA
Theoretical median0000000
Actual median0.46030.47750.49110.51130.51110.49760.5097
Number of values10101010101010
Sum of signed ranks (W)55555555555555
Sum of positive ranks55555555555555
Sum of negative ranks0000000
P value (two tailed)0.0020.0020.0020.0020.0020.0020.002
Exact or estimate?ExactExactExactExactExactExactExact
Significant (alpha = 0.05)?YesYesYesYesYesYesYes
Discrepancy0.46030.47750.49110.51130.51110.49760.5097
Table 9. Evaluation of the classifiers results before applying feature selection.
Table 9. Evaluation of the classifiers results before applying feature selection.
D1p-Valuep-ValueF-Score AccuracySensitivitySpecificity
SGD0.7290.5000.4410.5330.3800.776
DT0.7560.7300.5780.6800.7070.638
KNN0.7250.7630.6250.6930.8040.517
GNB0.8130.8090.6950.7670.8040.707
SVM0.7610.8390.8380.7800.9350.534
LR0.8000.8330.7600.7870.8700.655
RF0.8400.8490.7680.8130.8590.741
Table 10. Evaluation of the results of the classifiers after applying the proposed feature selection method and before optimizing the parameters of the classifiers.
Table 10. Evaluation of the results of the classifiers after applying the proposed feature selection method and before optimizing the parameters of the classifiers.
D1p-Valuep-ValueF-ScoreAccuracySensitivitySpecificity
SGD0.5880.8420.7140.6570.9090.432
DT0.6820.8420.7790.7300.9090.533
KNN0.7690.8570.8470.7910.9430.545
GNB0.8060.8570.8700.8190.9430.600
SVM0.8570.8420.8820.8520.9090.762
LR0.8510.8570.8890.8530.9300.720
RF0.9090.8420.9090.8850.9090.842
Table 11. Statistical analysis of the results of the classification results.
Table 11. Statistical analysis of the results of the classification results.
D1DBERDTO + RFBER + RFDTO + RFPSO + RFWAO + RFGWO + RFFA + RF
Number of values10101010101010
Minimum0.9860.9510.9500.9330.9280.9270.923
25% Percentile0.9860.9610.9590.9430.9380.9370.927
Median0.9860.9610.9590.9430.9380.9370.927
75% Percentile0.9860.9610.9590.9430.9410.9370.927
Maximum0.9920.9610.9600.9530.9580.9400.936
Range0.0060.0100.0100.0200.0300.0120.013
10% Percentile0.9860.9520.9510.9340.9290.9280.923
90% Percentile0.9910.9610.9600.9520.9570.9400.935
Actual confidence level97.85%97.85%97.85%97.85%97.85%97.85%97.85%
Lower confidence limit0.9860.9590.9590.9430.9380.9370.927
Upper confidence limit0.9860.9610.9590.9430.9480.9370.927
Mean0.9860.9600.9580.9430.9400.9370.927
Std. Deviation0.0020.0030.0030.0050.0080.0030.003
Std. Error of Mean0.0010.0010.0010.0010.0020.0010.001
Lower 95% CI of mean0.9850.9580.9560.9400.9350.9340.925
Upper 95% CI of mean0.9880.9620.9600.9470.9460.9390.929
Coefficient of variation0.1911%0.3289%0.3028%0.4997%0.8391%0.3562%0.3518%
Geometric mean0.9860.9600.9580.9430.9400.9370.927
Geometric SD factor1.0021.0031.0031.0051.0081.0041.004
Lower 95% CI of geo. mean0.9850.9580.9560.9400.9350.9340.925
Upper 95% CI of geo. mean0.9880.9620.9600.9470.9460.9390.929
Harmonic mean0.9860.9600.9580.9430.9400.9370.927
Lower 95% CI of harm. mean0.9850.9580.9560.9400.9350.9340.925
Upper 95% CI of harm. mean0.9880.9620.9600.9470.9460.9390.929
Quadratic mean0.9860.9600.9580.9430.9400.9370.927
Lower 95% CI of quad. mean0.9850.9580.9560.9400.9340.9340.925
Upper 95% CI of quad. mean0.9880.9620.9600.9470.9460.9390.929
Skewness3.162−2.939−3.0900.0001.290−2.8192.223
Kurtosis10.0008.8189.6954.5002.9858.7057.021
Sum9.8629.6009.5819.4339.4019.3659.270
Table 12. ANOVA test applied to the results of the optimized RF classifier.
Table 12. ANOVA test applied to the results of the optimized RF classifier.
D1SSDFMSF (DFn, DFd)p-Value
Treatment0.0232760.003878F (6, 63) = 211.8p < 0.0001
Residual0.00115630.000018
Total0.0244269
Table 13. Wilcoxon test applied to the results of the optimized RF classifier.
Table 13. Wilcoxon test applied to the results of the optimized RF classifier.
D1DBERDTO + RFBER + RFDTO + RFPSO + RFWAO + RFGWO + RFFA + RF
Number of values10101010101010
Actual median0.98560.96120.95890.94330.93810.93730.9265
Theoretical median0000000
Sum of negative ranks0000000
Sum of signed ranks (W)55555555555555
Sum of positive ranks55555555555555
P value (two tailed)0.0020.0020.0020.0020.0020.0020.002
Exact or estimate?ExactExactExactExactExactExactExact
Discrepancy0.98560.96120.95890.94330.93810.93730.9265
Significant (alpha = 0.05)?YesYesYesYesYesYesYes
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Alhussan, A.A.; Abdelhamid, A.A.; Towfek, S.K.; Ibrahim, A.; Eid, M.M.; Khafaga, D.S.; Saraya, M.S. Classification of Diabetes Using Feature Selection and Hybrid Al-Biruni Earth Radius and Dipper Throated Optimization. Diagnostics 2023, 13, 2038. https://doi.org/10.3390/diagnostics13122038

AMA Style

Alhussan AA, Abdelhamid AA, Towfek SK, Ibrahim A, Eid MM, Khafaga DS, Saraya MS. Classification of Diabetes Using Feature Selection and Hybrid Al-Biruni Earth Radius and Dipper Throated Optimization. Diagnostics. 2023; 13(12):2038. https://doi.org/10.3390/diagnostics13122038

Chicago/Turabian Style

Alhussan, Amel Ali, Abdelaziz A. Abdelhamid, S. K. Towfek, Abdelhameed Ibrahim, Marwa M. Eid, Doaa Sami Khafaga, and Mohamed S. Saraya. 2023. "Classification of Diabetes Using Feature Selection and Hybrid Al-Biruni Earth Radius and Dipper Throated Optimization" Diagnostics 13, no. 12: 2038. https://doi.org/10.3390/diagnostics13122038

APA Style

Alhussan, A. A., Abdelhamid, A. A., Towfek, S. K., Ibrahim, A., Eid, M. M., Khafaga, D. S., & Saraya, M. S. (2023). Classification of Diabetes Using Feature Selection and Hybrid Al-Biruni Earth Radius and Dipper Throated Optimization. Diagnostics, 13(12), 2038. https://doi.org/10.3390/diagnostics13122038

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop