Next Article in Journal
Extraction, Isolation, Identification, and Characterization of Anthocyanin from Banana Inflorescence by Liquid Chromatography-Mass Spectroscopy and Its pH Sensitivity
Previous Article in Journal
Multi-Level Thresholding Color Image Segmentation Using Modified Gray Wolf Optimizer
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Adaptive Sand Cat Swarm Optimization Algorithm for Feature Selection and Global Optimization

1
College of Information Science and Technology, Shihezi University, Shihezi 832000, China
2
College of Medicine, Shihezi University, Shihezi 832000, China
3
College of Science, Shihezi University, Shihezi 832000, China
*
Author to whom correspondence should be addressed.
Biomimetics 2024, 9(11), 701; https://doi.org/10.3390/biomimetics9110701
Submission received: 9 October 2024 / Revised: 12 November 2024 / Accepted: 12 November 2024 / Published: 15 November 2024

Abstract

:
Feature selection (FS) constitutes a critical stage within the realms of machine learning and data mining, with the objective of eliminating irrelevant features while guaranteeing model accuracy. Nevertheless, in datasets featuring a multitude of features, choosing the optimal feature poses a significant challenge. This study presents an enhanced Sand Cat Swarm Optimization algorithm (MSCSO) to improve the feature selection process, augmenting the algorithm’s global search capacity and convergence rate via multiple innovative strategies. Specifically, this study devised logistic chaotic mapping and lens imaging reverse learning approaches for population initialization to enhance population diversity; balanced global exploration and local development capabilities through nonlinear parameter processing; and introduced a Weibull flight strategy and triangular parade strategy to optimize individual position updates. Additionally, the Gaussian–Cauchy mutation strategy was employed to improve the algorithm’s ability to overcome local optima. The experimental results demonstrate that MSCSO performs well on 65.2% of the test functions in the CEC2005 benchmark test; on the 15 datasets of UCI, MSCSO achieved the best average fitness in 93.3% of the datasets and achieved the fewest feature selections in 86.7% of the datasets while attaining the best average accuracy across 100% of the datasets, significantly outperforming other comparative algorithms.

1. Introduction

During the big data era, with the sharp increase in feature dimensions in intelligent information processing tasks, a large number of unnecessary or unrelated features appear in the dataset, which not only results in reduced model accuracy but also adds to the computational time burden [1]. Feature selection (FS), as a key data dimensionality reduction method, aims to select the most informative feature subset in a given feature set to improve the performance of machine learning models [2]. FS helps to cut down on computational costs while also increasing the model’s interpretability. Especially when dealing with large-scale datasets, it can significantly improve computational efficiency and reduce the complexity of model training.
The traditional feature selection methods can be divided into the filtering method, wrapping method, and embedding method [3]. Filtering methods are selected by independent evaluation of each feature, with common evaluation metrics including information gain (IG), the chi-square test, and mutual information (MI). These methods are computationally efficient but often ignore the correlation between features. By combining feature subsets with specific learning algorithms to evaluate their performance, the package method selects the optimal feature subsets. Although the wrapping method can consider the relationship between features, it is computationally expensive, especially when the feature dimension is high. The embedding method integrates the feature selection process into the model training process and directly selects features through regularization technology or model structures, such as ridge regression (RR) and decision trees (DTs), etc. This method has a good performance in computational efficiency and accuracy.
When it comes to handling large-scale datasets or high-dimensional data, the drawbacks of these traditional methods become particularly prominent. Hence, researchers have embarked on seeking more flexible algorithms capable of handling complex feature interactions. Metaheuristic algorithms, as an emergent approach, have emerged as an effective tool for resolving feature selection issues with their potent global search capabilities and adaptability to complex matters [4].
Metaheuristic algorithms draw inspiration from natural processes and human behavioral patterns. Based on the heuristic approach mechanism, these algorithms fall into five categories, namely evolutionary-based algorithms, population-based algorithms, physics/chemistry-based algorithms, human behavior-based algorithms, and mathematics-based algorithms. Evolutionary-based algorithms mimic natural selection and genetic processes, with typical instances including Genetic Algorithms (GAs) [5], Differential Evolution (DE) [6], and Evolutionary Strategies (ESs) [7]. Population-based algorithms simulate the behavior of biological populations, such as the Krill Herd Algorithm (KHA) [8], the Honeybee Algorithm (HBA) [9], and the Fox Optimization Algorithm (FOX) [10]. Physics/chemistry-based algorithms draw from physical and chemical phenomena in nature, like Simulated Annealing (SA) [11], the Crystal Structure Algorithm (CryStAl) [12], and the Gravitational Search Algorithm (GSA) [13]. Human behavior-based algorithms borrow characteristics of human thinking, behavior, or social structures, such as the Love Evolutionary Algorithm (LEA) [14], the Rider Optimization Algorithm (ROA) [15], and the Stadium Audience Optimization Algorithm (SSO) [16]. Finally, mathematics-based algorithms are designed based on mathematical theories, such as Hyperbolic Sine Cosine Optimization (SCHO) [17], Triangular Topology Aggregation Optimization (TTAO) [18], and Exponential Distribution Optimization (EDO) [19].
According to the “No Free Lunch Theorem” for optimization [20], no single algorithm performs best in every situation, which explains the continuous emergence of diverse optimization algorithms.
The Sand Cat Swarm Optimization (SCSO) algorithm was proposed in [21], which mimics how sand cats survive in the natural world, characterized by its simple principle, few parameters, and strong global search capability. This algorithm has shown excellent performance on 20 famous test functions and 10 CEC2019 test functions and has performed well in seven engineering design problems. However, the original algorithm still encounters issues such as limited optimization accuracy, slow convergence, and a tendency to become trapped in local optima when addressing high-dimensional highly intricate optimization tasks. To overcome these issues, several improvements have been made to SCSO, and the enhanced version is named MSCSO.
This study makes the following key contributions:
  • An innovative multi-objective feature selection method, the improved Sand Cat Swarm Optimization algorithm, is proposed for the feature selection problem. This method enhances feature selection efficiency while considering both the classification error rate and feature selection ratio as objectives, aiming to obtain the optimal feature subset, reduce redundant features, and improve classification accuracy.
  • The population initialization method, nonlinear parameter handling, and position update strategy have been innovatively improved, and the Gaussian–Cauchy mutation strategy has been introduced. These enhancements significantly boost the algorithm’s global search capability, convergence speed, and optimization accuracy, while effectively avoiding the issue of local optima.
  • The global optimization performance of MSCSO was evaluated through the CEC2005 benchmark test, with results showing that it performed excellently on 65.2% of the test functions. Additionally, in feature selection experiments conducted on 15 datasets from UCI, MSCSO achieved the optimal average fitness in 93.3% of the datasets, demonstrated the optimal number of feature selections in 86.7% of the datasets, and reached the highest average accuracy in 100% of the datasets, all superior to other comparative algorithms.
The structure of the remainder of this study is as follows: Section 2 examines the relevant research progress in feature selection, Section 3 describes the improvement methods of SCSO and its complexity analysis, Section 4 presents the analysis and discussion of experimental results, and Section 5 reviews the key findings of this research and suggests potential avenues for future investigation.

2. Related Work

2.1. Feature Selection Based on Traditional Methods

Feature selection (FS) is regarded as an NP-hard problem because of the complexity of its search space. With increasing data dimensionality, the number of potential feature subsets increases exponentially, making it extremely challenging to identify the optimal subset. To tackle this issue, researchers have developed numerous approaches to feature selection, which are typically divided into three main types, namely filtering methods, wrapper methods, and embedded methods.
Filter, wrapper, and embedded methods are three widely used feature selection techniques, each with distinct characteristics. Filter methods evaluate the relationship between features and the target variable to screen feature subsets, operating independently of model training. Univariate methods are computationally efficient but may overlook interactions among features, while multivariate approaches can capture complex inter-feature relationships. For example, Ref. [22] utilized variance–covariance distance to eliminate features with minimal variance for dimensionality reduction, while Ref. [23] employed MI and the variance inflation factor (VIF) to eliminate multicollinearity, thereby refining the feature set. Wrapper methods, on the other hand, evaluate feature subsets based directly on model performance, using metrics such as forward selection, backward elimination, and recursive feature elimination (RFE). Ref. [24] applied forward feature selection to iteratively construct transitional index maps, and Ref. [25] combined recursive feature elimination with cross-validation (RFECV) to optimize feature selection for decision tree models. Embedded methods perform feature selection during model training, allowing for synchronous feature selection and model fitting. For instance, Ref. [26] leveraged LASSO to select features highly correlated with liver disease, enhancing classification accuracy. Additionally, Ref. [27] employed multivariate variational mode decomposition (MVMD) and kernel extreme learning machine (KELM) models to predict plant and soil moisture evaporation, effectively addressing complex feature selection tasks.

2.2. Feature Selection Based on Metaheuristic Algorithm

In traditional feature selection methods, the filtering method, wrapper method, and embedded method each have their strengths and weaknesses. Although these methods perform well in addressing feature selection issues, they may still encounter limitations when dealing with high-dimensional data, complex interactions between features, and limited computational resources. To overcome these limitations, researchers are gradually turning to metaheuristic algorithms. These algorithms simulate heuristic processes in the natural world and can effectively search the feature space to find feature subsets suitable for specific problems.
The introduction of meta-heuristic algorithms provides a new way to solve the difficult problem of feature selection, especially in the case of high feature dimensionality or complex relationships between features. Next, the application of several typical meta-heuristic algorithms in feature selection will be discussed. Ref. [28] proposed the binary arithmetic optimization algorithm (BAOA) to address the feature selection problem. The algorithm transforms continuous variables into discrete variables through S-shaped and V-shaped shift functions and combines four logical operations and parametric models based on sine and cosine functions to optimize the feature selection process. Ref. [29] adopted an improved slime mold algorithm (ISMA) for feature selection. This algorithm significantly improves classification performance by simultaneously optimizing parameters and feature selection of a support vector machine (SVM). Ref. [30] used the Binary Macaque Optimization Algorithm (BMOA) for feature selection, which improves classification accuracy by screening for the most relevant and effective features. The improved version of the BMOA (IBMOA) further enhances exploration and development capabilities. Ref. [31] designed an enhanced raccoon optimization algorithm (mCoatiOA) to enhance its exploration and development capabilities by introducing adaptive s-best mutation operators, directional mutation rules, and global optimal search control. Ref. [32] improved the Gray Wolf Optimization algorithm (GWO) for feature selection, improved the initial population quality by introducing the ReliefF algorithm and Coupla entropy, and optimized the search process using competition guidance and DE’s lead wolf enhancement strategy. This method effectively reduces the feature set and enhances classification accuracy. Ref. [33] presented an advanced variant of the Harris Eagle Optimization algorithm (EHHO). By introducing a hierarchical structure, the algorithm shows a stronger capacity for handling complex problems. In feature selection, EHHO enhances classification performance, decreases feature count, and cuts down on execution time. The above algorithms perform well in feature selection, but there are some limitations in practical applications. First, they often require selecting a large number of features, which results in low computational efficiency, especially when dealing with large-scale datasets, leading to slower processing speeds. Additionally, due to their high computational cost and sensitivity to parameters, these algorithms have limited practicality and are difficult to apply in real-time systems or complex environments.
Some scholars use hybrid algorithms for feature selection, which combine two or more different algorithms to take full advantage of each. Ref. [34] proposed the mixed sine–cosine Firehawk algorithm (HSCFHA). The model combines the Fire Eagle algorithm (FHO) and the sine and cosine algorithm (SCA) to obtain better results in feature selection problems by exploiting the advantages of both algorithms in exploration and development. Ref. [35] used a particle swarm-guided Condor search algorithm (PS-BES) for feature selection. The algorithm guides the condor’s search by combining the speed of the particle swarm and utilizes the attack–retreat–surrender technique to increase the capacity for escaping local optima, thus improving the overall effect of the feature selection process. Ref. [36] created a new integrated binary metaheuristic that combines the Deep Throat Optimization algorithm (DTO) and the SCA. By using the SCA to improve the exploration process and combining DTO to accelerate convergence, the algorithm improves the efficiency and accuracy of feature selection. Ref. [37] combined particle swarm optimization (PSO) and the firefly algorithm (FA), combining the advantages of each algorithm for problems related to feature selection. Ref. [38] proposed a hybrid optimization approach (GWDTO) based on GWO and DTO for feature selection. This method improves the effectiveness of feature selection by effectively balancing the exploration and development steps in the optimization process. These hybrid algorithms perform well in feature selection, but there are some issues. Firstly, they have high computational complexity, leading to low efficiency when processing large datasets. Secondly, they are prone to getting trapped in local optima, making it difficult to find the global optimum. Finally, the algorithms are sensitive to parameters and tuning them is challenging, which makes them less flexible and practical in real-world applications.
Compared to other complex meta-heuristic algorithms, SCSO has the advantage of fewer parameters, making it easier to adjust. This results in a simpler optimization process and reduced complexity in parameter tuning. Additionally, SCSO maintains a good balance between exploration and exploitation, effectively avoiding local optima. By simplifying the adjustment process, SCSO improves feature selection efficiency and adapts well to complex datasets. Therefore, SCSO was chosen as the base algorithm for improvement in this study.

2.3. Motivation

Feature selection is a critical step in machine learning and data mining, aimed at removing irrelevant features to ensure model accuracy, thereby improving interpretability and generalization. However, existing feature selection methods have notable limitations, necessitating more efficient and accurate solutions.
(1)
Traditional feature selection methods struggle with feature correlation and computational efficiency. Filter methods, though simple, overlook correlations between features; wrapper methods consider feature relationships but are computationally intensive, making them unsuitable for high-dimensional data; and embedded methods are often affected by model dependency. Addressing these issues requires an efficient method capable of accurately identifying feature correlations.
(2)
In heuristic algorithm applications, many algorithms exhibit low accuracy and fail to identify the optimal subset in feature selection, limiting their practical utility. Thus, designing superior optimization strategies to improve accuracy and applicability has become a current research focus.
(3)
Additionally, heuristic algorithms in feature selection often suffer from premature convergence, getting trapped in local optima and failing to reach the global optimum, thereby limiting their effectiveness. The key to overcoming this lies in enhancing convergence speed and solution diversity to ensure a comprehensive search and optimal solutions.
To address these issues, this study proposes a feature selection technique based on MSCSO. This algorithm minimizes redundant features while balancing error rates and the ratio between selected and original features, aiming for both accuracy and efficiency in feature selection. The proposed improvements include diversified population initialization via logistic chaotic mapping and lens imaging reverse learning, nonlinear parameter adjustment, and individual position updates optimized with Weibull flight and triangular marching strategies. Additionally, Gaussian–Cauchy mutation is integrated to enhance the algorithm’s ability to escape local optima. These enhancements provide a robust mechanism for feature selection, promising to improve stability and effectiveness in practical applications.

3. Improved Sand Cat Swarm Optimization Algorithm

3.1. Improved Population Initialization

As one of the most classical chaotic mapping methods, logistic chaotic mapping enhances the diversity of initial solutions, helping the algorithm perform a more effective global search. The formula for logistic chaotic mapping is given by
X t + 1 = μ X t ( 1 X t )
where μ 0 , 4 , t represents the current iteration count, and X t 0 , 1 .
Reverse learning improves the search capability of the algorithm by generating the reverse solution of the current solution and enlarges the search space of the algorithm. Lens imaging reverse learning is an improvement of reverse learning. By simulating the imaging principle of optical lenses, the reverse solution of the solution is generated so as to further enhance the variability within the population and speed up the algorithm’s convergence. The formula is shown as follows:
X = u b + l b 2 + u b + l b 2 k X k
where u b denotes the solution’s upper bound and l b its lower bound. The term u b + l b 2 in the formula represents the center of the solution space, which is the average of the upper and lower bounds. The terms u b + l b 2 k and X k adjust the current solution X through the scaling factor k . In this study, k is set to 0.75. When k = 1 , the formula reduces to standard opposition-based learning; when k 1 , by simulating the principle of optical lens imaging, the formula changes the distribution range of the opposite solutions, further enhancing the diversity of the population.
First, the initial sand cat population was generated by logistic chaos mapping, and then the corresponding reverse population was generated by lens imaging reverse learning. The two populations were combined, the evaluation of each individual’s fitness value was performed, and they were ranked from smallest to largest. Next, the top N individuals with the highest fitness scores were chosen to constitute the final sand cat population. Compared with the original SCSO, this approach not only accelerates the search process but also greatly increases the search space, thereby enhancing the algorithm’s capability to perform a global search.

3.2. Nonlinear Parameterization

The foraging behavior of sand cats consists of two main stages, namely search and attack. During the search phase, the sand cat conducts global exploration, scanning the entire feasible domain for potential targets. In the attack phase, fine-grained search is focused on a specific area, similar to local optimization. How to coordinate the switching between these two stages effectively is the key to improve the effectiveness of the algorithm. This coordination process relies on the control of the key parameter R . The parameter R is directly related to r g . In the original SCSO algorithm, r g decreases linearly from 2 to 0 as the number of iterations increases. However, to enhance the method’s global exploration capability and avoid local optima traps, the parameters need to be nonlinearized, increasing the number of iterations during the search stage. This helps prevent premature convergence and improves the algorithm’s ability to escape local optima. The specific formula is as follows:
r g = S log 10 S × log 10 ( 2 t T max ) 2
where the initial value of S is 2, t represents the current iteration number, while T max denotes the maximum iterations. The change in r is associated with r g , r is the sensitivity range of each sand cat, r a n d is a value randomly chosen between 0 and 1, and r is calculated by the following formula:
r = r g × r a n d

3.3. Enhanced Location Update Strategy

When R > 1, the sand cat begins searching for prey. The sand cat adjusts its position according to a random candidate location P b c ( t ) , the current location P c ( t ) , the range of sensitivity r , and a r a n d value between 0 and 1. The location update formula is given by
P ( t + 1 ) = r ( P b c ( t ) r a n d P c ( t ) )
During the prey search phase, the location is updated mainly by the random candidate location P b c ( t ) and current location P c ( t ) . If the individuals in the algorithm are too concentrated in the position that is currently considered to be optimal, especially if this position is only a local optimal solution, the amplitude of the overall disturbance will gradually decrease as the iteration progresses. This will cause the whole population to become more and more close to this local optimal region, which will significantly reduce the variety within the population, causing the algorithm to be prone to premature convergence and eventually affecting the optimization accuracy. To solve this problem, the Weibull flight strategy is designed to create new individual locations to introduce more variability into the search process and prevent getting trapped in a local optimal solution prematurely. In searching for prey, the position update formula combined with Weibull’s flight strategy is as follows:
P n e w ( t + 1 ) = r R w ( P b c ( t ) r a n d P c ( t ) )
where P b c ( t ) is the random candidate location, P c ( t ) is the current location, r is the sensitivity range, and R w is a random value drawn from a Weibull distribution. The probability density formula for the Weibull distribution is given by
f ( x , λ , k ) = k λ × x λ k 1 × exp x λ k
where x is the value of the random variable, λ is the scale parameter, k is the shape parameter, and the values of λ and k determine the shape and scale of the distribution. In this study, λ is 0.5 and k is 4.
When R 1 , the sand cat attacks its prey by first using the best location P b ( t ) and the current location P c ( t ) to generate a random location P r n d . Assuming that the sensitivity range of the sand cat is a circle, the roulette method is used to randomly select an angle θ for each sand cat, where the range of θ is from 0 degrees to 360 degrees, and finally, the prey is attacked by Equation (9). To maintain proximity to the prey, the algorithm uses random positioning for the sand cat. Additionally, random angles are used to avoid local optima entrapment.
P r n d = r a n d P b ( t ) P c ( t )
P ( t + 1 ) = P b ( t ) r P r n d cos ( θ )
In attacking prey, the continuous local search can lead to a decrease in population diversity, thus making the algorithm lose enough exploration ability to discover new solution space regions. To address this issue, the triangle parade strategy was added. By introducing a random walk element into local search, the triangular parade enhances the exploration of the algorithm, making it more likely to escape local optima, and helps the sand cat to explore in a wider solution space. The formula for updating positions combined with the triangular parade strategy is as follows:
P n e w ( t + 1 ) = P b ( t ) + r a l p h
where P b ( t ) is the best position, r is the sensitivity range, and a l p h is the offset of the new position. The formula for a l p h is
a l p h = L 2 + L P 2 2 × L P × L cos ( 2 × π × r a n d )
L P = L × r a n d
L = P b ( t ) P c ( t )
where L represents the distance from the sand cat to its current best solution, L P is used to introduce randomness, and r a n d is a random number between 0 and 1.

3.4. Out of Optimal Value

After one iteration, Gaussian and Cauchy variations were used to disturb the optimal position of the sand cat population and update the position of the whole population. Gaussian variation helps individuals fine-tune their position in the solution space by introducing small amplitude random changes and enhances local search ability and optimizes solutions more effectively while avoiding premature convergence. In contrast, Cauchy variation introduces large random perturbations that allow individuals to jump out of local optimal solutions, thereby improving global exploration capabilities, and the formula for the perturbation position of Gaussian–Cauchy variation is
P h ( t ) = P b ( t ) × ( 1 + Δ P ( t ) )
Δ P ( t ) = α × G a u s s i a n ( 1 , 0 ) + β × c a u c h y ( 1 , 0 )
α = t / T max
β = 1 α
where P h ( t ) is the position after perturbation by Cauchy–Gaussian variation, P b ( t ) is the best position, t represents the current iteration, T max denotes the total number of iterations, and G a u s s i a n ( 1 , 0 ) is a random variable based on Gaussian distribution. c a u c h y ( 1 , 0 ) denotes a random variable that follows a Cauchy distribution.
In order to further optimize the population after the position disturbance of Gaussian–Cauchy variation, an optimal strategy, the greedy strategy, was adopted. By comparing the fitness scores of the modified individual and the original individual, the strategy selects the individual with better fitness. If the mutated individual has better fitness, it replaces the original individual, thereby improving the overall performance of the algorithm and ensuring that each generation moves toward the optimal solution by accurately selecting better solutions and robustly converging toward the global optimum. The formula for the greedy strategy is
P c ( t ) = P c ( t ) f ( P h ( t ) ) f ( P c ( t ) ) P h ( t ) f ( P h ( t ) ) < f ( P c ( t ) )
where P c ( t ) is the current position, P h ( t ) is the position after Cauchy–Gaussian variation perturbation, and f is the fitness function.

3.5. MSCSO for Feature Selection

Suppose S is a dataset with K samples and M features and F is a set of M features. The purpose of feature selection is to optimize the objective function by selecting the optimal feature subset Y b ( Y b M ) , since the setting of the objective function is to maintain a balance between minimizing the error rate and minimizing the number of features selected. Therefore, this study chooses the linear combination of error rate and the number of selected features as the objective function, and the formula is as follows:
f = λ × e r r o r + ( 1 λ ) × ( 1 Y M )
In order to apply MSCSO to feature selection, it is necessary to binarize the sand cat individual. The details are shown in Formula (23).
Y i = y i 1 , y i 2 , .... y i M , y i j 0 , 1
where y i j = 1 means that the j feature in the i feature subset Y i is selected, and y i j = 0 means that this feature is not selected. The feature selection problem in this study can be expressed as an optimization problem, and the formula is as follows:
min f ( Y ) s . t . Y i = ( y i 1 , y i 2 , .... y i M ) , y i j { 0 , 1 } j = 1 , 2 , .... M 1 | Y | M
where min f ( Y ) is the minimization objective function and s . t . is the constraint.
MSCSO was originally used to solve the problem of continuous variables, but the value of the feature subset is limited to {0, 1}, so to convert continuous values between [0, 1] to binary values, 0.5 is used as the threshold to decide whether to select a feature. When the value in the individual is greater than 0.5, it is converted to 1, indicating that the feature has been selected. When the value is less than or equal to 0.5, it is converted to 0, indicating that the feature is not selected [2,3,31].

3.6. Complexity Analysis of MSCSO

The time complexity of this algorithm consists of the initialization phase and multiple steps in each iteration. In the initialization phase, the algorithm first generates 2 × N positions and initializes the population using chaotic mapping and opposition-based learning, with a time complexity of O ( 2 × N × D ) , where N is the number of search agents and D is the feature dimension. Then, it calculates the fitness of each individual and performs sorting, with a time complexity of O ( 2 × N × D + 2 × N l o g ( 2 × N ) ) . In each iteration, position updates, binarization, fitness calculations, and boundary checks are performed for each individual, with a time complexity of O ( N × D ) . The position update includes calculating random factors, Lévy flight, Weibull flight, and other O ( D ) operations. Gaussian and Cauchy mutations are applied to all individuals for perturbation, with a complexity of O ( N × D ) . Combined with fitness recalculations and greedy selection, the overall complexity remains O ( N × D ) . Assuming a total number of iterations T max , the overall time complexity is O ( T m a x × N × D ) , with the number of iterations being the primary computational cost.
In this section, the original sand cat algorithm is mainly improved, and the mathematical model of MSCSO is depicted in Algorithm 1, with the overall flowchart depicted in Figure 1.
Algorithm 1 Pseudocode for the MSCSO optimization algorithm
1: Initialization parameters: population size N , maximum iteration count T max , upper and lower limits of solutions u b and l b .
2: Initialize the optimal position of the sand cat P b and the optimal fitness value P b e s t .
3: Fitness function f ( x ) : see Equation (19).
4: Population initialization: to initialize the sand cat population, use Equations (1) and (2) to compute fitness value f ( P i ) of every sand cat individual, and take the first N to form a new sand cat population.
5: While t < T max do
6: for every individual sand cat P i do
7:  To ensure that each individual sand cat P i is in the limits of the solution and to calculate the P i , the fitness value of the f ( P i ) is used.
8:   if f ( P i ) < P b e s t , then
9:    Assign the value of f ( P i ) to P b e s t , updating the optimal location P b .
10:  end if
11: end for
12: Set S to 2 and initialize rg and r according to Equations (3) and (4).
13:  if | R | > 1 then
14:  Update individual positions according to Equation (5).
15:  Use the Weibull flight strategy to obtain a new position, see Equation (6).
16: else
17:  Adjust positions based on Formula (9).
18:  Use the triangle parade to obtain a new position, see Equation (10).
19: end if
20: Calculate the fitness of the individual value f ( P i ) and the new position of fitness value f ( P n e w i ) .
21: if f ( P n e w i ) < f ( P i ) , then
22:   Assign f ( P n e w i ) to P i .
23: end if
24: Use Formulas (15) and (18) to update the location of the entire sand cat population.
25: Use Formula (19) to ensure excellence.
26: If the number of iterations increases by 1, t = t + 1.
27: end while

4. Performance Metrics of Experiment

Two key experiments are conducted in this section to thoroughly examine the effectiveness and precision of the MSCSO approach introduced in this research. The first experiment aimed to assess how well the MSCSO algorithm performs on global optimization problems, using the CEC2005 reference function set, which covers a variety of optimization problems from simple to complex, including unimodal functions, multimodal functions, and fixed-dimensional multimodal functions. Through these benchmark functions, we can methodically assess how well the MSCSO algorithm performs and maintains stability across various optimization challenges. The second experiment focused on feature selection, using 15 classical datasets covering different domains and feature selection difficulties. By applying the MSCSO algorithm to these datasets, its performance in actual data processing can be evaluated. Feature selection, as a key step in machine learning and data mining, has a major effect on the performance and complexity of the model, so it is of great significance to evaluate the effect of the MSCSO algorithm in this task. All experiments were conducted on the 64-bit version of MATLAB R2022a to ensure a consistent experimental environment. In these two experiments, the results of the MSCSO algorithm are evaluated against the original SCSO algorithm and several other commonly used meta-heuristic algorithms in detail, and a variety of performance indicators are used to evaluate these algorithms so as to fully understand the relative advantages and application potential of the MSCSO algorithm.

4.1. Performance Metrics of Experiment

To assess the effectiveness of MSCSO, the following evaluation indicators were used in this study:
Average fitness: Fitness is a joint evaluation of the classification error rate and the proportion of selected features relative to the total number of features. The smaller the fitness, the better the effect, which means reducing the classification error rate while minimizing the selection of feature count to achieve the optimal effect. The average fitness is calculated as the mean of several separate runs, and its expression is shown as follows:
f m e a n = i = 1 R f i R
where R is the number of independent runs. In global optimization, R is assigned a value of 30 for the number of independent trials, whereas in feature selection, R is assigned a value of 20 and f i denotes the fitness value obtained from the initial run i .
Standard deviation: This is used to assess the consistency and robustness of the optimization algorithm across multiple runs. A reduced standard deviation indicates a more stable optimization algorithm. The formula is as follows:
s t d = i = 1 R ( x i x m e a n ) 2 R
where R is the number of independent runs and x i and x m e a n represent the measured value and average value in the i run, respectively.
Average accuracy: this is the average of classification accuracy after the R run, defined as
a c c = i = 1 R a c c i R
where a c c i is the accuracy of the first run of i .
Average feature selection: this represents the mean number of selected features across R runs, defined as
f e a t u r e = i = 1 R s i z e ( i ) R
where s i z e ( i ) indicates the count of features chosen during the i -th run.

4.2. Parameter Setting of Algorithm

Table 1 presents the parameter configurations for MSCSO, SCSO, Moth Flame Optimization (MFO) [39], the Salp Swarm Algorithm (SSA) [40], Four-Vector Optimization (FVIM) [41], the Arithmetic Optimization Algorithm (AOA) [42], the SCA [43], PSO [44], Dandelion Optimization (DO) [45], and Sailfish Optimization (SFO) [46]. These settings significantly impact algorithm efficiency and the outcomes of the experiments. Specifically, the table includes the number of iterations, population size, and the specific parameter configurations for each algorithm.

4.3. Experiment 1: Global Optimization of CEC2005 Test Functions

This section assesses how well the MSCSO algorithm performs with the CEC2005 benchmark functions, as shown in Supplementary Materials Table S1. Among the 23 test functions, F1 to F7 are unimodal benchmark functions, F8 to F13 are multimodal benchmark functions, and F14 to F23 are fixed-dimension multimodal benchmark functions. Unimodal benchmark functions mainly test the algorithm’s ability to perform local searches, while multimodal benchmark functions examine the algorithm’s ability to escape local optima and find the global optimum. Fixed-dimension multimodal benchmark functions evaluate how well the algorithm performs in a global search at specific dimensions. To thoroughly assess the effectiveness of the MSCSO algorithm, nine commonly used heuristic algorithms are chosen for comparison, including standard SCSO, MFO, the SSA, FVIM, the AOA, the SCA, PSO, DO, and SFO. All algorithms are assessed under the same experimental conditions to maintain the integrity and consistency of the results. Each algorithm uses 30 search agents and performs 1000 iterations. To minimize the effects of random initialization on the results, each algorithm is executed 30 times separately, with the performance assessed by averaging fitness values and calculating the standard deviation.

4.3.1. Numerical and Statistical Analysis of Experiment 1

Table 2 presents a comparison of the mean fitness and standard deviation between MSCSO and existing optimization algorithms for the CEC2005 test functions. It can be seen from the ranking of average fitness that MSCSO outperforms other optimization algorithms on 15 test functions. Specifically, MSCSO performs best on F1–F5, F8–F11, F15, F16, and F20–F23; performs second best on F7, F12, F13, and F14; ranks third for F17, F18, and F19; and has relatively weaker competitiveness in F6. Overall, MSCSO demonstrates the best performance among all compared algorithms.

4.3.2. Comparative Analysis of MSCSO and Other Algorithms on Test Functions

In the CEC2005 test functions, MSCSO performed significantly better than SCSO. Specifically, MSCSO outperformed SCSO in 17 test functions, while SCSO only outperformed MSCSO in F17 and F18, both of which performed equally on F9–F11 and F16. MSCSO vs. MFO: MSCSO performed well on 19 test functions, with the same performance on F16, only underperforming compared to MFO on F17–F19. MSCSO compared to the SSA: the SSA had better average fitness on F6, F14, F17, F18, and F19, with the same average fitness on F16, and MSCSO performed best on the remaining 17 test functions. MSCSO vs. FVIM: MSCSO outperforms FVIM on 23 test functions. MSCSO vs. the AOA: the AOA won only on F7, achieving the same score on F2, F9, and F10, and MSCSO achieved better results on the remaining 19 test functions. MSCSO vs. the SCA: MSCSO is 100 percent better than the SCA and has a lower mean fitness. Comparison between MSCSO and PSO: Among the 23 test functions, MSCSO achieved optimal results in 17 test functions; PSO was superior to MSCSO in F6, F14, F17, F18, and F19, and the two algorithms had the same results in F16. MSCSO vs. DO: DO outperforms MSCSO in F6, F12, F13, F14, F17, F18, and F19 and performs the same in F16, and MSCSO performs better in the remaining 15 test functions. MSCSO vs. SFO: SFO failed to surpass MSCSO in any of the test functions, only scoring the same as MSCSO in F8 and F11.
Table 2 further shows that MSCSO ranks highest among all comparison algorithms, while the SCA ranks lowest. The order is MSCSO > SCSO = DO > FVIM > SSA > PSO > SFO > AOA > MFO > SCA.
Wilcoxon tests are widely used in statistical analysis to compare the differences between two algorithms. Table 3 records the p-values of the Wilcoxon test between MSCSO and existing optimization algorithms. These p-values indicate whether significant differences exist between the performance of the two algorithms across various test functions. When the p-value falls below 0.05, it indicates that the difference between MSCSO and comparison algorithms in this test function is statistically significant. Conversely, if the p-value is 0.05 or higher, it suggests that there is no significant performance disparity between the two algorithms for the given test function. p-values above 0.05 are shown as underlined below.

4.3.3. Graphic Analysis of Experiment 1

This section primarily conducts an in-depth analysis of the performance of various optimization algorithms through convergence curve diagrams, assessing their performance in solving the CEC2005 test functions. By observing the convergence curves, one can intuitively understand the convergence speed and final convergence state of each algorithm during the iteration process. Figure 2 illustrates the convergence behavior of MSCSO compared to other optimization methods.
From the results of single-peak reference functions F1–F7, MSCSO has significant performance advantages, especially in the performance of F1–F4, where the MSCSO algorithm’s convergence curve quickly reaches the optimal value, while other algorithms remain trapped in local optima, with their curves no longer showing a downward trend as iterations increase. For F5, MSCSO also achieved relatively optimal results. However, in F6 and F7, the optimization results of MSCSO are not good, showing only a certain degree of improvement from the original algorithm. In the multimodal benchmark functions F8–F13, MSCSO performs well in the optimization of multimodal complex dimensional functions. In F8–F11, MSCSO outperforms other optimization algorithms in both convergence accuracy and speed, while the other methods exhibit oscillations in their convergence curves. For F12 and F13, MSCSO’s final optimization results are not as good as DO. In the fixed-dimensional multimodal reference functions F14–F23, these test functions have low dimensionality and are more similar to real-world optimization problems. MSCSO shows better convergence accuracy and more stable convergence curve for F20–F23.
Integrating the above analysis, MSCSO demonstrates, in comparison with other swarm intelligence optimization algorithms, such as SCSO, MFO, the SSA, FVIM, the AOA, the SCA, PSO, DO, and SFO, a generally faster convergence speed, higher convergence accuracy, and more stable convergence performance.

4.4. Experiment 2: Feature Selection

4.4.1. Feature Selection Dataset Collection and Preprocessing

This study’s feature selection experiments used 15 datasets from the well-established UCI Machine Learning Repository, a resource that spans various application domains and supports the development and evaluation of machine learning algorithms. The selected 15 datasets span a range of data characteristics and complexities, ensuring the broad applicability and representativeness of the experimental results. Table 4 provides detailed information for each dataset, including names, sample count, feature count, and class count.
For datasets containing missing values, numerical missing values are filled by the median and categorical missing values are filled by the mode. The dataset is split into training and test sets with an 80/20 ratio, and the KNN algorithm is applied with K = 5. All algorithms are evaluated under the same experimental conditions to ensure fairness and comparability. Each algorithm uses 30 search agents for 100 iterations. To mitigate the impact of random initialization, each algorithm is independently executed 20 times.

4.4.2. Numerical and Statistical Analysis of Experiment 2

This section assesses MSCSO’s performance in feature selection using various indicators. Table 5, Table 6 and Table 7 display the comparison of MSCSO with other optimization algorithms across different datasets. In order to highlight the best performance of the algorithm, the best values in the table are identified in bold font. To provide a clearer comparison of overall performance, the final rows summarize each algorithm’s average and overall rankings across 15 datasets.
Table 5 presents the average fitness and standard deviation for various optimization methods across each dataset. The data in the analysis table shows that MSCSO performs well on most datasets. Of the 15 datasets tested, MSCSO achieved the best average fitness on 14 datasets, accounting for up to 93.33%, showing a significant advantage. The only exception was the M-of-n dataset, where the mean fitness of MSCSO was slightly lower than that of the FVIM optimization algorithm, although this difference was not significant and had a more limited impact on the overall results. Further analysis revealed that MSCSO performed particularly well in comparison to the original SCSO, outperforming SCSO in all datasets. In addition, the overall ranking shows that MSCSO is the top and best performing algorithm, while PSO is the lowest ranking algorithm. The ranking order is MSCSO > AOA > SCSO > DO > FVIM > MFO > SCA > SSA > SFO > PSO. Referring further to Figure 3, which visually shows the average fitness of different optimization algorithms across 15 datasets, it is clear that MSCSO has a lower mean fitness value and a better effect.
Table 5. Comparison table of mean fitness and standard deviations between MSCSO and other optimization algorithms in feature selection (Bold indicates better data performance).
Table 5. Comparison table of mean fitness and standard deviations between MSCSO and other optimization algorithms in feature selection (Bold indicates better data performance).
DatasetMeasuresMSCSOSCSOMFOSSAFVIMAOASCAPSODOSFO
ZooMean0.04110.04310.04440.04760.04280.04240.04830.05880.04120.0701
Std0.00640.00680.00640.00640.00610.00470.00740.09530.00620.0081
WineMean0.03160.03240.03390.03720.03330.03180.03430.04690.03290.0440
Std0.00370.00340.00440.00250.00420.00370.00430.09640.00420.0094
VoteMean0.08260.09030.09550.10420.09060.08970.09980.11520.08760.1073
Std0.00920.00830.00860.00750.01070.00770.01150.09010.00940.0164
LymphographyMean0.10210.10990.11370.11960.11100.10890.11150.12740.10650.1468
Std0.01340.01240.01080.00800.01110.00980.01470.08860.01170.0085
HeartEWMean0.11310.11970.11420.13200.11610.11850.12730.13950.11670.1553
Std0.01120.01180.01230.01060.01150.00780.01440.08760.00970.0132
SonarMean0.04520.04870.05200.05320.04980.05120.05140.06420.05130.0611
Std0.00340.00390.00320.00320.00220.00170.00600.09460.00310.0116
SpectEWMean0.14930.15800.16130.16520.16220.15930.16550.17210.16260.1862
Std0.01140.00920.00980.00840.00900.00750.01820.08420.01170.0103
Lung-CancerMean0.05150.05530.05830.06450.06160.05160.05570.07060.05480.1541
Std0.02350.01980.02190.02570.06160.00800.01290.09590.01870.0260
BreastEWMean0.03120.03310.03570.03720.00110.03550.03580.04950.03460.0379
Std0.00260.00210.00170.00110.00200.00140.00370.09610.00160.0064
CongressEWMean0.03150.03490.03790.04120.03480.03340.04020.05670.03590.0483
Std0.00540.00670.00590.00410.00550.00440.00590.09550.00620.0112
Clean1Mean0.12000.12720.13360.13860.13150.12790.13460.14930.12460.1417
Std0.00660.00620.00580.00370.00490.00350.01400.08620.00610.0113
ExactlyMean0.06350.06900.08120.10260.06620.06380.08700.10050.06520.2544
Std0.03590.04240.05460.04470.04120.02930.03800.10150.04300.0096
Rank15694278310
Exactly2Mean0.17270.17400.17700.18120.17570.17390.17820.19160.17400.1833
Std0.00340.00360.00400.00380.00400.00310.01810.08180.00380.0147
M-of-nMean0.05410.05610.05920.06540.05370.05410.06540.07340.05760.1169
Std0.01480.01600.01830.01750.01410.01290.01580.09500.01550.0105
VPMean0.08950.09150.09420.09480.09400.09450.09320.11010.09420.0935
Std0.00160.00070.00040.00010.00020.00010.00940.08990.00020.0065
Avg. rank1.073.475.537.734.403.206.409.203.608.93
Table 6 provides detailed data on the number of features selected by each optimization method. The count of features selected is a crucial index for assessing the effectiveness of optimization methods. From the table data, it is clear that MSCSO selected the least number of features on average in 13 datasets, and only in the SpectEW and Clean1 datasets was the number of features selected lower than that of SFO, ranking second. On the Wine dataset, MSCSO and the AOA each select an average of 3.65 features, while the DO algorithm selects an average of 5.95 features in the Zoo dataset. In summary, MSCSO ranked first in the number of feature selection, SCSO ranked second, and PSO ranked lowest. In terms of standard deviation, MSCSO showed the lowest standard deviation among the nine datasets, indicating excellent stability and robustness; that is, consistent performance over multiple runs. Figure 4 further highlights the advantage of MSCSO in feature selection. The figure illustrates that MSCSO selects the least number of features on average in multiple datasets, which further proves the superiority of MSCSO in the process of feature number selection.
Table 6. Comparison table of average feature selection size and standard deviation between MSCSO and other optimization algorithms (Bold indicates better data performance).
Table 6. Comparison table of average feature selection size and standard deviation between MSCSO and other optimization algorithms (Bold indicates better data performance).
DatasetMeasuresMSCSOSCSOMFOSSAFVIMAOASCAPSODOSFO
ZooMean5.956.26.157.0566.27.17.155.958.15
Std0.39400.61560.58710.75920.64890.83350.55250.48940.60482.1831
WineMean3.653.83.94.553.93.654.34.253.855.15
Std0.48940.61560.30780.60480.44720.48940.57120.63870.48941.0894
VoteMean4.34.84.44.954.64.94.754.854.454.8
Std0.65691.15170.94031.27630.99471.02080.96650.81270.75921.3992
LymphographyMean8.258.458.99.459.19.258.59.18.759.15
Std1.48240.99871.29371.05011.02081.11801.39551.07111.20851.4965
HeartEWMean44.854.255.754.254.955.856.14.955.35
Std0.85841.26801.01951.40960.85071.76141.26801.55261.50351.1821
SonarMean25.927.6529.930.1528.6529.9529.853129.1528.1
Std1.48322.10952.26882.23082.45542.08942.15881.55601.89942.5526
SpectEWMean10.4511.412.5512.612.412.7513.1512.1512.29.7
Std2.23552.16192.23552.62382.34861.91602.15882.79612.04171.5927
Lung-CancerMean23.926.128.229.4527.6526.928.830.0527.2525.55
Std2.53192.91821.90842.48102.27753.16063.69352.11451.91603.0517
BreastEWMean8.859.3510.110.659.91010.4510.959.710.9
Std0.81270.81270.91191.53131.20961.12391.19100.99870.97871.1653
CongressEWMean2.6533.154.23.153.14.254.452.954.65
Std0.48940.79470.67081.00520.81270.55250.63870.88700.94451.6311
Clean1Mean89.6595.89799.799.195.3594.5102.695.680.15
Std9.149510.28348.404310.45848.97318.78569.73065.43287.97636.8077
ExactlyMean6.056.256.57.156.26.256.956.956.16.5
Std0.22360.44430.51300.48940.41040.44430.39400.51040.30782.1643
Exactly2Mean11.251.351.851.251.152.152.21.052.45
Std00.44430.48940.74520.44430.36630.58710.89440.22360.8256
M-of-nMean6.16.156.356.856.156.156.86.96.37.1
Std0.30780.36630.48940.67080.36630.36630.61560.44720.47021.6827
VPMean52.2555.658.85859.159.7557.2567.357.6554.25
Std2.24493.69073.98154.36494.94074.32715.43742.97536.51545.4374
Avg. rank1.133.275.077.334.535.136.007.803.475.47
Additionally, to evaluate MSCSO’s effectiveness, the average accuracy of feature subsets selected by various optimization algorithms is compared, with the results presented in Table 7. The analysis showed that MSCSO achieved the highest accuracy across all 15 datasets. In contrast, SCSO and MFO performed well in nine datasets, the AOA, DO, and FVIM performed well in eight datasets, PSO achieved high accuracy in six datasets, the SSA and SCA performed well in five datasets, and SFO performed well only in Elactly2. Overall, MSCSO outperforms other optimization methods in average accuracy, which fully proves that it still has an excellent capability of relevant feature recognition even when the feature subset size is not dominant. For example, in the SpectEW dataset, while SFO selected the lowest average number of features at 9.7, MSCSO achieved a higher accuracy of 0.8991 with a slightly higher number of features at 10.45 compared to SFO’s accuracy of 0.8434. In the Clean1 dataset, SFO selected an average of 80.15 features, while MSCSO selected 89.65 features, with accuracy rates of 0.8979 and 0.9337, respectively. In addition, MSCSO selected the lowest average number of features in some datasets but still achieved the highest accuracy. As a result, MSCSO achieves a better balance between maximizing accuracy and minimizing feature subsets. Figure 5 illustrates the average accuracy across 15 datasets, further highlighting the benefits of MSCSO.
Table 7. Comparison table of average accuracy and standard deviations of MSCSO and other optimization algorithms (Bold indicates better data performance).
Table 7. Comparison table of average accuracy and standard deviations of MSCSO and other optimization algorithms (Bold indicates better data performance).
DatasetMeasuresMSCSOSCSOMFOSSAFVIMAOASCAPSODOSFO
ZooMean1111111110.9800
Std0000000000.0251
WineMean1111111110.9971
Std0000000000.0088
VoteMean0.94670.94250.93330.92750.94330.94170.92750.92920.94420.9167
Std0.00680.01010.01080.01350.01000.00860.01240.01420.00820.0162
LymphographyMean0.95340.94310.94140.93280.94660.94830.93620.93280.94830.8948
Std0.01690.01690.01620.00770.01760.01770.01260.00770.01770.0136
HeartEWMean0.92410.92220.92130.91300.92220.92310.91480.91940.92310.8750
Std0.00570.00760.00820.00870.00760.00680.0093 0.00910.00680.0158
SonarMean1111111110.9866
Std0000000000.0124
SpectEWMean0.89910.89340.89810.88870.89250.89530.89150.89150.89620.8434
Std0.01110.01760.01280.0161 0.01510.01560.01480.01610.01150.0195
Lung-CancerMean1111111110.8833
Std0000000000.0784
BreastEWMean1110.99960.99910.99910.99910.99960.99960.9991
Std0000.00200.00270.00270.00270.00200.00200.0027
CongressEWMean0.98840.98840.98840.98780.98840.98840.98780.98660.98840.9802
Std0000.0046000.00260.004300.0114
Clean1Mean0.93370.93050.92580.91740.92630.92580.91470.91950.93320.8979
Std0.01030.00930.1110.01040.00840.00930.00900.00520.01290.0128
ExactlyMean1110.98280.999510.99050.991510.7748
Std0000.02120.002200.01500.017100.0983
Exactly2Mean0.820.820.820.820.820.820.820.820.820.82
Std00000000.003100
M-of-nMean1110.9975110.99830.999310.9323
Std0000.0062000.00540.003400.0513
VPMean0.94740.94700.94700.94510.94700.94700.94510.94740.94590.9451
Std00.00170.00170.00350.00170.00170.003500.00310.0035
Avg. rank122.23.932.22.073.673.21.64.87

4.4.3. Graphical Analysis of Experiment 2

Figure 6 shows the convergence curve in the feature selection task. By analyzing the convergence curve, the convergence speed and final convergence effect of different algorithms can be evaluated. The figure clearly shows that MSCSO achieved the lowest mean fitness values on eight datasets, which were Vote, Lymphography, Sonar, SpectEW, Lung-Cancer, BreastEW, CongressEW, and Clean1. This indicates that the algorithm excels with these datasets. Further combining the results of Table 5, MSCSO also performed well on the Zoo, Wine, HeartEW, Exactly, Exactly2, and VP datasets, only slightly worse than FVIM optimization on the M-of-n dataset. In addition, it is worth noting that MSCSO performed better on mid-dimensional datasets (such as Sonar, Lung-Cancer, and BreastEW) than it did on low-dimensional datasets and that the performance gap between MSCSO and other algorithms was greater in mid-dimensional datasets than in low-dimensional datasets. These results show that the MSCSO algorithm has significant advantages in feature selection tasks, especially on medium-dimensional datasets.

5. Conclusions and Future Directions

The MSCSO algorithm proposed in this study has demonstrated significant advantages in global optimization and feature selection tasks. In the CEC2005 benchmark tests, MSCSO excelled on 65.2% of the test functions, achieving the best average fitness and outperforming nine common heuristics in terms of both convergence speed and precision. The Wilcoxon test further confirmed the significant performance differences between MSCSO and other algorithms. During feature selection, through experiments on 15 UCI datasets, MSCSO achieved the best average fitness in 93.3% of the datasets and selected the least number of features in 86.7% of the datasets. Moreover, MSCSO also achieved the highest classification accuracy across all datasets. These results indicate that MSCSO has shown excellent effectiveness in global optimization and feature selection tasks.
In the future, the further development of MSCSO can be carried out in the following directions: firstly, MSCSO can be combined with other optimization algorithms to form a hybrid algorithm to give full play to their respective advantages. Secondly, exploring a variety of binarization methods may further improve the efficiency and accuracy of feature selection. Finally, the application potential of MSCSO is not limited to the current experimental field and can also be extended to image processing, bioinformatics, financial data analysis, and other fields to verify its applicability and effectiveness in a wider range of scenarios.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/biomimetics9110701/s1, Table S1: CEC2005 test function details.

Author Contributions

R.L.: data curation, conceptualization, methodology, software, writing—original draft, visualization, validation. R.F.: supervision, data curation, visualization, validation. T.Z.: investigation, project administration, validation. H.F.: data curation, supervision, validation. Q.Q.: data curation, software. P.Z.: investigation, software. L.X.: supervision, validation. W.L.: supervision, writing—review and editing, funding acquisition, validation. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by [National Natural Science Foundation of China] grant number [32460290], [the Xinjiang Production and Construction Corps Science and Technology Program] grant number [2023CB008-23] and [the Key Areas Science and Technology Research Plan of the Xinjiang Production And Construction Corps Financial Science and Technology Plan Project] grant number [grant number 2023AB048] And The APC was funded by [National Natural Science Foundation of China].

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zebari, R.; Abdulazeez, A.; Zeebaree, D.; Zebari, D.; Saeed, J. A comprehensive review of dimensionality reduction techniques for feature selection and feature extraction. J. Appl. Sci. Technol. Trends 2020, 1, 56–70. [Google Scholar] [CrossRef]
  2. Houssein, E.H.; Oliva, D.; Celik, E.; Emam, M.M.; Ghoniem, R.M. Boosted sooty tern optimization algorithm for global optimization and feature selection. Expert Syst. Appl. 2023, 213, 119015. [Google Scholar] [CrossRef]
  3. Chhabra, A.; Hussien, A.G.; Hashim, F.A. Improved bald eagle search algorithm for global optimization and feature selection. Alex. Eng. J. 2023, 68, 141–180. [Google Scholar] [CrossRef]
  4. He, J.; Qu, L.; Wang, P.; Li, Z. An oscillatory particle swarm optimization feature selection algorithm for hybrid data based on mutual information entropy. Appl. Soft Comput. 2024, 152, 111261. [Google Scholar] [CrossRef]
  5. Sampson, J.R. Adaptation in Natural and Artificial Systems (John H. Holland); The MIT Press: Cambridge, MA, USA, 1976. [Google Scholar]
  6. Storn, R.; Price, K. Differential evolution–a simple and efficient heuristic for global optimization over continuous spaces. J. Glob. Optim. 1997, 11, 341–359. [Google Scholar] [CrossRef]
  7. Huning, A. Evolutionsstrategie. Optimierung Technischer Systeme Nach Prinzipien der Biologischen Evolution; Franz Steiner Verlag: Stuttgart, Germany, 1976. [Google Scholar]
  8. Gandomi, A.H.; Alavi, A.H. Krill herd: A new bio-inspired optimization algorithm. Commun. Nonlinear Sci. Numer. Simul. 2012, 17, 4831–4845. [Google Scholar] [CrossRef]
  9. Hashim, F.A.; Houssein, E.H.; Hussain, K.; Mabrouk, M.S.; Al-Atabany, W. Honey badger algorithm: New metaheuristic algorithm for solving optimization problems. Math. Comput. Simul. 2022, 192, 84–110. [Google Scholar] [CrossRef]
  10. Mohammed, H.; Rashid, T. Fox: A fox-inspired optimization algorithm. Appl. Intell. 2023, 53, 1030–1050. [Google Scholar] [CrossRef]
  11. Kirkpatrick, S.; Gelatt, C.D., Jr.; Vecchi, M.P. Optimization by simulated annealing. Science 1983, 220, 671–680. [Google Scholar] [CrossRef]
  12. Talatahari, S.; Azizi, M.; Tolouei, M.; Talatahari, B.; Sareh, P. Crystal structure algorithm (crystal): A metaheuristic optimization method. IEEE Access 2021, 9, 71244–71261. [Google Scholar] [CrossRef]
  13. Rashedi, E.; Nezamabadi-Pour, H.; Saryazdi, S. Gsa: A gravitational search algorithm. Inf. Sci. 2009, 179, 2232–2248. [Google Scholar] [CrossRef]
  14. Gao, Y.; Zhang, J.; Wang, Y.; Wang, J.; Qin, L. Love evolution algorithm: A stimulus–value–role theory-inspired evolutionary algorithm for global optimization. J. Supercomput. 2024, 80, 12346–12407. [Google Scholar] [CrossRef]
  15. Binu, D.; Kariyappa, B. Ridenn: A new rider optimization algorithm-based neural network for fault diagnosis in analog circuits. IEEE Trans. Instrum. Meas. 2018, 68, 2–26. [Google Scholar] [CrossRef]
  16. Nemati, M.; Zandi, Y.; Agdas, A.S. Application of a novel meta-heuristic algorithm inspired by stadium spectators in global optimization problems. Sci. Rep. 2024, 14, 3078. [Google Scholar]
  17. Bai, J.; Li, Y.; Zheng, M.; Khatir, S.; Benaissa, B.; Abualigah, L.; Wahab, M.A. A sinh cosh optimizer. Knowl.-Based Syst. 2023, 282, 111081. [Google Scholar] [CrossRef]
  18. Zhao, S.; Zhang, T.; Cai, L.; Yang, R. Triangulation topology aggregation optimizer: A novel mathematics-based meta-heuristic algorithm for continuous optimization and engineering applications. Expert Syst. Appl. 2024, 238, 121744. [Google Scholar] [CrossRef]
  19. Abdel-Basset, M.; El-Shahat, D.; Jameel, M.; Abouhawwash, M. Exponential distribution optimizer (edo): A novel math-inspired algorithm for global optimization and engineering problems. Artif. Intell. Rev. 2023, 56, 9329–9400. [Google Scholar] [CrossRef]
  20. Wolpert, D.H.; Macready, W.G. No free lunch theorems for optimization. IEEE Trans. Evol. Comput. 1997, 1, 67–82. [Google Scholar] [CrossRef]
  21. Seyyedabbasi, A.; Kiani, F. Sand cat swarm optimization: A nature inspired algorithm to solve global optimization problems. Eng. Comput. 2023, 39, 2627–2651. [Google Scholar] [CrossRef]
  22. Karami, S.; Saberi-Movahed, F.; Tiwari, P.; Marttinen, P.; Vahdati, S. Unsupervised feature selection based on variance–covariance subspace distance. Neural Netw. 2023, 166, 188–203. [Google Scholar] [CrossRef]
  23. Cheng, J.; Sun, J.; Yao, K.; Xu, M.; Cao, Y. A variable selection method based on mutual information and variance inflation factor. Spectrochim. Acta Part A Mol. Biomol. Spectrosc. 2022, 268, 120652. [Google Scholar] [CrossRef] [PubMed]
  24. Shafizadeh-Moghadam, H.; Minaei, M.; Pontius, R.G., Jr.; Asghari, A.; Dadashpoor, H. Integrating a forward feature selection algorithm, random forest, and cellular automata to extrapolate urban growth in the Tehran-Karaj region of Iran. Comput. Environ. Urban Syst. 2021, 87, 101595. [Google Scholar] [CrossRef]
  25. Awad, M.; Fraihat, S. Recursive feature elimination with cross-validation with decision tree: Feature selection method for machine learning-based intrusion detection systems. J. Sens. Actuator Netw. 2023, 12, 67. [Google Scholar] [CrossRef]
  26. Afrin, S.; Shamrat, F.J.M.; Nibir, T.I.; Muntasim, M.F.; Moharram, M.S.; Imran, M.; Abdulla, M. Supervised machine learning based liver disease prediction approach with lasso feature selection. Bull. Electr. Eng. Inform. 2021, 10, 3369–3376. [Google Scholar] [CrossRef]
  27. Malik, A.; Jamei, M.; Ali, M.; Prasad, R.; Karbasi, M.; Yaseen, Z.M. Multi-step daily forecasting of reference evapotranspiration for different climates of India: A modern multivariate complementary technique reinforced with ridge regression feature selection. Agric. Water Manag. 2022, 272, 107812. [Google Scholar] [CrossRef]
  28. Li, X.D.; Wang, J.S.; Liu, Y.; Song, H.M.; Wang, Y.C.; Hou, J.N.; Zhang, M.; Hao, W.K. Classification feature selection and dimensionality reduction based on logical binary sine-cosine function arithmetic optimization algorithm. Egypt. Inform. J. 2024, 26, 100472. [Google Scholar] [CrossRef]
  29. Qiu, Y.; Li, R.; Zhang, X. Simultaneous svm parameters and feature selection optimization based on improved slime mould algorithm. IEEE Access 2024, 12, 18215–18236. [Google Scholar] [CrossRef]
  30. Hussien, R.M.; Abohany, A.A.; Abd El-Mageed, A.A.; Hosny, K.M. Improved binary meerkat optimization algorithm for efficient feature selection of supervised learning classification. Knowl. Based Syst. 2024, 292, 111616. [Google Scholar] [CrossRef]
  31. Hashim, F.A.; Houssein, E.H.; Mostafa, R.R.; Hussien, A.G.; Helmy, F. An efficient adaptive-mutated coati optimization algorithm for feature selection and global optimization. Alex. Eng. J. 2023, 85, 29–48. [Google Scholar] [CrossRef]
  32. Pan, H.; Chen, S.; Xiong, H. A high-dimensional feature selection method based on modified gray wolf optimization. Appl. Soft Comput. 2023, 135, 110031. [Google Scholar] [CrossRef]
  33. Peng, L.; Cai, Z.; Heidari, A.A.; Zhang, L.; Chen, H. Hierarchical Harris Hawks optimizer for feature selection. J. Adv. Res. 2023, 53, 261–278. [Google Scholar] [CrossRef]
  34. Moosavi, S.K.R.; Saadat, A.; Abaid, Z.; Ni, W.; Li, K.; Guizani, M. Feature selection based on dataset variance optimization using hybrid sine cosine–firehawk algorithm (hscfha). Future Gener. Comput. Syst. 2024, 155, 272–286. [Google Scholar] [CrossRef]
  35. Kwakye, B.D.; Li, Y.; Mohamed, H.H.; Baidoo, E.; Asenso, T.Q. Particle guided metaheuristic algorithm for global optimization and feature selection problems. Expert Syst. Appl. 2024, 248, 123362. [Google Scholar] [CrossRef]
  36. Abdelhamid, A.A.; El-Kenawy, E.S.M.; Ibrahim, A.; Eid, M.M.; Khafaga, D.S.; Alhussan, A.A.; Mirjalili, S.; Khodadadi, N.; Lim, W.H.; Shams, M.Y. Innovative feature selection method based on hybrid sine cosine and dipper throated optimization algorithms. IEEE Access 2023, 11, 79750–79776. [Google Scholar] [CrossRef]
  37. Ragab, M. Hybrid firefly particle swarm optimisation algorithm for feature selection problems. Expert Syst. 2024, 41, e13363. [Google Scholar] [CrossRef]
  38. Alkanhel, R.; El-kenawy, E.S.M.; Abdelhamid, A.A.; Ibrahim, A.; Alohali, M.A.; Abotaleb, M.; Khafaga, D.S. Network intrusion detection based on feature selection and hybrid metaheuristic optimization. Comput. Mater. Contin. 2023, 74, 2677–2693. [Google Scholar] [CrossRef]
  39. Mirjalili, S. Moth-flame optimization algorithm: A novel nature-inspired heuristic paradigm. Knowl.-Based Syst. 2015, 89, 228–249. [Google Scholar] [CrossRef]
  40. Mirjalili, S.; Gandomi, A.H.; Mirjalili, S.Z.; Saremi, S.; Faris, H.; Mirjalili, S.M. Salp swarm algorithm: A bio-inspired optimizer for engineering design problems. Adv. Eng. Softw. 2017, 114, 163–191. [Google Scholar] [CrossRef]
  41. Fakhouri, H.N.; Awaysheh, F.M.; Alawadi, S.; Alkhalaileh, M.; Hamad, F. Four vector intelligent metaheuristic for data optimization. Computing 2024, 106, 2321–2359. [Google Scholar] [CrossRef]
  42. Abualigah, L.; Diabat, A.; Mirjalili, S.; Abd Elaziz, M.; Gandomi, A.H. The arithmetic optimization algorithm. Comput. Methods Appl. Mech. Eng. 2021, 376, 113609. [Google Scholar] [CrossRef]
  43. Mirjalili, S. Sca: A sine cosine algorithm for solving optimization problems. Knowl. Based Syst. 2016, 96, 120–133. [Google Scholar] [CrossRef]
  44. Kennedy, J.; Eberhart, R. Particle swarm optimization. In Proceedings of the ICNN’95-International Conference on Neural Networks, Perth, Australia, 27 November–1 December 1995; IEEE: Piscataway, NJ, USA, 1995; pp. 1942–1948. [Google Scholar]
  45. Zhao, S.; Zhang, T.; Ma, S.; Chen, M. Dandelion optimizer: A nature-inspired metaheuristic algorithm for engineering applications. Eng. Appl. Artif. Intell. 2022, 114, 105075. [Google Scholar] [CrossRef]
  46. Shadravan, S.; Naji, H.R.; Bardsiri, V.K. The sailfish optimizer: A novel nature-inspired metaheuristic algorithm for solving constrained engineering optimization problems. Eng. Appl. Artif. Intell. 2019, 80, 20–34. [Google Scholar] [CrossRef]
Figure 1. Flow chart of MSCSO optimization algorithm.
Figure 1. Flow chart of MSCSO optimization algorithm.
Biomimetics 09 00701 g001
Figure 2. Comparison of convergence curves of MSCSO and other optimization algorithms for global optimization.
Figure 2. Comparison of convergence curves of MSCSO and other optimization algorithms for global optimization.
Biomimetics 09 00701 g002aBiomimetics 09 00701 g002b
Figure 3. Average fitness across 15 datasets.
Figure 3. Average fitness across 15 datasets.
Biomimetics 09 00701 g003
Figure 4. Average number of features across 15 datasets.
Figure 4. Average number of features across 15 datasets.
Biomimetics 09 00701 g004
Figure 5. Average accuracy across 15 datasets.
Figure 5. Average accuracy across 15 datasets.
Biomimetics 09 00701 g005
Figure 6. Comparison of convergence curves of MSCSO feature selection with other optimization algorithms.
Figure 6. Comparison of convergence curves of MSCSO feature selection with other optimization algorithms.
Biomimetics 09 00701 g006
Table 1. Parameter information of each algorithm.
Table 1. Parameter information of each algorithm.
AlgorithmParameter
AllGlobal optimization parameters: Population size = 30, Maximum iterations = 1000, Run count = 30
Feature selection parameters: Population size = 30, Maximum iterations = 100, Run count = 20
MSCSOk = 0.75, p = [1:360], S = 2
SCSOp = [1:360], S = 2
MFOb = 1
SSA-
FVIMalpha = 1.5
AOAMOP_Max = 1, MOP_Min = 0.2, Alpha = 5, Mu = 0.499
SCAa = 2
PSOw = 0.85, c1 = 1.2, c2 = 1.2
DO-
SFOPD = 2/3
Table 2. Comparison of mean fitness and standard deviation between MSCSO and existing optimization algorithms in CEC2005 test function (Bold indicates better data performance).
Table 2. Comparison of mean fitness and standard deviation between MSCSO and existing optimization algorithms in CEC2005 test function (Bold indicates better data performance).
FunctionMeasuresMSCSOSCSOMFOSSAFVIMAOASCAPSODOSFO
F1Mean01.6732 × 10−226.68351.2217 × 10−87.8252 × 10−551.0115 × 10−185.011 × 10−28.8258 × 10−71.2027 × 10−85.9286 × 10−19
Std002.53662.8644 × 10−92.3666 × 10−545.5403 × 10−181.2384 × 10−11.2243 × 10−68.8765 × 10−91.2626 × 10−18
F2Mean01.2274 × 10−113.33338.7898 × 10−16.0333 × 10−3208.8029 × 10−53.4267 × 10−17.9067 × 10−52.9971 × 10−9
Std06.5811 × 10−112.02288.4123 × 10−17.6058 × 10−3204.1444 × 10−41.82443.3934 × 10−52.3602 × 10−9
F3Mean07.1882 × 10−191.47052.63452.1786 × 10−62.6499 × 10−33.3952 × 1032.56771.02501.2079 × 10−16
Std001.08372.14617.2399 × 10−67.6668 × 10−33.4423 × 1031.11241.20482.1024 × 10−16
F4Mean09.6592 × 10−106.84308.05247.1856 × 10−21.6863 × 10−21.9320 × 1013.80519.6583 × 10−22.0926 × 10−10
Std04.8151 × 10−108.54603.53303.1181 × 10−22.1198 × 10−21.3087 × 1011.39455.1918 × 10−21.8565 × 10−10
F5Mean1.33182.8212 × 1016.88121.80623.0309 × 1012.8215 × 1015.1516 × 1029.48872.8709 × 1012.8706 × 101
Std1.28556.9124 × 10−12.26513.60871.1645 × 1014.3180 × 10−11.6177 × 1031.13851.6104 × 1011.4421 × 10−3
F6Mean1.671.822.33671.2538 × 10−83.58372.76934.74849.4045 × 10−79.1274 × 10−73.1592
Std6.32226.1323 × 10−16.80342.4312 × 10−98.3087 × 10−12.4581 × 10−18.6016 × 10−11.0517 × 10−64.1544 × 10−72.2552
F7Mean6.15299.1979 × 10−54.01131.0180 × 10−17.4159 × 10−33.9189 × 10−52.9036 × 10−22.4734 × 10−29.3052 × 10−38.1435 × 10−5
Std6.11661.1144 × 10−48.62513.7415 × 10−23.1360 × 10−34.5493 × 10−52.2559 × 10−27.8311 × 10−34.9979 × 10−36.9022 × 10−5
F8Mean−1.203−6.9303−8.9559−7.3619−5.3718−5.8265−3.8623−7.6221−8.2970−3.9342
Std1.16798.6264 × 1029.32536.92461.1014 × 1033.7253 × 1022.8576 × 1027.33276.2155 × 1021.1031 × 103
F9Mean001.77635.80063.5712 × 10101.6014 × 1014.84571.5862 × 1010
Std004.31032.26659.330401.9496 × 1011.38251.1639 × 1010
F10Mean4.4409 × 10−14.4409 × 10−11.24522.08408.1416 × 10−154.4409 × 10−11.4426 × 1014.5028 × 10−12.4648 × 10−55.2770 × 10−10
Std008.58618.4850 × 10−11.8853 × 10−1508.14416.5171 × 10−11.1472 × 10−55.3965 × 10−10
F11Mean001.51581.1401 × 10−25.5451 × 10−39.9232 × 10−22.7640 × 10−11.9719 × 10−21.5936 × 10−20
Std003.44521.1605 × 10−28.8110 × 10−38.2692 × 10−22.8250 × 10−12.5921 × 10−21.6144 × 10−20
F12Mean2.60457.3425 × 10−21.53874.8849 × 1018.1274 × 10−14.0717 × 10−16.5974 × 1028.5519 × 10−26.9630 × 10−82.7687 × 10−1
Std1.36742.9459 × 10−24.66683.08816.0639 × 10−14.6520 × 10−23.5664 × 1032.0937 × 10−13.6371 × 10−83.8328 × 10−1
F13Mean4.78952.43951.36691.68892.28872.78194.5932 × 1026.2330 × 10−31.0343 × 10−61.0019 × 10−2
Std1.04454.2140 × 10−17.48676.22363.9923 × 10−11.2718 × 10−12.4504 × 1031.1425 × 10−23.7267 × 10−73.2748 × 10−2
F14Mean1.03116.08422.34839.98 × 10−19.18411.0218 × 1011.46189.98 × 10−19.98 × 10−18.1515
Std1.81484.50541.92512.6562 × 10−164.65763.14878.5309 × 10−14.1233 × 10−174.9736 × 10−164.8078
F15Mean3.19874.0878 × 10−41.7980 × 10−38.3495 × 10−45.0951 × 10−32.2018 × 10−28.5939 × 10−43.6175 × 10−31.1016 × 10−31.8113 × 10−2
Std3.92312.7855 × 10−43.5397 × 10−32.8094 × 10−48.6189 × 10−33.5493 × 10−23.5571 × 10−41.1204 × 10−23.6516 × 10−33.0145 × 10−2
F16Mean−1.031628−1.0316284−1.0316284−1.0316284−1.0316282−1.0316283−1.0316002−1.0316284−1.0316284−1.0316282
Std4.0375 × 10−12.4429 × 10−106.7752 × 10−167.2408 × 10−158.8799 × 10−71.0438 × 10−72.7392 × 10−56.7752 × 10−168.2357 × 10−141.0214 × 10−6
F17Mean0.39788730.397887370.397887360.397887360.397887550.405608830.398929790.397887360.397887360.39791132
Std5.17702.0828 × 10−805.2723 × 10−154.3609 × 10−77.5248 × 10−39.4806 × 10−403.4351 × 10−121.0370 × 10−4
F18Mean3.00000073.00000123333.0000331010.96526483.0000393133.0000000012.0000119
Std1.52521.8880 × 10−61.3297 × 10−151.3386 × 10−135.1040 × 10−51.2387 × 1018.7638 × 10−58.8049 × 10−162.2231 × 10−92.4901 × 101
F19Mean3.8627805−3.8616140−3.8627821−3.8627821−3.8627257−3.8536551−3.8550555−3.8627821−3.8627821−3.8035099
Std2.3680 × 10−62.7517 × 10−32.7101 × 10−155.2138 × 10−146.4765 × 10−52.6938 × 10−31.8405 × 10−32.6823 × 10−152.3382 × 10−81.9563 × 10−1
F20Mean−3.322−3.2477−3.1982−3.2238−3.2821−3.0976−2.8989−3.2679−3.2744−2.9596
Std2.1588 × 10−61.1059 × 10−15.7419 × 10−25.0081 × 10−25.7438 × 10−27.5351 × 10−23.0548 × 10−16.9818 × 10−25.9241 × 10−21.9696 × 10−1
F21Mean−9.983−5.0860−6.3886−7.3946−7.1978−3.7087−2.2721−5.7286−6.0454−5.7652
Std9.3075 × 10−11.22303.46393.30913.20431.12441.96823.32433.32252.8908
F22Mean−9.694−6.8608−8.7772−9.5422−9.5982−4.2464−3.1372−6.5563−7.3459−5.8515
Std1.83772.54753.05212.27472.27761.62291.72113.50193.65722.9746
F23Mean−9.996−6.4471−7.4247−8.2498−9.7547−4.2953−3.8125−7.6379−6.7571−6.0959
Std1.65013.18673.66013.58212.38551.13881.93843.67883.72163.1840
Avg. rank1.573.616.224.524.485.657.704.653.615.39
Table 3. Comparison of Wilcoxon test p-values between MSCSO and existing optimization algorithms (Underline indicates no significant difference between MSCSO and the algorithm).
Table 3. Comparison of Wilcoxon test p-values between MSCSO and existing optimization algorithms (Underline indicates no significant difference between MSCSO and the algorithm).
FunctionSCSOMFOSSAFVIMAOASCAPSODOSFO
F11.21 × 10−121.21 × 10−121.21 × 10−124.57 × 10−121.21 × 10−121.21 × 10−121.21 × 10−121.21 × 10−121.21 × 10−12
F21.21 × 10−121.21 × 10−121.21 × 10−121.001.21 × 10−121.21 × 10−121.21 × 10−121.21 × 10−121.21 × 10−12
F31.21 × 10−121.21 × 10−121.21 × 10−121.21 × 10−121.21 × 10−121.21 × 10−121.21 × 10−121.21 × 10−121.21 × 10−12
F41.21 × 10−121.21 × 10−121.21 × 10−121.21 × 10−121.21 × 10−121.21 × 10−121.21 × 10−121.21 × 10−121.21 × 10−12
F51.19 × 10−63.02 × 10−112.57 × 10−71.29 × 10−64.08 × 10−55.49 × 10−111.43 × 10−50.0748271.07 × 10−7
F63.02 × 10−115.11 × 10−13.02 × 10−113.02 × 10−113.02 × 10−113.02 × 10−113.02 × 10−113.02 × 10−116.70 × 10−11
F72.84 × 10−13.02 × 10−113.02 × 10−113.02 × 10−113.92 × 10−23.02 × 10−113.02 × 10−113.02 × 10−110.22257
F83.02 × 10−113.82 × 10−93.02 × 10−113.02 × 10−113.02 × 10−113.02 × 10−116.70 × 10−118.99 × 10−113.02 × 10−11
F91.001.21 × 10−121.21 × 10−121.21 × 10−121.001.21 × 10−121.21 × 10−121.21 × 10−121.00
F101.001.21 × 10−121.21 × 10−126.13 × 10−141.001.21 × 10−121.21 × 10−121.21 × 10−121.21 × 10−12
F111.001.21 × 10−121.21 × 10−121.37 × 10−31.21 × 10−121.21 × 10−121.21 × 10−121.21 × 10−121.00
F123.02 × 10−116.07 × 10−113.02 × 10−113.02 × 10−113.02 × 10−113.02 × 10−110.610013.02 × 10−118.15 × 10−11
F133.02 × 10−112.32 × 10−69.94 × 10−13.02 × 10−113.02 × 10−113.02 × 10−110.00728843.02 × 10−110.0076171
F148.84 × 10−79.47 × 10−11.41 × 10−114.07 × 10−113.02 × 10−112.87 × 10−101.72 × 10−121.44 × 10−104.97 × 10−11
F152.32 × 10−23.01 × 10−118.99 × 10−118.15 × 10−51.33 × 10−106.70 × 10−110.520140.129671.61 × 10−10
F162.17 × 10−11.21 × 10−122.98 × 10−116.01 × 10−83.02 × 10−113.02 × 10−111.21 × 10−121.33 × 10−102.23 × 10−9
F178.53 × 10−11.21 × 10−122.70 × 10−118.84 × 10−73.02 × 10−113.02 × 10−111.21 × 10−123.02 × 10−119.53 × 10−7
F182.81 × 10−21.27 × 10−113.02 × 10−112.37 × 10−104.12 × 10−11.16 × 10−77.87 × 10−121.17 × 10−90.021506
F192.61 × 10−21.21 × 10−123.01 × 10−111.25 × 10−73.02 × 10−113.02 × 10−112.36 × 10−127.77 × 10−91.73 × 10−7
F209.12 × 10−19.92 × 10−76.77 × 10−52.25 × 10−43.02 × 10−113.02 × 10−110.179020.185773.02 × 10−11
F211.78 × 10−104.27 × 10−13.33 × 10−13.82 × 10−94.08 × 10−113.69 × 10−110.0346250.0992587.39 × 10−11
F226.97 × 10−32.00 × 10−46.05 × 10−75.87 × 10−43.16 × 10−109.92 × 10−110.331540.264339.76 × 10−10
F236.10 × 10−33.52 × 10−16.97 × 10−39.21 × 10−52.37 × 10−109.92 × 10−110.143950.830266.12 × 10−10
Table 4. Feature selection dataset details.
Table 4. Feature selection dataset details.
DatasetFeature CountSample CountClasses
Zoo161017
Wine131783
Vote163002
Lymphography181484
HeartEW132702
Sonar602082
SpectEW222672
Lung-Cancer56323
BreastEW305682
CongressEW164342
Clean11664762
Exactly1310002
Exactly21310002
M-of-n1310002
VP1286693
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liu, R.; Fang, R.; Zeng, T.; Fei, H.; Qi, Q.; Zuo, P.; Xu, L.; Liu, W. A Novel Adaptive Sand Cat Swarm Optimization Algorithm for Feature Selection and Global Optimization. Biomimetics 2024, 9, 701. https://doi.org/10.3390/biomimetics9110701

AMA Style

Liu R, Fang R, Zeng T, Fei H, Qi Q, Zuo P, Xu L, Liu W. A Novel Adaptive Sand Cat Swarm Optimization Algorithm for Feature Selection and Global Optimization. Biomimetics. 2024; 9(11):701. https://doi.org/10.3390/biomimetics9110701

Chicago/Turabian Style

Liu, Ruru, Rencheng Fang, Tao Zeng, Hongmei Fei, Quan Qi, Pengxiang Zuo, Liping Xu, and Wei Liu. 2024. "A Novel Adaptive Sand Cat Swarm Optimization Algorithm for Feature Selection and Global Optimization" Biomimetics 9, no. 11: 701. https://doi.org/10.3390/biomimetics9110701

APA Style

Liu, R., Fang, R., Zeng, T., Fei, H., Qi, Q., Zuo, P., Xu, L., & Liu, W. (2024). A Novel Adaptive Sand Cat Swarm Optimization Algorithm for Feature Selection and Global Optimization. Biomimetics, 9(11), 701. https://doi.org/10.3390/biomimetics9110701

Article Metrics

Back to TopTop