Next Article in Journal
Enabling High-Resolution Micro-Vibration Detection Using Ground-Based Synthetic Aperture Radar: A Case Study for Pipeline Monitoring
Previous Article in Journal
Clutter and Interference Cancellation in River Surface Velocity Measurement with a Coherent S-Band Radar
Previous Article in Special Issue
Experimental Tests and Simulations on Correction Models for the Rolling Shutter Effect in UAV Photogrammetry
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Hybrid Binary Dragonfly Algorithm with an Adaptive Directed Differential Operator for Feature Selection

1
School of Computer Science and Engineering, Wuhan Institute of Technology, Wuhan 430073, China
2
Hubei Key Laboratory of Intelligent Robot (Wuhan Institute of Technology), Wuhan 430073, China
3
Hubei Engineering Research Center of Intelligent Production Line Equipment (Wuhan Institute of Technology), Wuhan 430073, China
4
School of Computer Science, China University of Geosciences, Wuhan 430073, China
5
Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518052, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(16), 3980; https://doi.org/10.3390/rs15163980
Submission received: 14 June 2023 / Revised: 28 July 2023 / Accepted: 9 August 2023 / Published: 11 August 2023
(This article belongs to the Special Issue New Advancements in Remote Sensing Image Processing)

Abstract

:
Feature selection is a typical multiobjective problem including two conflicting objectives. In classification, feature selection aims to improve or maintain classification accuracy while reducing the number of selected features. In practical applications, feature selection is one of the most important tasks in remote sensing image classification. In recent years, many metaheuristic algorithms have attempted to explore feature selection, such as the dragonfly algorithm (DA). Dragonfly algorithms have a powerful search capability that achieves good results, but there are still some shortcomings, specifically that the algorithm’s ability to explore will be weakened in the late phase, the diversity of the populations is not sufficient, and the convergence speed is slow. To overcome these shortcomings, we propose an improved dragonfly algorithm combined with a directed differential operator, called BDA-DDO. First, to enhance the exploration capability of DA in the later stages, we present an adaptive step-updating mechanism where the dragonfly step size decreases with iteration. Second, to speed up the convergence of the DA algorithm, we designed a new differential operator. We constructed a directed differential operator that can provide a promising direction for the search, then sped up the convergence. Third, we also designed an adaptive paradigm to update the directed differential operator to improve the diversity of the populations. The proposed method was tested on 14 mainstream public UCI datasets. The experimental results were compared with seven representative feature selection methods, including the DA variant algorithms, and the results show that the proposed algorithm outperformed the other representative and state-of-the-art DA variant algorithms in terms of both convergence speed and solution quality.

1. Introduction

With the advances in data mining and machine learning techniques, many research problems now involve the analysis of multiple datasets. These datasets frequently contain numerous irrelevant, noisy, and/or redundant features, which significantly affect the classification performance. Specifically, irrelevant or noisy features can significantly degrade the classification performance due to their misleading information [1]. Feature selection (FS) aims to enhance classification performance, reduce data dimensionality, save storage space, improve computational efficiency, and facilitate data visualization and understanding by selecting a small subset of relevant features. The increase in data and dimensionality leads to an increasingly difficult feature selection [2].
In recent years, feature selection has found widespread application in various research domains, including text classification, remote sensing [3], intrusion detection, gene analysis, and image retrieval [4]. Notably, feature selection plays a crucial role in remote sensing image classification tasks. Remote sensing images often contain a vast amount of pixel data and diverse spectral information [5,6], yet only a subset of features contributes significantly to the classification process. Hence, feature selection techniques are used to identify the most informative and relevant features, reduce the complexity of the classification tasks, and improve the overall accuracy. By effectively selecting pertinent features, the interference of redundant information and noise can be minimized, while emphasizing the spatial, spectral, and textural characteristics of objects in remote sensing images. Consequently, feature selection holds paramount significance in achieving precise land object classification and monitoring in remote sensing applications [7,8]. To this end, the development of efficient feature selection methods becomes imperative, aiming to optimize classification accuracy and computational efficiency.
Feature selection methods are mainly divided into two categories: filters and wrappers [9]. Filter-based methods use the intrinsic information of features, such as correlation, information gain, and consistency [10], to determine features [11]. The wrapper-based approach is used to generate feature subsets through a classifier and a learning model and perform accuracy evaluation by a learning algorithm to determine the feature subsets [12,13]. Although wrapper-based methods can produce better feature subsets, they are computationally expensive, especially when the search space is large and complex. In contrast, filter-based methods are usually less computationally expensive, but have lower classification accuracy and cannot select a good subset of features [14].
Although feature selection is an indispensable process, it is also a complex and challenging combinatorial optimization problem. The challenges of feature selection mainly revolve around three aspects. Firstly, the search space for feature selection exponentially grows with the number of features, as N features can yield  2 N  feature subsets [15,16]. Therefore, an exhaustive search of all possible subsets is impractical, especially when dealing with numerous features. Secondly, features have complex interactions with each other. Thirdly, feature selection is inherently a multiobjective problem. The two primary objectives of feature selection are to maximize classification performance and to minimize the number of selected features. However, these objectives are often conflicting [17]. A robust and powerful search algorithm forms the foundation for addressing feature selection problems effectively.
In recent years, metaheuristics have been successfully applied to feature selection methods. Metaheuristic algorithms are inspired by natural phenomena, such as particle swarm optimization (PSO) [18,19,20], genetic algorithms (GA) [21], the differential evolution algorithm (DE) [22,23], the bat algorithm (BA) [24,25], the gray wolf optimization algorithm (GWO) [26,27,28], the dragonfly algorithm (DA) [29], cuckoo search (CS) [30], the salp swarm algorithm (SSA) [31], Harris hawks optimization (HHO) [32], etc. When using or designing metaheuristic algorithms, there needs to be an effective way to maintain a balance between development and exploration. Metaheuristic algorithms need to search as much of the search space as possible in the early stages and develop the optimal region in the later stages, so a balance between exploration and development is essential.
The dragonfly algorithm (DA) is a recently developed optimization algorithm inspired by the collective behavior of dragonflies in nature. Initially designed for continuous optimization tasks, a binary version called the binary dragonfly algorithm (BDA) was later introduced to address discrete problems [33]. While BDA has shown strong performance on various datasets, it may suffer from limited exploration capabilities, potentially leading to local optimization problems.
In this study, we propose a hybrid method that combines the BDA algorithm with the directed differential operator to improve the performance of BDA. An adaptive step-updating mechanism is also proposed to enhance the exploration capability of the algorithm which improves the search performance. The proposed method has been tested on 14 mainstream datasets and compared with seven representative feature selection methods, including the DA variant algorithms. The main contributions of this study are as follows:
  • The algorithm incorporates an adaptive step-updating mechanism to optimize the search process. It adjusts the step size based on the stage of exploration: a longer step size benefits an early search, while a smaller step size is advantageous for later exploration. To achieve this, the dragonfly’s step size gradually decreases with each iteration, adapting to the evolving search landscape. This adaptive mechanism enhances the algorithm’s ability to balance exploration and exploitation, resulting in improved overall performance.
  • Based on the renowned differential evolution algorithm, we propose a directed differential operator. This operator utilizes information from both the best and worst positions of the dragonflies to guide the algorithm towards promising regions, facilitating more effective search and convergence to optimal solutions. Additionally, the size of the directed differential operator influences the balance between targeted exploitation and broad exploration. A larger operator emphasizes guidance, while a smaller one allows for greater exploration of the solution space. To achieve this balance, we design an adaptive updating mechanism to adjust the directed differential operator, playing a crucial role in optimizing exploration and exploitation within the search space.
  • We enhance BDA’s exploration capability by integrating it with the differential evolution algorithm (DE). During the position-updating phase of the dragonfly algorithm, individual positions are combined with the directed differential operator to guide the search in promising directions, accelerating convergence. To maintain population diversity, we introduce an adaptive method for updating the directed differential operator. Additionally, an adaptive update mechanism for the dragonfly step dynamically adjusts the step size to improve exploration in later stages. This integration significantly improves BDA’s performance and exploration capabilities.
The paper’s structure is as follows: In Section 2, related works are discussed, focusing on recent feature selection algorithms and the binary dragonfly algorithm (BDA). Section 3 provides a detailed explanation of the proposed hybrid BDA-DDO algorithm, including the three improvement mechanisms and the algorithm’s specific process. Section 4 describes the experimental setup and analysis of the obtained results. Finally, Section 5 presents the conclusion and proposes future research directions.

2. Related Work

In this section, an extensive review of feature selection algorithms is presented, with a specific emphasis on the binary dragonfly algorithm (BDA). We provide a comprehensive overview of existing approaches to feature selection, highlighting their methodologies and performance. Furthermore, we delve into the intricacies of BDA, elucidating its principles and underlying mechanisms.

2.1. Related Work on Feature Selection

Metaheuristic algorithms have gained popularity in the past decade due to their advantages, and they have been widely applied to feature selection and optimization problems. As a result, several feature selection algorithms based on metaheuristics have been proposed. This section presents a comprehensive review of some noteworthy feature selection algorithms that have emerged in recent years [14].
In [34], Xue et al. proposed a feature selection approach that combines multiobjective particle swarm optimization (PSO) with NSGA-II. The algorithm efficiently explores the solution space by integrating the nondominated sorting mechanism into PSO, resulting in a set of nondominated solutions rather than a single one. However, a limitation of the method is the rapid loss of diversity in the later stages, which affects its feature selection performance. The particle-ranking-based PSO method [35] divides the objective space into several subregions using uniform and nonuniform partitioning techniques. It calculates particle rank and feature rank to update particle velocity and position in each generation. This approach accelerates optimization and guides the particles to better solutions. Luo et al. [36] proposed a memetic algorithm based on particle swarm optimization to tackle high-dimensional feature selection. The algorithm includes information-entropy-based initialization and adaptive local search techniques to enhance search efficiency. Additionally, a novel velocity-updating mechanism is introduced to consider solution convergence and diversity. The algorithm demonstrates strong performance in improving feature subset quality and size. The algorithm’s performance on low-dimensional datasets may be limited, necessitating further enhancement of its robustness.
Xue et al. [37] introduced a hybrid algorithm that combines PSO and DE for feature selection. DE is used to generate promising individuals to guide the search process. This approach maintains diversity and directs particles to promising regions, enabling effective escape from local optima. The algorithm performs well on small-scale datasets, but its performance on large-scale datasets remains unverified. Nelson et al. [22] introduced a hybrid DE algorithm that combines binary differential evolution (BDE) for obtaining feature subsets and incorporates a local search method to minimize classification error. The algorithm’s essence lies in using a classical feature selection algorithm as the local search strategy to enhance the solutions derived from DE. In this way, the hybrid method aims to strike a balance between exploration and exploitation in achieving improved solutions. In their work [38], Wang et al. proposed SaWDE, a novel weighted differential evolution algorithm that addresses large-scale feature selection problems. SaWDE employs a multipopulation mechanism for enhanced diversity and introduces an adaptive strategy to capture diverse features from historical dataset information. Additionally, a weighted model is used to identify important features, leading to effective feature selection solutions. The algorithm performs remarkably well in reducing feature dimensions and demonstrates strong performance on large-scale datasets. However, further research is needed to fine-tune the number of subpopulations and relevant parameters in the algorithm. Xue et al. proposed a multiobjective differential evolution approach in [39] to search for multiple optimal feature subsets. They used a method that considers feature correlations for initialization to provide a strong starting point. The population was divided into several subarchives using a clustering approach, and within each subarchive, a sophisticated crowding distance was utilized to ensure diversity by considering both the search and objective spaces. Nondominated solutions from all subarchives were stored in another archive to guide the evolutionary feature selection process. The proposed algorithm achieved improved feature subsets, but the training process may take longer due to the increased computational time.
The dragonfly algorithm (DA) is a recent metaheuristic algorithm introduced by Mirjalili et al. In [33], they proposed a wrapper feature selection algorithm based on the binary dragonfly algorithm (BDA). This algorithm uses a classical transfer function to convert the continuous search space into a discrete space, which effectively solves feature selection problems and outperforms particle swarm optimization (PSO) and genetic algorithms (GA). In their subsequent work [40], Mirjalili et al. introduced three different update mechanisms to balance exploration and exploitation in the dragonfly algorithm. The updated algorithm showed improved performance compared to the binary dragonfly algorithm. The reduced exploration capability in the later stages of the algorithm is still a limitation, leading to the possibility of encountering local optimizations. Too et al. [41] proposed the hyper-learning binary dragonfly algorithm (HLBDA) to mitigate the binary dragonfly algorithm’s susceptibility to local optima. The algorithm employs a hyper-learning strategy to improve search behavior and avoid getting trapped in local optimization. HLBDA was applied to the COVID-19 dataset, and the results show its superior performance in enhancing prediction accuracy. Hamouda et al. [42] proposed a hybrid feature selection algorithm that combines the dragonfly algorithm (DA) with simulated annealing to address the issue of local optimization in DA. By integrating simulated annealing, the algorithm aims to improve its ability to select the optimal feature subset for classification tasks and mitigate the problem of local optima. The approach shows promising results in enhancing classification accuracy. Duan et al. [43] proposed hybrid DA-DE, a novel algorithm that combines the dragonfly algorithm (DA) and differential evolution (DE) to address global optimization problems. The algorithm introduces a mutation operator from DA and adapts the scaling factor in an individual-dependent and self-adaptive manner. By integrating DE’s development capability and DA’s exploration ability, the algorithm achieves optimal global solutions effectively, especially on high-dimensional functions.
In response to the limitations of the dragonfly algorithm (DA) and based on previous research progress, we propose a hybrid approach by combining the binary dragonfly algorithm with the directed differential operator. Our goal is to enhance the algorithm’s search capability and improve its performance in feature selection problems.

2.2. Binary Dragonfly Algorithm

The Dragonfly algorithm (DA) is a swarm intelligence optimization algorithm proposed by Seyedali Mirjalili [29] in 2015, inspired by the swarm intelligent behavior of dragonflies. It was observed that dragonfly swarming behavior comprises two main patterns: hunting and migration [29,42]. The hunting mechanism involves dragonflies swarming in search of food sources, while the migration mechanism involves large groups of dragonflies migrating over long distances in a specific direction. In 2016, Mirjalili introduced the binary dragonfly algorithm (BDA), which is a discrete version based on the original DA and adapted for binary optimization problems.
The dragonfly algorithm employs five primary behaviors, which is essential for updating the dragonfly positions. Each of these behaviors is described as follows:
  • Separation, in order to avoid static collisions between individuals and other nearby individuals. The mathematical model of separation behavior is represented by Equation (1).
    S i = j = 1 N X X j
    In the presented model, X represents the current search agent, and  X j  denotes its j-th neighbor. The parameter N indicates the total number of neighbors.
  • Alignment, which represents the speed match between the current individual and other surrounding individuals. The mathematical model can be represented by Equation (2).
    A i = j = 1 N V j N
    where  V j  denotes the velocity of the j-th neighbor.
  • Cohesion, which implies that the current individual tends to move closer to the mass center of the surrounding group. Mathematically, this cohesion behavior is expressed by Equation (3).
    C i = j = 1 N x j N X
  • Attraction, which entails the attraction of a food source to an individual and its movement towards the food source. Mathematically, it is expressed by Equation (4).
    F i = F l o c a t i o n X
    where  F l o c a t i o n  represents the location of the food source.
  • Distraction, which means the individual moves outward away from the enemy’s position. The behavior of the i-th individual principle enemy is represented by Equation (5).
    E i = E l o c a t i o n + X
    where  E l o c a t i o n  represents the location of the current enemy.
The dragonfly algorithm follows a framework similar to that of particle swarm optimization (PSO). The position update process primarily involves two vectors: the step vector ( Δ X ) and the position vector (X). The step size vector serves a role similar to the speed vector in PSO, determining both the movement direction and step size of the dragonfly. The step vector is defined as follows:
Δ X t + 1 = ( s S i + a A i + c C i + f F i + e E i ) + ω Δ X t
where s is the separation weight, a is the alignment weight, c is the cohesion weight, f is the food weight, e is the enemy weight,  ω  is the inertia weight, and t is the current number of iterations. Equation (7) demonstrates adaptively adjusting the value of  ω  to balance exploration and exploitation. Hammouri et al. [40] proposed three adaptive methods to adjust these parameters, aiming to enhance BDA’s performance. Further explanation of these adaptive methods will be provided below.
ω = 0.9 I t e r ( 0.9 0.4 ) M a x _ i t e r
The position of an individual is updated as in Equation (8):
X t + 1 = X t + Δ X t + 1
In LBDA, a linear model is utilized to update parameter values within specific ranges, as shown in Equation (9). The values of s, a, and c are linearly updated in the range of 0 to 0.2. Meanwhile, the value of e is updated within the range of 0.1 to 0, and the value of f is updated in the range of 0 to 2.
s = 0.2 ( 0.2 × i t e r / m a x _ i t e r ) e = 0.1 ( 0.1 × i t e r / m a x _ i t e r ) a = 0.0 + ( 0.2 × i t e r / m a x _ i t e r ) c = 0.0 + ( 0.2 × i t e r / m a x _ i t e r ) f = 0.0 + ( 2.0 × i t e r / m a x _ i t e r )
In QBDA, a quadratic model is utilized to update the value of each parameter in a nonlinear manner. The parameter value range remains consistent with that of LBDA.
s = ( 0.2 ( 0.2 × i t e r / m a x _ i t e r ) ) 2 e = ( 0.1 ( 0.1 × i t e r / m a x _ i t e r ) ) 2 a = ( 0.0 + ( 0.2 × i t e r / m a x _ i t e r ) ) 2 c = ( 0.0 + ( 0.2 × i t e r / m a x _ i t e r ) ) 2 f = ( 0.0 + ( 2.0 × i t e r / m a x _ i t e r ) ) 2
In SBDA, a sinusoidal model is used to update the values of the coefficients in a nonlinear manner. The parameter value range remains the same as that of LBDA.
s = 0.10 + ( 0.10 × cos ( i t e r / m a x _ i t e r × 4 × π β ) ) e = 0.05 + ( 0.05 × cos ( i t e r / m a x _ i t e r × 4 × π β ) ) a = 0.10 ( 0.05 × cos ( i t e r / m a x _ i t e r × 4 × π β ) ) c = 0.10 ( 0.05 × cos ( i t e r / m a x _ i t e r × 4 × π β ) ) f = 2.00 ( 1.00 × cos ( i t e r / m a x _ i t e r × 4 × π β ) )
Feature selection poses distinct challenges in binary and continuous search spaces due to its discrete nature. In the continuous search space, the position update is achieved by adding the updated step vector to the position vector. However, in the discrete search space, where the position vector can only take binary values (0 or 1), a transfer function is essential to convert the continuous space into the discrete space. BDA adopts a V-shaped transfer function, represented by the following equation, where  X i d ( t )  represents the d-dimensional position of the i-th dragonfly, and  Δ X  represents the step vector.
X i d ( t + 1 ) = 1 X i d ( t ) r a n d < T F ( Δ X i d ( t + 1 ) ) X i d ( t ) r a n d T F ( Δ X i d ( t + 1 ) )
T F ( Δ X ) = Δ X Δ X 2 + 1

3. The Proposed Feature Selection Method

In this section, we present the binary dragonfly algorithm with a hybrid directed differential operator (BDA-DDO) and its application to feature selection. While BDA exhibits powerful search capability, it faces challenges such as limited late exploration, slow convergence, and a lack of population diversity. To effectively address these issues, we propose three improvement mechanisms. In BDA-DDO, we introduce a directed differential operator that combines BDA-generated individuals with the differential operator, resulting in faster convergence. Additionally, we propose an adaptive approach to update this operator, thereby increasing population diversity. Furthermore, we design an adaptive step-updating mechanism to promote exploration in later stages. We provide a detailed explanation of these improvements and describe the fitness function used to evaluate the quality of the solution.

3.1. Directed Differential Operator

Differential evolution (DE), introduced by Storn and Price [44] in 1997, has gained significant attention from researchers due to its simple structure, fast convergence speed, and ease of implementation. In order to enhance the convergence speed of BDA, we have devised a new differential operator based on the DE algorithm. This novel differential operator combines the optimal and worst solutions of the dragonfly, forming a directed differential operator. By doing so, it provides a promising direction in the search space, guiding individuals towards better solutions and facilitating faster convergence to the optimal region. Below, we provide a step-by-step description of the implementation of the directed differential operator.
  • Mutation: In the proposed algorithm, mutation is applied to the ith dragonfly individual to generate a mutation vector. Storn and Price [45] introduced several mutation operations, with “DE/rand/1” being the most typical one:
    V i = X i + F × ( X F o o d X E n e m y )
    Among them,  X i  represents the position of the individual after updating through BDA,  X F o o d  represents the location of the dragonfly’s food (the best location in history), and  X E n e m y  represents the location of dragonfly enemies. The scaling factor, denoted by F, controls the amplification of the difference vector and significantly influences the convergence speed. We will provide a detailed explanation of F later in this section.
  • Crossover: After mutation, a crossover operation is performed by randomly selecting either the mutant individual  V i , j  or the original individual  X i , j  to generate the experimental individual. Among the three classical crossover operators [46]—binomial crossover, exponential crossover, and rotation-invariant arithmetic crossover—we are using the binomial crossover operator, as shown in the following equation:
    U i , j = V i , j r a n d < C R o r r a n d i ( 1 , d ) = j X i , j r a n d > C R o r r a n d i ( 1 , d ) j
    where rand is a random variable between [0, 1] and jrand is an integer randomly selected from [1, D], ensuring that at least one of  U i , j  comes from  V i , j . CR is the crossover probability, selected from [0, 1], which controls the population diversity.
    After the crossover operation, the resulting  U i , j  vector is transformed into a discrete space vector using the transfer function  T 1 . Then, a selection operation is carried out to determine whether it can survive to the next generation. The transfer function  T 1  is defined as follows:
    T 1 = 1 1 + e ( U i d )
    U i , j = U i , j = 0 r a n d < T 1 U i , j = 1 r a n d > T 1
  • Selection: After the mutation and crossover operations, selection is performed to determine the survival of individuals  U i  and  X i  in the next generation.
    X i = U i f ( U i ) f ( X i ) X i o t h e r s
    where  f ( U i )  and  f ( X i )  represent the fitness functions corresponding to  U i  and  X i , respectively.

3.2. Time-Varying Differential Vector

To address the issue of population diversity in the late stage of the BDA algorithm, we propose a time-varying differential vector approach. The value of the differential vector decreases from its initial value as iterations progress. The directed differential operator introduced earlier provides directional information. In the early stage, a larger differential vector is used to provide more valuable information for individual search. In contrast, a smaller differential vector is applied in the late stage to improve population diversity in the optimal region, there by improving the overall search performance. The time-varying differential vector is designed as shown in Equation (19), where F ranges from [1, 0.5] and gradually decreases with iterations.
F = 0.5 1 + ( 0.5 ) × e 0.5 × i t e r

3.3. Adaptive Step-Updating Mechanism

We propose an adaptive step-updating mechanism in the BDA algorithm to address the fixed step size issue. A fixed step size can cause oscillations if too large, or slow convergence if too small, leading to local optima. To overcome this, our adaptive approach dynamically adjusts the dragonfly step size during iterations. Equations (19) and (20) are used for this purpose. The BDA-DDO algorithm, which integrates the adaptive step updating mechanism and the directed differential operator, is given in Algorithm 1. This comprehensive approach aims to enhance exploration and improve population diversity in the late stages of optimization.
Δ X = F × Δ X
Algorithm 1 Pseudocode of the BDA-DDO
Input: The number of populations N, the maximum number of iterations
Output: The best solution
1:
Initialize population position  X i ( i = 1 , 2 , , N )
2:
Initialize the step vectors,  Δ X i ( i = 1 , 2 , , N )
3:
while Maximum number of iterations not reached do
4:
    Calculate the fitness value of all dragonfly individuals
5:
    Update the  F l o c a t i o n , and the  E l o c a t i o n
6:
    Update s, a, c, f, e, and  ω  (Using Equations (9) or (10) or (11), Using Equation (7))
7:
    for  i = 1  to Number of dragonflies, N do
8:
          Calculate the values of S, A, C, F, E (Using Equations (1)–(5))
9:
          Update step vector using Equation (6)
10:
        Update step vector using Equation (20)
11:
        Update the position of dragonfly (i-th) using Equations (12) and (13)
12:
    Individual position mutations using Equations (14) and (19)
13:
    Crossover operation using Equations (15)–(17)
14:
    Select operation using Equation (18)
15:
return the best solution

3.4. Fitness Evaluation

Feature selection is a multiobjective problem, where maximum classification accuracy and the minimum subset of features are the goals that need to be achieved. We balance these two objectives with a fitness function, by setting weighting factors. As shown in the following equation [33,40,41,42]:
f i t n e s s = α × E R R + β × R N
where  E R R  is the classification error rate obtained using the KNN classifier, R is the number of subsets of features selected by the search agent, and N is the total number of features in the datasets.  α  and  β  are weight factors for balancing classification accuracy and feature subsets. The value range of  α  is between [0, 1] and the value of  β  is (1- α ). To maximize the classification accuracy, we set  α  to 0.99 and  β  to 0.01 in our experiments. We use the parameter settings described in [40] and compare our results with those obtained using the BDA method.

4. Experiments and Results Analysis

In this section, we evaluate the performance of three proposed BDA-DDO algorithms, namely LBDA-DDO, QBDA-DDO, and SBDA-DDO, which are the improved versions based on BDA. Section 4.1 presents the details of the mainstream datasets. The parameter information of related algorithms is introduced in Section 4.2. Section 4.3 introduces the performance comparison analysis of the proposed algorithm and LBDA. Section 4.4 presents a comparison between the proposed algorithm and QBDA. Section 4.5 introduces the SBDA-DDO algorithm for comparison with SBDA. Section 4.6 presents a comparison of BDA-DDO with other binary versions of metaheuristic-based feature selection algorithms.

4.1. Datasets

In this section, the performance of the proposed algorithm was evaluated on 14 popular datasets collected from the UCI repository [47]. Table 1 provides detailed information about each of the datasets, which are commonly used by researchers to study feature selection methods. From Table 1, it can be observed that the datasets vary in terms of the number of instances, features, and dimensions. This demonstrates that the proposed algorithm’s performance has been tested on datasets with different structures. Each dataset in Table 1 was randomly divided into two sets: 80% for training and 20% for testing. To account for the algorithm’s randomness, we conducted 30 independent runs for each algorithm.

4.2. Parameter Settings

In this study, K-nearest neighbors (KNN) was utilized as the classification algorithm to evaluate the accuracy of the selected features. The datasets were divided into training and test sets using 10-fold cross-validation, and the classification error was calculated using K-nearest neighbors (KNN) with k = 5. This experiment was conducted on a WIN10 system with NVIDIA GTX 1660 graphics card, Inter Core i5-11400 processor, 2.6 GHz main frequency, and 16 GB RAM; the 2021a version of matlab was used. To ensure the fairness of the experiments, we used the same parameter settings and biomimetic environment as described in the original paper. The relevant parameters of the algorithm are set as shown in Table 2 below:

4.3. Comparison with LBDA Method

In this section, we obtain results on 14 mainstream datasets using the proposed FS method. The obtained results are compared with LBDA in three aspects: classification accuracy, number of selected features and fitness value. Table 3 shows the comparison between BDA-DDO and LBDA in terms of classification accuracy, and it should be noted that LBDA1 indicates that the proposed algorithm does not apply an adaptive step-updating mechanism. Table 4 shows the comparison of the proposed algorithm with LBDA in selecting the number of features. Table 5 shows that BDA-DDO compares with LBDA in terms of fitness value.
Table 3 shows the classification accuracy obtained by LBDA-DDO, LBDA1, and LBDA on the 14 datasets. The results indicate that LBDA1 achieves higher classification accuracy on most datasets compared to LBDA, while LBDA-DDO consistently outperforms LBDA in terms of classification accuracy on all 14 datasets. Furthermore, considering the standard deviations, LBDA-DDO demonstrates better robustness than LBDA, indicating its ability to handle variations in the datasets more effectively. Therefore, LBDA-DDO can achieve better results by utilizing an adaptive step size strategy, which enhances the algorithm’s search capability in the later stages and balances the trade-off between exploration and exploitation.
Table 4 illustrates the average number of selected features by the proposed LBDA-DDO algorithm on the datasets, with the best results highlighted in bold. The findings in Table 4 indicate that LBDA-DDO consistently selects fewer features than LBDA and LBDA1 across all 14 datasets. Furthermore, LBDA1 demonstrates a lower feature count compared to LBDA. These results demonstrate the ability of the proposed algorithm to effectively reduce the number of selected features in comparison to LBDA, thereby eliminating noisy or irrelevant features that may have been chosen by the LBDA method. This outcome aligns perfectly with our main objective.
Table 5 illustrates the average fitness results obtained by the proposed LBDA-DDO algorithm. As before, the best results are shown in bold. The results show that the fitness values obtained by LBDA-DDO on the nine datasets are lower than those of LBDA and LBDA1. On the two datasets Breastcancer and penglungEW, LBDA-DDO and LBDA1 exhibit the same fitness values. On Lymphography, LBDA1 achieves the best results. Overall, LBDA-DDO performs better than LBDA.
Figure 1 presents the convergence behavior of the proposed LBDA-DDO algorithm on 14 mainstream datasets, comparing it with LBDA and LBDA1. LBDA refers to the binary dragonfly algorithm with linear model updating, while LBDA1 represents the proposed algorithm(LBDA-DDO) without the adaptive step-updating mechanism. The X-axis shows the number of iterations, and the Y-axis displays the fitness value. The results show that the proposed algorithm achieves better convergence on most datasets. This indicates an enhanced search capability, especially with the inclusion of the adaptive step-updating mechanism. The proposed algorithm effectively avoids getting stuck in local optima.
Additionally, it can be observed that LBDA exhibits slower convergence on certain datasets, necessitating more iterations to attain satisfactory fitness values. Consequently, the proposed algorithm holds an advantage in terms of convergence behavior, enabling it to swiftly discover high-quality solutions. This highlights the efficacy of the introduced directional difference operator strategy in accelerating convergence speed, while the adaptive step-updating mechanism enhances search capabilities. In summary, the proposed algorithm outperforms LBDA by achieving faster convergence and yielding solutions with higher fitness values.

4.4. Comparison with QBDA Method

In this section, the performance of the proposed QBDA-DDO algorithm is examined, which combines the QBDA method with DE. QBDA1 is also included as an algorithm without the adaptive step size mechanism. For the performance of the algorithm, a comparison is conducted with QBDA in three aspects: classification accuracy, selected features, and fitness values. Table 6 presents the comparison of the classification accuracy between the proposed algorithm and QBDA. Table 7 shows the comparison of selected features between QBDA-DDO and QBDA. Additionally, Table 8 provides a comparison of fitness values between the QBDA-DDO algorithm and QBDA. Furthermore, Figure 2 shows the convergence speed of the three algorithms on the 14 mainstream datasets for visual analysis.
Table 6 displays the average classification accuracy achieved by the proposed QBDA-DDO algorithm. QBDA-DDO achieves the best results on seven datasets, while the results of the three algorithms are equal on three datasets. QBDA1 and BDA-DDO obtain the same results on Breastcancer and CongressEW. For BreastEW and SonarEW, QBDA1 performs the best. Therefore, the directional differential operator and adaptive step size, when applied to QBDA, can effectively improve the classification accuracy of feature selection, highlighting the effectiveness of the algorithm innovation.
Table 7 shows the average number of selected features obtained by the proposed BDA-DDO algorithm, with the best results highlighted in bold. The results demonstrate that the BDA-DDO algorithm achieves the best results on all 13 datasets, while QBDA1 outperforms QBDA on most datasets. This indicates that the directional differential operator mechanism assists QBDA in converging to favorable solutions, while the adaptive step size enhances the algorithm’s search capability in the later stages. In summary, these improvements facilitate the removal of redundant and noisy features, resulting in enhancing algorithm performance.
Table 8 presents the average fitness values obtained by the QBDA-DDO algorithm, with the best results highlighted in bold. It is evident from the table that QBDA-DDO consistently outperforms QBDA in terms of average fitness values on the majority of datasets. The fitness value serves as an indicator of the combined performance in terms of classification accuracy and the number of selected features. Thus, the proposed algorithm demonstrates superior performance compared to QBDA, showcasing its effectiveness in better overall results.
Figure 2 depicts the convergence behavior of the proposed QBDA-DDO algorithm compared to QBDA and QBDA1 on 14 mainstream datasets. QBDA represents the binary dragonfly algorithm with quadratic model-based parameter updates. The X-axis represents the number of iterations, and the Y-axis represents the fitness values. The results demonstrate that the proposed QBDA-DDO algorithm exhibits superior search capability in the later stages compared to QBDA for most datasets. The proposed algorithm effectively avoids getting trapped in local optima.

4.5. Comparison with SBDA Method

In this section, we carried out tests on the proposed method from three perspectives: classification accuracy, number of selected features, and fitness value. We also compared it with the SBDA algorithm, where SBDA1 represents the algorithm without the application of an adaptive step size mechanism. Table 9 presents the classification accuracy achieved by the proposed method on the 14 mainstream datasets. Table 10 displays the number of selected features obtained by SBDA-DDO. Table 11 shows the fitness values achieved by SBDA-DDO on the 14 datasets. Additionally, Figure 3 is used to compare the convergence speed of SBDA-DDO and SBDA on the 14 mainstream datasets.
In terms of classification accuracy, according to the results in Table 9, the SBDA-DDO algorithm achieves the best results on seven datasets, which are highlighted in bold. It is noteworthy that SBDA1 achieves the same classification accuracy as QBDA-DDO on five datasets, and all three algorithms show the same accuracy on three datasets. This indicates that the integration of the direction differential operator and the adaptive step size mechanism further improves the performance of the SBDA algorithm. Therefore, SBDA-DDO demonstrates superior ability to classify the datasets.
In terms of the number of selected features, it can be seen from Table 10 that BDA-DDO can achieve better performance, BDA-DDO achieves the best results on eight datasets, and SBDA1 obtained the best results on six datasets. Therefore, BDA-DDO is able to select fewer features, proving that the adaptive step can improve the performance of the algorithm, while the results obtained by SBDA1 are better than the original SBDA algorithm.
Table 11 illustrates the results obtained by the proposed SBDA-DDO algorithm in terms of fitness values, with the best results highlighted in bold. It can be observed that SBDA-DDO outperforms other algorithms on 13 datasets, while SBDA1 achieves the best results on 6 datasets. Remarkably, SBDA-DDO and SBDA1 yield the same results on five datasets. These findings confirm the effectiveness of our innovative approach, which incorporates direction differential operators and adaptive step size mechanisms into three different methods. The combination of these innovations with each algorithm demonstrates superior performance compared to the original algorithms. Consequently, our proposed approach enhances the algorithm’s performance by achieving improved classification accuracy while selecting fewer features.
Figure 3 depicts the convergence behavior of the proposed SBDA-DDO algorithm on 14 datasets, and the comparison with SBDA and SBDA1. SBDA represents the binary dragonfly algorithm with parameter updates using the cosine model. The X-axis represents the number of iterations, and the Y-axis displays the corresponding fitness values. Results show that SBDA-DDO exhibits faster convergence than SBDA for most datasets, thanks to the direction differential operator guiding individuals towards more promising directions. In addition, the later search ability of this algorithm is obviously stronger than that of SBDA, especially after using the adaptive step size mechanism.
The result of Figure 1, Figure 2 and Figure 3 confirms the significant impact of the proposed innovative mechanisms on enhancing the performance of BDA with various update strategies (linear, quadratic, and cosine). The results indicate that the introduced improvements effectively boost the performance of the three BDA variants, particularly improving their search capabilities in later stages and facilitating escape from local optima. This highlights the versatility and effectiveness of the proposed mechanisms in various scenarios, affirming the algorithm’s robustness and adaptability.

4.6. Comparison with Other Binary Optimization Algorithms

In the previous section, we compared the proposed BDA-DDO algorithm and three enhanced versions of BDA (LBDA, QBDA, SBDA) [40] in terms of classification accuracy, selected feature count, and fitness values. Now, we extend the comparison to four other typical metaheuristic feature selection algorithms: binary grey wolf optimization (BGWO) [48], the binary bat algorithm (BBA) [49], binary atomic search optimization (BASO) [50], and the binary gravitational search algorithm (BGSA) [49]. The results are presented in Table 12, Table 13 and Table 14 for classification accuracy, selected feature count, and fitness values, respectively. By considering these additional algorithms, we aim to provide a comprehensive evaluation of the performance of the proposed BDA-DDO algorithm.
Based on the results in Table 12, we compared our proposed method with BGWO, BBA, BASO, and BGSA in terms of classification accuracy. The best results are highlighted in bold. QBDA-DDO achieved the highest classification accuracy on 12 datasets, while LBDA-DDO and SBDA-DDO obtained the best results on 8 and 6 datasets, respectively. Therefore, QBDA outperformed other algorithms in improving classification accuracy.
The superior performance of QBDA-DDO can be attributed to its specific algorithmic strategies. By integrating the strengths of the QBDA algorithm and incorporating the direction bias operator and adaptive step size mechanism, QBDA-DDO provides efficient and accurate search capabilities. These strategies guide individuals in more promising directions, leading to improved classification accuracy. Furthermore, both LBDA-DDO and SBDA-DDO also demonstrate competitive performance on their respective datasets, achieving relatively good results.
Table 13 shows the average number of selected features for the proposed algorithm and four comparison algorithms on the 14 datasets. LBDA-DDO, QBDA-DDO, and BASO achieved similar performance, obtaining the best results on four datasets. On the other hand, the BBA algorithm demonstrated its advantage in feature selection by obtaining the best results on two datasets. The comparable performance of LBDA-DDO, QBDA-DDO, and BASO indicates their effectiveness in selecting a reasonable number of features on different datasets.
Overall, the results in Table 13 emphasize the importance of algorithm selection in the feature selection task. This comparison provides us with the advantages and disadvantages of different algorithms and helps researchers and practitioners to choose the most suitable method according to the specific dataset and needs.
Table 14 presents a comprehensive comparison of our proposed BDA-DDO algorithm and four comparison algorithms (BGWO, BBA, BASO, and BGSA) based on their average fitness values across 14 datasets. The results clearly demonstrate that QBDA-DDO outperforms the other algorithms, achieving the best fitness values on all 13 datasets. Additionally, LBDA-DDO and SBDA-DDO also show strong competitiveness by obtaining the best results on seven and three datasets, respectively. This highlights the effectiveness of the proposed algorithm compared to the four representative feature selection methods, demonstrating its superior performance.
Based on the analysis of the classification accuracy, number of selected features, and fitness, the proposed BDA-DDO algorithm consistently outperforms other algorithms across the majority of datasets. Among them, the QBDA-DDO algorithm demonstrates the highest performance, followed by LBDA-DDO and SBDA-DDO. The results highlight the effectiveness of our proposed method in selecting a smaller subset of features while achieving superior classification accuracy. Moreover, the proposed algorithm exhibits faster convergence speed and higher solution quality compared to three improved BDA algorithms. Furthermore, it proves to be highly competitive when compared to other feature selection algorithms.

5. Conclusions and Future Work

The purpose of this paper is to enhance the performance of the BDA algorithm through a hybrid approach. This goal has been successfully achieved by introducing three improvement mechanisms. Firstly, a novel differential operator, called the directed differential operator, is designed. Combining BDA with the directed differential operator provides a correct direction for the search process, resulting in faster convergence. Secondly, an adaptive update method is devised to enhance population diversity by updating the directed differential vector. Lastly, an adaptive step-updating mechanism is proposed to enhance the algorithm’s exploration capability by adjusting the dragonfly step.
The proposed algorithm is evaluated on 14 mainstream datasets from the UCI library and compared with seven representative feature selection algorithms. The experimental results show that the proposed BDA-DDO algorithm outperforms LBDA on 10 datasets in terms of classification accuracy, while achieving the same accuracy on 4 datasets. Additionally, BDA-DDO selects smaller feature subsets than LBDA on all 14 datasets. Compared to QBDA, BDA-DDO achieves higher classification accuracy on 10 datasets and the same accuracy on some low-dimensional datasets (reaching 1). Moreover, BDA-DDO selects smaller feature subsets than QBDA on all 14 datasets. When compared with SBDA, BDA-DDO achieves higher classification accuracy on nine datasets and selects fewer features on 13 datasets. In conclusion, BDA-DDO demonstrates its superiority over the three BDA algorithms (LBDA, QBDA, and SBDA) by consistently achieving higher classification accuracy while selecting smaller feature subsets on most datasets. Moreover, when compared to four other typical feature selection algorithms (BGWO, BBA, BASO, and BGSA), it also achieves higher classification accuracy.
Although the proposed improvement mechanisms have successfully enhanced the performance of the BDA algorithm, there are still some limitations. Specifically, the algorithm’s performance may be constrained when dealing with complex optimization scenarios, such as high-dimensional or large-scale datasets. Additionally, the feature selection problem is a multimodal problem, where the same number of features may correspond to different feature subsets. Currently, the algorithm may not be able to find all possible feature subsets. Moreover, the algorithm has not been tested in real-world applications, such as remote sensing tasks. Therefore, further research and experimentation is needed to address these issues and ensure the algorithm’s applicability and effectiveness in practical scenarios.
In future research, our main focus will be to explore the practical applications of the proposed algorithm, particularly in the field of remote sensing [51]. Remote sensing datasets often exhibit high dimensionality and large scale, with a substantial number of features, samples, and classes, and may contain feature correlations or redundancies. These characteristics make the feature selection problem in remote sensing datasets more complex and challenging, requiring consideration of feature interactions and impacts, as well as the efficiency and stability of feature selection algorithms.

Author Contributions

Conceptualization, B.G.; Methodology, B.G. and Y.C.; Software, B.G.; Validation, B.G.; Investigation, B.G.; Writing—original draft, B.G.; Writing—review & editing, Y.C., T.L., H.L., Y.W., D.Z. and X.L.; Supervision, Y.C., T.L., H.L., Y.W., D.Z. and X.L.; Project administration, Y.C.; Funding acquisition, Y.C., T.L. and X.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the National Key Research and Development Program of China under Grant NO. 2020YFB1313900, the Scientific Research Foundation of Wuhan Institute of Technology under Grant NO. 21QD53, the Innovation Fund of Hubei Key Laboratory of Intelligent Robot under Grant NO. HBIRL202107 and NO. HBIR 202105, the Science and Technology Research Project of Education Department of Hubei Province under Grant NO. Q20221501, the Graduate Innovative Fund of Wuhan Institute of Technology under Grant NO. CX2022342, Shenzhen Science and Technology Program under Grant NO. JCYJ20200109115201707 and JCYJ20220818101408019.

Data Availability Statement

The datasets used in this study are available at the UCI database at http://archive.ics.uci.edu/datasets (accessed on 4 May 2022).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Jiao, R.; Nguyen, B.H.; Xue, B.; Zhang, M. A Survey on Evolutionary Multiobjective Feature Selection in Classification: Approaches, Applications, and Challenges. IEEE Trans. Evol. Comput. 2023; Early Access. [Google Scholar] [CrossRef]
  2. Wang, J.; Zhang, Y.; Hong, M.; He, H.; Huang, S. A self-adaptive level-based learning artificial bee colony algorithm for feature selection on high-dimensional classification. Soft Comput. 2022, 26, 9665–9687. [Google Scholar] [CrossRef]
  3. Yang, S.; Gu, L.; Li, X.; Jiang, T.; Ren, R. Crop classification method based on optimal feature selection and hybrid CNN-RF networks for multi-temporal remote sensing imagery. Remote Sens. 2020, 12, 3119. [Google Scholar] [CrossRef]
  4. Kumar, V.; Minz, S. Feature selection: A literature review. SmartCR 2014, 4, 211–229. [Google Scholar] [CrossRef]
  5. Hennessy, A.; Clarke, K.; Lewis, M. Hyperspectral classification of plants: A review of waveband selection generalisability. Remote Sens. 2020, 12, 113. [Google Scholar] [CrossRef] [Green Version]
  6. Guo, A.; Huang, W.; Ye, H.; Dong, Y.; Ma, H.; Ren, Y.; Ruan, C. Identification of wheat yellow rust using spectral and texture features of hyperspectral images. Remote Sens. 2020, 12, 1419. [Google Scholar] [CrossRef]
  7. Mohamed, S.A.; Metwaly, M.M.; Metwalli, M.R.; AbdelRahman, M.A.; Badreldin, N. Integrating Active and Passive Remote Sensing Data for Mapping Soil Salinity Using Machine Learning and Feature Selection Approaches in Arid Regions. Remote Sens. 2023, 15, 1751. [Google Scholar] [CrossRef]
  8. Kiala, Z.; Mutanga, O.; Odindi, J.; Peerbhay, K. Feature selection on sentinel-2 multispectral imagery for mapping a landscape infested by parthenium weed. Remote Sens. 2019, 11, 1892. [Google Scholar] [CrossRef] [Green Version]
  9. Bing, X.; Zhang, M.; Browne, W.N.; Xin, Y. A Survey on Evolutionary Computation Approaches to Feature Selection. IEEE Trans. Evol. Comput. 2016, 20, 606–626. [Google Scholar]
  10. Chaudhuri, A.; Sahu, T.P. Feature selection using Binary Crow Search Algorithm with time varying flight length. Expert Syst. Appl. 2021, 168, 114288. [Google Scholar] [CrossRef]
  11. Javed, K.; Babri, H.A.; Saeed, M. Feature selection based on class-dependent densities for high-dimensional binary data. IEEE Trans. Knowl. Data Eng. 2010, 24, 465–477. [Google Scholar] [CrossRef]
  12. Xue, B.; Zhang, M.; Browne, W.N. A comprehensive comparison on evolutionary feature selection approaches to classification. Int. J. Comput. Intell. Appl. 2015, 14, 1550008. [Google Scholar] [CrossRef]
  13. Dhiman, G.; Oliva, D.; Kaur, A.; Singh, K.K.; Vimal, S.; Sharma, A.; Cengiz, K. BEPO: A novel binary emperor penguin optimizer for automatic feature selection. Knowl.-Based Syst. 2021, 211, 106560. [Google Scholar] [CrossRef]
  14. Rostami, M.; Berahmand, K.; Nasiri, E.; Forouzandeh, S. Review of swarm intelligence-based feature selection methods. Eng. Appl. Artif. Intell. 2021, 100, 104210. [Google Scholar] [CrossRef]
  15. Cheng, F.; Cui, J.; Wang, Q.; Zhang, L. A Variable Granularity Search-Based Multiobjective Feature Selection Algorithm for High-Dimensional Data Classification. IEEE Trans. Evol. Comput. 2023, 27, 266–280. [Google Scholar] [CrossRef]
  16. Paniri, M.; Dowlatshahi, M.B.; Nezamabadi-Pour, H. MLACO: A multi-label feature selection algorithm based on ant colony optimization. Knowl.-Based Syst. 2020, 192, 105285. [Google Scholar] [CrossRef]
  17. Wang, Z.; Gao, S.; Zhou, M.; Sato, S.; Cheng, J.; Wang, J. Information-Theory-based Nondominated Sorting Ant Colony Optimization for Multiobjective Feature Selection in Classification. IEEE Trans. Cybern. 2023, 53, 5276–5289. [Google Scholar] [CrossRef] [PubMed]
  18. Alhasan, W.M.; Ibrahim, S.; Hefny, H.A.; Shaheen, S.I. LDWMeanPSO: A new improved particle swarm optimization technique. In Proceedings of the 2011 Seventh International Computer Engineering Conference (ICENCO’2011), Giza, Egypt, 27–28 December 2011; pp. 37–43. [Google Scholar]
  19. Li, G.; Yan, L.; Qu, B. Multi-Objective Particle Swarm Optimization Based on Gaussian Sampling. IEEE Access 2020, 8, 209717–209737. [Google Scholar] [CrossRef]
  20. Lin, A.; Sun, W.; Yu, H.; Wu, G.; Tang, H. Global genetic learning particle swarm optimization with diversity enhancement by ring topology. Swarm Evol. Comput. 2019, 44, 571–583. [Google Scholar] [CrossRef]
  21. Deb, K.; Pratap, A.; Agarwal, S.; Meyarivan, T. A fast and elitist multiobjective genetic algorithm: NSGA-II. IEEE Trans. Evol. Comput. 2002, 6, 182–197. [Google Scholar] [CrossRef] [Green Version]
  22. Varghese, N.V.; Singh, A.; Suresh, A.; Rahnamayan, S. Binary hybrid differential evolution algorithm for multi-label feature selection. In Proceedings of the 2020 IEEE International Conference on Systems, Man, and Cybernetics (SMC), Toronto, ON, Canada, 11–14 October 2020; pp. 4386–4391. [Google Scholar]
  23. Zhang, Y.; Gong, D.W.; Rong, M. Multi-objective differential evolution algorithm for multi-label feature selection in classification. In Advances in Swarm and Computational Intelligence: 6th International Conference, ICSI 2015, Held in Conjunction with the Second BRICS Congress, CCI 2015, Beijing, China, 25–28 June 2015, Proceedings, Part I 6; Springer: Cham, Switzerland, 2015; pp. 339–345. [Google Scholar]
  24. Yang, X.S.; Gandomi, A.H. Bat algorithm: A novel approach for global engineering optimization. Eng. Comput. 2012, 29, 464–483. [Google Scholar] [CrossRef] [Green Version]
  25. Nakamura, R.Y.M.; Pereira, L.A.M.; Rodrigues, D.; Costa, K.A.P.; Papa, J.P.; Yang, X.S. Binary bat algorithm for feature selection. In Swarm Intelligence and Bio-Inspired Computation; Elsevier: Amsterdam, The Netherlands, 2013; pp. 225–237. [Google Scholar]
  26. Mirjalili, S.; Mirjalili, S.M.; Lewis, A. Grey wolf optimizer. Adv. Eng. Softw. 2014, 69, 46–61. [Google Scholar] [CrossRef] [Green Version]
  27. Arora, S.; Singh, H.; Sharma, M.; Sharma, S.; Anand, P. A new hybrid algorithm based on grey wolf optimization and crow search algorithm for unconstrained function optimization and feature selection. IEEE Access 2019, 7, 26343–26361. [Google Scholar] [CrossRef]
  28. Ibrahim, R.A.; Abd Elaziz, M.; Lu, S. Chaotic opposition-based grey-wolf optimization algorithm based on differential evolution and disruption operator for global optimization. Expert Syst. Appl. 2018, 108, 1–27. [Google Scholar] [CrossRef]
  29. Mirjalili, S. Dragonfly algorithm: A new meta-heuristic optimization technique for solving single-objective, discrete, and multi-objective problems. Neural Comput. Appl. 2016, 27, 1053–1073. [Google Scholar] [CrossRef]
  30. Zhang, Z.; Ding, S.; Jia, W. A hybrid optimization algorithm based on cuckoo search and differential evolution for solving constrained engineering problems. Eng. Appl. Artif. Intell. 2019, 85, 254–268. [Google Scholar] [CrossRef]
  31. Tubishat, M.; Idris, N.; Shuib, L.; Abushariah, M.A.; Mirjalili, S. Improved Salp Swarm Algorithm based on opposition based learning and novel local search algorithm for feature selection. Expert Syst. Appl. 2020, 145, 113122. [Google Scholar] [CrossRef]
  32. Hussain, K.; Neggaz, N.; Zhu, W.; Houssein, E.H. An efficient hybrid sine-cosine Harris hawks optimization for low and high-dimensional feature selection. Expert Syst. Appl. 2021, 176, 114778. [Google Scholar] [CrossRef]
  33. Mafarja, M.M.; Eleyan, D.; Jaber, I.; Hammouri, A.; Mirjalili, S. Binary dragonfly algorithm for feature selection. In Proceedings of the 2017 International Conference on New Trends in Computing Sciences (ICTCS), Amman, Jordan, 11–13 October 2017; pp. 12–17. [Google Scholar]
  34. Xue, B.; Zhang, M.; Browne, W.N. Particle swarm optimization for feature selection in classification: A multi-objective approach. IEEE Trans. Cybern. 2012, 43, 1656–1671. [Google Scholar] [CrossRef]
  35. Rashno, A.; Shafipour, M.; Fadaei, S. Particle ranking: An efficient method for multi-objective particle swarm optimization feature selection. Knowl.-Based Syst. 2022, 245, 108640. [Google Scholar] [CrossRef]
  36. Luo, J.; Zhou, D.; Jiang, L.; Ma, H. A particle swarm optimization based multiobjective memetic algorithm for high-dimensional feature selection. Memetic Comput. 2022, 14, 77–93. [Google Scholar] [CrossRef]
  37. Chen, K.; Xue, B.; Zhang, M.; Zhou, F. Hybridising Particle Swarm optimisation with Differential Evolution for Feature Selection in Classification. In Proceedings of the IEEE Congress on Evolutionary Computation (CEC), Glasgow, UK, 19–24 July 2020. [Google Scholar]
  38. Wang, X.; Wang, Y.; Wong, K.C.; Li, X. A self-adaptive weighted differential evolution approach for large-scale feature selection. Knowl.-Based Syst. 2022, 235, 107633. [Google Scholar] [CrossRef]
  39. Wang, P.; Xue, B.; Liang, J.; Zhang, M. Multiobjective Differential Evolution for Feature Selection in Classification. IEEE Trans. Cybern. 2023, 53, 4579–4593. [Google Scholar] [CrossRef]
  40. Hammouri, A.I.; Mafarja, M.; Al-Betar, M.A.; Awadallah, M.A.; Abu-Doush, I. An improved dragonfly algorithm for feature selection. Knowl.-Based Syst. 2020, 203, 106–131. [Google Scholar] [CrossRef]
  41. Too, J.; Mirjalili, S. A hyper learning binary dragonfly algorithm for feature selection: A COVID-19 case study. Knowl.-Based Syst. 2021, 212, 106553. [Google Scholar] [CrossRef]
  42. Chantar, H.; Tubishat, M.; Essgaer, M.; Mirjalili, S. Hybrid Binary Dragonfly Algorithm with Simulated Annealing for Feature Selection. SN Comput. Sci. 2021, 2, 295. [Google Scholar] [CrossRef] [PubMed]
  43. Duan, M.; Yang, H.; Yang, B.; Wu, X.; Liang, H. Hybridizing dragonfly algorithm with differential evolution for global optimization. IEICE Trans. Inf. Syst. 2019, 102, 1891–1901. [Google Scholar] [CrossRef] [Green Version]
  44. Storn, R.; Price, K. Differential evolution—A simple and efficient heuristic for global optimization over continuous spaces. J. Glob. Optim. 1997, 11, 341–359. [Google Scholar] [CrossRef]
  45. Price, K.; Storn, R.M.; Lampinen, J.A. Differential Evolution: A Practical Approach to Global Optimization; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2006. [Google Scholar]
  46. Das, S.; Abraham, A.; Chakraborty, U.K.; Konar, A. Differential evolution using a neighborhood-based mutation operator. IEEE Trans. Evol. Comput. 2009, 13, 526–553. [Google Scholar] [CrossRef] [Green Version]
  47. Dua, D.; Graff, C.; Bache, K.; Lichman, M. UCI Machine Learning Repository. 2017. Available online: http://archive.ics.uci.edu/datasets (accessed on 4 May 2022).
  48. Emary, E.; Zawba, H.M.; Hassanien, A.E. Binary grey wolf optimization approaches for feature selection. Neurocomputing 2016, 172, 371–381. [Google Scholar] [CrossRef]
  49. Mafarja, M.; Aljarah, I.; Heidari, A.A.; Hammouri, A.I.; Faris, H.; Ala’M, A.Z.; Mirjalili, S. Evolutionary population dynamics and grasshopper optimization approaches for feature selection problems. Knowl.-Based Syst. 2018, 145, 25–45. [Google Scholar] [CrossRef] [Green Version]
  50. Too, J.; Rahim Abdullah, A. Binary atom search optimisation approaches for feature selection. Connect. Sci. 2020, 32, 406–430. [Google Scholar] [CrossRef]
  51. Wang, M.; Wu, C.; Wang, L.; Xiang, D.; Huang, X. A feature selection approach for hyperspectral image based on modified ant lion optimizer. Knowl.-Based Syst. 2019, 168, 39–48. [Google Scholar] [CrossRef]
Figure 1. Convergence behavior of the proposed algorithm based on LBDA on 14 mainstream datasets.
Figure 1. Convergence behavior of the proposed algorithm based on LBDA on 14 mainstream datasets.
Remotesensing 15 03980 g001aRemotesensing 15 03980 g001b
Figure 2. Convergence behavior of the proposed algorithm based on QBDA on 14 mainstream datasets.
Figure 2. Convergence behavior of the proposed algorithm based on QBDA on 14 mainstream datasets.
Remotesensing 15 03980 g002
Figure 3. Convergence behavior of the proposed algorithm based on SBDA on 14 mainstream datasets.
Figure 3. Convergence behavior of the proposed algorithm based on SBDA on 14 mainstream datasets.
Remotesensing 15 03980 g003
Table 1. Details of datasets.
Table 1. Details of datasets.
DataSetNo. of AttributesNo. of ObjectsNo. of Classes
Breastcancer96992
BreastEW305692
Exactly1310002
HeartEW132702
Lymphography181484
PenglungEW325737
SonarEW602082
SpectEW222672
CongressEW164352
IonosphereEW343512
KrvskpEW3631962
WaveformEW4050003
WineEW131783
Zoo161017
Table 2. The parameter settings of algorithms.
Table 2. The parameter settings of algorithms.
ParameterValue
Population size10
The maximum number of iterations100
K parameters in KNN5
  C R 0.3
β  of SBDA   π 3
a of BGWOfrom 2 to 0
G 0  of BGSA100
α  of BGSA20
Table 3. Average classification accuracy based on the proposed LBDA-DDO algorithm.
Table 3. Average classification accuracy based on the proposed LBDA-DDO algorithm.
DataSetMetricLBDALBDA1LBDA-DDO
BreastcancerAvg0.9990.9990.999
StDev0.0030.0020.003
BreastEWAvg0.9930.9940.994
StDev0.0040.0040.004
ExactlyAvg0.9520.9721
StDev0.0530.0330
HeartEWAvg0.9030.9210.935
StDev0.0200.0180.019
LymphographyAvg0.9560.9700.964
StDev0.0190.0120.014
PenglungEWAvg111
StDev000
SonarEWAvg0.9540.9600.964
StDev0.0190.0130.015
SpectEWAvg0.9220.9290.929
StDev0.0140.0140.017
CongressEWAvg0.9980.9991
StDev0.0030.0040
IonosphereEWAvg0.9600.9660.967
StDev0.0080.0110.011
KrvskpEWAvg0.9700.9680.985
StDev0.0070.0050.003
WaveformEWAvg0.8310.8290.845
StDev0.0050.0050.005
WineEWAvg111
StDev000
ZooAvg111
StDev000
Table 4. Average selected features based on proposed LBDA-DDO algorithm.
Table 4. Average selected features based on proposed LBDA-DDO algorithm.
DataSetMetricLBDALBDA1LBDA-DDO
BreastcancerAvg4.974.574.46
StDev1.301.221.20
BreastEWAvg14.1713.9312.57
StDev2.072.822.01
ExactlyAvg7.307.176
StDev0.780.580
HeartEWAvg5.704.774.33
StDev1.241.541.42
LymphographyAvg7.738.136.77
StDev2.062.442.06
PenglungEWAvg126.13106.13105.30
StDev19.175.939.54
SonarEWAvg27.6324.4323.00
StDev3.666.003.85
SpectEWAvg9.839.508.07
StDev2.654.093.10
CongressEWAvg6.504.834.33
StDev2.191.801.81
IonosphereEWAvg15.3312.8011.17
StDev3.273.673.10
KrvskpEWAvg19.9019.0716.53
StDev2.753.082.30
WaveformEWAvg22.3722.6719.83
StDev3.113.382.55
WineEWAvg4.433.702.67
StDev0.880.820.65
ZooAvg4.904.403.93
StDev0.700.660.72
Table 5. Average fitness based on the proposed LBDA-DDO algorithm.
Table 5. Average fitness based on the proposed LBDA-DDO algorithm.
DataSetMetricLBDALBDA1LBDA-DDO
BreastcancerAvg0.0070.0060.006
StDev0.0030.0020.002
BreastEWAvg0.0110.0100.009
StDev0.0040.0040.004
ExactlyAvg0.0520.0320.005
StDev0.0530.0580
HeartEWAvg0.1010.0810.066
StDev0.0200.0170.019
LymphographyAvg0.0470.0340.039
StDev0.0190.0110.014
PenglungEWAvg0.0040.0030.003
StDev0.00100
SonarEWAvg0.0500.0440.039
StDev0.0180.0120.014
SpectEWAvg0.0820.0750.073
StDev0.0130.0130.017
CongressEWAvg0.0060.0040.002
StDev0.0040.0030.001
IonosphereEWAvg0.0440.0380.036
StDev0.0080.0100.011
KrvskpEWAvg0.0360.0370.019
StDev0.0070.0050.003
WaveformEWAvg0.1730.1750.157
StDev0.0050.0050.005
WineEWAvg0.0030.0030.002
StDev0.0010.0010
ZooAvg0.0030.0030.002
StDev000
Table 6. The average classification accuracy of the proposed algorithm QBDA-DDO.
Table 6. The average classification accuracy of the proposed algorithm QBDA-DDO.
DataSetMetricQBDAQBDA1QBDA-DDO
BreastcancerAvg0.9980.9990.999
StDev0.0020.0020.002
BreastEWAvg0.9940.9950.994
StDev0.0040.0040.004
ExactlyAvg0.9740.9901
StDev0.0440.0230
HeartEWAvg0.9150.9230.945
StDev0.0170.0180.019
LymphographyAvg0.9630.9690.971
StDev0.0230.0180.020
PenglungEWAvg111
StDev000
SonarEWAvg0.9580.9680.965
StDev0.0150.0170.016
SpectEWAvg0.9200.9300.940
StDev0.0140.0130.013
CongressEWAvg0.99911
StDev0.0020.0020
IonosphereEWAvg0.9590.9650.970
StDev0.0090.0110.011
KrvskpEWAvg0.9720.9750.984
StDev0.0040.0050.005
WaveformEWAvg0.8340.8350.849
StDev0.0060.0040.010
WineEWAvg111
StDev000
ZooAvg111
StDev000
Table 7. The average number of the selected features of the proposed algorithm QBDA-DDO.
Table 7. The average number of the selected features of the proposed algorithm QBDA-DDO.
DataSetMetricQBDAQBDA1QBDA-DDO
BreastcancerAvg4.804.574.13
StDev1.221.170.76
BreastEWAvg14.7313.7311.77
StDev2.432.951.83
ExactlyAvg7.006.636.03
StDev0.890.600.17
HeartEWAvg5.94.534.53
StDev1.851.821.17
LymphographyAvg8.377.307.03
StDev2.232.181.74
PenglungEWAvg128.00112.17108.90
StDev6.675.015.73
SonarEWAvg27.1323.0022.60
StDev3.304.603.92
SpectEWAvg9.0010.677.97
StDev1.712.872.96
CongressEWAvg6.175.074.10
StDev1.971.361.12
IonosphereEWAvg14.0012.4010.37
StDev3.294.023.02
KrvskpEWAvg19.6719.3317.90
StDev2.383.052.00
WaveformEWAvg22.8721.8321.20
StDev2.282.503.15
WineEWAvg4.533.633.06
StDev1.150.660.51
ZooAvg5.034.233.87
StDev0.910.670.42
Table 8. The average fitness value of the proposed algorithm QBDA-DDO.
Table 8. The average fitness value of the proposed algorithm QBDA-DDO.
DataSetMetricQBDAQBDA1QBDA-DDO
BreastcancerAvg0.0060.0060.005
StDev0.0020.0020.002
BreastEWAvg0.0100.0090.009
StDev0.0040.0040.004
ExactlyAvg0.0310.0150.005
StDev0.0440.0240
HeartEWAvg0.0880.0790.058
StDev0.0160.0170.019
LymphographyAvg0.0410.0350.032
StDev0.0220.0170.020
PenglungEWAvg0.0040.0030.003
StDev0.00200
SonarEWAvg0.0450.0350.038
StDev0.0150.0170.016
SpectEWAvg0.0820.0730.063
StDev0.0140.0130.013
CongressEWAvg0.0040.0030.002
StDev0.0020.0020
IonosphereEWAvg0.0440.0380.032
StDev0.0090.0100.012
KrvskpEWAvg0.0330.0300.020
StDev0.0040.0050.005
WaveformEWAvg0.1700.1690.154
StDev0.0060.0040.010
WineEWAvg0.0030.0030.002
StDev000
ZooAvg0.0030.0030.002
StDev000
Table 9. The average classification accuracy of the proposed algorithm SBDA-DDO.
Table 9. The average classification accuracy of the proposed algorithm SBDA-DDO.
DataSetMetricSBDASBDA1SBDA-DDO
BreastcancerAvg0.9980.9990.999
StDev0.0030.0030.002
BreastEWAvg0.9930.9940.993
StDev0.0030.0040.004
ExactlyAvg0.9460.9600.987
StDev0.0600.0520.025
HeartEWAvg0.8970.9090.919
StDev0.0180.0200.019
LymphographyAvg111
StDev000
PenglungEWAvg111
StDev000
SonarEWAvg0.9600.9630.964
StDev0.0130.0200.015
SpectEWAvg0.9240.9290.929
StDev0.0180.0140.018
CongressEWAvg0.9970.9991
StDev0.0050.0030
IonosphereEWAvg0.9590.9610.968
StDev0.0100.0110.011
KrvskpEWAvg0.9680.9650.974
StDev0.0050.0050.004
WaveformEWAvg0.8320.8280.834
StDev0.0060.0070.006
WineEWAvg111
StDev000
ZooAvg111
StDev000
Table 10. The average number of the selected features of the proposed algorithm SBDA-DDO.
Table 10. The average number of the selected features of the proposed algorithm SBDA-DDO.
DataSetMetricSBDASBDA1SBDA-DDO
BreastcancerAvg4.904.035.07
StDev1.321.051.34
BreastEWAvg14.4013.6313.17
StDev2.393.732.74
ExactlyAvg7.477.136.77
StDev0.620.800.61
HeartEWAvg4.874.134.80
StDev1.231.171.25
LymphographyAvg8.777.977.73
StDev2.172.642.00
PenglungEWAvg136.50103.03108.47
StDev15.324.984.72
SonarEWAvg28.9324.3023.73
StDev3.956.465.11
SpectEWAvg10.508.579.00
StDev2.563.803.63
CongressEWAvg6.634.435.10
StDev2.551.411.87
IonosphereEWAvg13.8011.1712.37
StDev2.262.583.65
KrvskpEWAvg20.0719.4718.63
StDev2.053.872.73
WaveformEWAvg22.9721.4320.50
StDev3.173.572.80
WineEWAvg4.703.603.23
StDev1.290.880.67
ZooAvg5.134.504.13
StDev0.760.880.88
Table 11. The average fitness value of the proposed algorithm SBDA-DDO.
Table 11. The average fitness value of the proposed algorithm SBDA-DDO.
DataSetMetricSBDASBDA1SBDA-DDO
BreastcancerAvg0.0070.0060.006
StDev0.0020.0020.002
BreastEWAvg0.0120.0110.011
StDev0.0030.0040.004
ExactlyAvg0.0590.0450.018
StDev0.0600.0530.025
HeartEWAvg0.1050.0930.083
StDev0.0180.0200.018
LymphographyAvg0.0460.0380.042
StDev0.0160.0210.018
PenglungEWAvg0.0040.0030.003
StDev000
SonarEWAvg0.0470.0400.039
StDev0.0130.0200.014
SpectEWAvg0.0790.0740.074
StDev0.0180.0140.017
CongressEWAvg0.0060.0030.003
StDev0.0040.0020.001
IonosphereEWAvg0.0450.0410.036
StDev0.0100.0110.010
KrvskpEWAvg0.0370.0400.031
StDev0.0050.0050.004
WaveformEWAvg0.1720.1760.169
StDev0.0060.0070.006
WineEWAvg0.0040.0030.002
StDev000
ZooAvg0.0030.0030.002
StDev000
Table 12. Comparison with other binary algorithms on average classification accuracy.
Table 12. Comparison with other binary algorithms on average classification accuracy.
DataSetMetricLBDA-DDOQBDA-DDOSBDA-DDOBGWOBBABASOBGSA
BreastcancerAvg0.9990.9990.9990.9780.9800.9710.948
StDev0.0030.0020.0020.0110.0090.0130.051
BreastEWAvg0.9940.9940.9930.9780.9830.9600.928
StDev0.0040.0040.0040.0130.0090.0180.014
ExactlyAvg110.9870.8050.9820.7270.732
StDev000.0250.0650.0690.0700.124
HeartEWAvg0.9350.9450.9190.8360.8870.8250.770
StDev0.0190.0190.0190.0570.0350.0420.066
LymphographyAvg0.9640.97110.8720.9110.8590.864
StDev0.0140.02000.0530.0490.0490.081
PenglungEWAvg1110.8830.9240.9240.949
StDev0000.0720.0690.0740.054
SonarEWAvg0.9640.9650.9640.9020.9390.8810.865
StDev0.0150.0160.0150.0440.0390.0490.047
SpectEWAvg0.9290.9400.9290.8630.8940.8310.785
StDev0.0170.0130.0180.0440.0360.0310.034
CongressEWAvg1110.9770.9830.9630.943
StDev0000.0180.0120.0230.026
IonosphereEWAvg0.9670.9700.9680.9080.9320.9090.869
StDev0.0110.0110.0110.0370.0250.0250.026
KrvskpEWAvg0.9850.9840.9740.9720.9830.9040.898
StDev0.0030.0050.0040.0080.0050.0410.051
WaveformEWAvg0.8450.8490.8340.8430.8450.7990.697
StDev0.0050.0100.0060.0110.0100.0150.021
WineEWAvg1110.9550.9860.9460.976
StDev0000.0450.0160.0490.035
ZooAvg1110.9430.9850.9330.995
StDev0000.0590.0260.0450.015
Table 13. Comparison with other binary algorithms on average number of selected features.
Table 13. Comparison with other binary algorithms on average number of selected features.
DataSetMetricLBDA-DDOQBDA-DDOSBDA-DDOBGWOBBABASOBGSA
BreastcancerAvg4.464.135.075.234.174.234.47
StDev1.200.761.341.451.211.231.01
BreastEWAvg12.5711.7713.1717.9311.3311.9014.93
StDev2.011.832.742.682.283.402
ExactlyAvg66.036.779.9765.577.67
StDev00.170.611.540.453.291.49
HeartEWAvg4.334.534.806.974.504.406.63
StDev1.421.1171.251.541.061.941.94
LymphographyAvg6.777.037.7310.235.476.009
StDev2.061.742.001.711.452.212.18
PenglungEWAvg105.30108.90108.47168.40127.9062.37145.1
StDev9.545.734.7221.798.3332.814.88
SonarEWAvg23.0022.6023.7333.9323.2718.2727.07
StDev3.853.925.113.263.897.973.64
SpectEWAvg8.077.979.0012.308.278.239.77
StDev3.102.963.633.001.864.072.3
CongressEWAvg4.334.105.108.035.334.177
StDev1.811.121.871.941.722.031.91
IonosphereEWAvg11.1710.3712.3718.4310.236.9314.9
StDev3.103.023.652.552.581.932.89
KrvskpEWAvg16.5317.9018.6328.4019.8018.2719.73
StDev2.302.002.732.582.454.402.36
WaveformEWAvg19.8321.2020.5031.2722.9321.2721.6
StDev2.553.152.801.862.534.163.69
WineEWAvg2.673.063.236.533.533.736.57
StDev0.650.510.671.761.151.211.36
ZooAvg3.933.874.1385.375.536.97
StDev0.720.420.881.731.051.801.25
Table 14. Comparison with other binary-based algorithms in terms of average fitness value.
Table 14. Comparison with other binary-based algorithms in terms of average fitness value.
DataSetMetricLBDA-DDOQBDA-DDOSBDA-DDOBGWOBBABASOBGSA
BreastcancerAvg0.0060.0050.0060.0270.0240.0330.027
StDev0.0020.0020.0020.0110.0090.0130.007
BreastEWAvg0.0090.0090.0110.0270.0200.0440.039
StDev0.0040.0040.0040.0120.0090.0180.01
ExactlyAvg0.0050.0050.0180.2000.0230.2750.253
StDev000.0250.0650.0680.0690.094
HeartEWAvg0.0660.0580.0830.1670.1150.1760.137
StDev0.0190.0190.0180.0560.0340.0420.03
LymphographyAvg0.0390.0320.0420.1320.0900.1430.081
StDev0.0140.0200.0180.0520.0490.0480.033
PenglungEWAvg0.0030.0030.0030.1210.0790.0770.004
StDev0000.0720.0680.0730
SonarEWAvg0.0390.0380.0390.1020.0640.1200.082
StDev0.0140.0160.0140.0440.0390.0480.023
SpectEWAvg0.0730.0630.0740.1410.1080.1710.153
StDev0.0170.0130.0170.0430.0350.0300.018
CongressEWAvg0.0020.0020.0030.0270.0200.0390.032
StDev0.00100.0010.0180.0120.0230.013
IonosphereEWAvg0.0360.0320.0360.0970.0700.0920.127
StDev0.0110.0120.0100.0370.0250.0240.011
KrvskpEWAvg0.0190.0200.0310.0350.0220.1000.099
StDev0.0030.0050.0040.0080.0050.0400.049
WaveformEWAvg0.1570.1540.1690.1630.1590.2040.251
StDev0.0050.0100.0060.0110.0100.0150.013
WineEWAvg0.0020.0020.0020.0490.0170.0570.009
StDev0000.0440.0160.0480.012
ZooAvg0.0020.0020.0020.0610.0180.0690.005
StDev0000.0580.0260.0450.001
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, Y.; Gao, B.; Lu, T.; Li, H.; Wu, Y.; Zhang, D.; Liao, X. A Hybrid Binary Dragonfly Algorithm with an Adaptive Directed Differential Operator for Feature Selection. Remote Sens. 2023, 15, 3980. https://doi.org/10.3390/rs15163980

AMA Style

Chen Y, Gao B, Lu T, Li H, Wu Y, Zhang D, Liao X. A Hybrid Binary Dragonfly Algorithm with an Adaptive Directed Differential Operator for Feature Selection. Remote Sensing. 2023; 15(16):3980. https://doi.org/10.3390/rs15163980

Chicago/Turabian Style

Chen, Yilin, Bo Gao, Tao Lu, Hui Li, Yiqi Wu, Dejun Zhang, and Xiangyun Liao. 2023. "A Hybrid Binary Dragonfly Algorithm with an Adaptive Directed Differential Operator for Feature Selection" Remote Sensing 15, no. 16: 3980. https://doi.org/10.3390/rs15163980

APA Style

Chen, Y., Gao, B., Lu, T., Li, H., Wu, Y., Zhang, D., & Liao, X. (2023). A Hybrid Binary Dragonfly Algorithm with an Adaptive Directed Differential Operator for Feature Selection. Remote Sensing, 15(16), 3980. https://doi.org/10.3390/rs15163980

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop