Next Article in Journal
Biased Random Walk Model of Neuronal Dynamics on Substrates with Periodic Geometrical Patterns
Next Article in Special Issue
A Comprehensive Review of Bio-Inspired Optimization Algorithms Including Applications in Microelectronics and Nanophotonics
Previous Article in Journal
Biomimetics with Trade-Offs
Previous Article in Special Issue
Improved Dipper-Throated Optimization for Forecasting Metamaterial Design Bandwidth for Engineering Applications
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Binary Restructuring Particle Swarm Optimization and Its Application

1
School of Computer Science and Mathematics, Fujian University of Technology, Fuzhou 350118, China
2
Fujian Provincial Key Laboratory of Big Data Mining and Applications, Fuzhou 350118, China
*
Author to whom correspondence should be addressed.
Biomimetics 2023, 8(2), 266; https://doi.org/10.3390/biomimetics8020266
Submission received: 16 May 2023 / Revised: 14 June 2023 / Accepted: 15 June 2023 / Published: 17 June 2023

Abstract

:
Restructuring Particle Swarm Optimization (RPSO) algorithm has been developed as an intelligent approach based on the linear system theory of particle swarm optimization (PSO). It streamlines the flow of the PSO algorithm, specifically targeting continuous optimization problems. In order to adapt RPSO for solving discrete optimization problems, this paper proposes the binary Restructuring Particle Swarm Optimization (BRPSO) algorithm. Unlike other binary metaheuristic algorithms, BRPSO does not utilize the transfer function. The particle updating process in BRPSO relies solely on comparison results between values derived from the position updating formula and a random number. Additionally, a novel perturbation term is incorporated into the position updating formula of BRPSO. Notably, BRPSO requires fewer parameters and exhibits high exploration capability during the early stages. To evaluate the efficacy of BRPSO, comprehensive experiments are conducted by comparing it against four peer algorithms in the context of feature selection problems. The experimental results highlight the competitive nature of BRPSO in terms of both classification accuracy and the number of selected features.

1. Introduction

The Particle Swarm Optimization (PSO) algorithm, initially proposed by Eberhart and Kennedy in 1995 [1], is a heuristic search algorithm grounded in evolutionary algorithms and artificial life. PSO exhibits a memetic nature, facilitating the retention of well-performing particles. Moreover, the parameters of PSO can be easily adjusted, enabling its applicability to a wide range of practical problems. Drawing inspiration from the foraging behavior of birds, PSO utilizes historical experiences and social knowledge to explore the solution space and identify optimal solutions. Each particle in PSO possesses position and velocity vectors, with the former representing a potential solution and the latter governing adjustments to the particle’s flight direction. The updating formula for the position and velocity vectors in the canonical PSO [2] is provided below.
v i d ( t + 1 ) = w × v i d ( t ) + c 1 × r 1 × p b e s t i d x i d ( t ) + c 2 × r 2 × g b e s t d x i d ( t )
x i d ( t + 1 ) = x i d ( t ) + v i d ( t + 1 )
where d denotes the dimensionality of the problem. i represents the number of particles in the swarm. The iteration number during the evolutionary process is denoted by t. w is the inertial weight, which controls the proportion inherited from the previous velocities in updating current velocity. Two acceleration factors, c 1 and c 2 control the impact of particle learning p b e s t and g b e s t on velocity. Additionally, two random numbers, r 1 and r 2 , uniformly distributed in the interval [0, 1], are utilized. The variable p b e s t i represents the historical optimal solution of the ith particle, while g b e s t represents the global optimal solution identified during the course of the evolutionary process.
While Particle Swarm Optimization (PSO) has predominantly been explored for optimization problems in continuous spaces, its application to optimization problems in binary spaces has received relatively limited attention. In 1997, Eberhart and Kennedy introduced a binary variant of PSO called Binary Particle Swarm Optimization (BPSO) specifically designed for addressing optimization problems in discrete binary spaces [3]. In BPSO, the position vector is encoded in a binary format. Building upon PSO, BPSO incorporates a sigmoid transfer function to map the velocity values onto the interval [0, 1]. The subsequent position value is determined by comparing the output of the transfer function to a randomly generated number within the [0, 1] interval, resulting in either “0” or “1”. The specific updating formula for BPSO is elucidated below.
T f v i d ( t + 1 ) = sigmoid v i d ( t + 1 ) = 1 1 + e v i d ( t + 1 )
x i d ( t + 1 ) = 1 if T f v i d ( t + 1 ) > rand ( ) 0 if T f v i d ( t + 1 ) rand ( )
Since the initial proposal of the BPSO algorithm, researchers have made notable advancements in its efficacy for solving diverse practical discrete combinatorial optimization problems [4,5,6]. Despite its merits, BPSO inherits certain limitations from the original PSO algorithm due to their shared velocity updating formula. Notably, BPSO exhibits a tendency to converge prematurely when confronted with complex problems. The performance of BPSO is significantly influenced by the design of the transfer function, and the sigmoid transfer function fails to adequately capture the characteristics of PSO. Consequently, researchers have sought to address these challenges by modifying the velocity updating formula or introducing alternative transfer functions. Numerous strategies have been employed and various new transfer functions were designed. While these strategies have enhanced the performance of BPSO in certain aspects, they have also introduced additional algorithmic complexity and increased computational overhead.
To address the prevailing challenges related to transfer functions and computational costs in existing algorithms, this study introduces a variant of Binary Particle Swarm Optimization called Binary Restructuring Particle Swarm Optimization (BRPSO). The BRPSO algorithm is developed based on the restructuring PSO (RPSO) framework proposed in 2022 [7]. By eliminating the use of transfer functions, BRPSO achieves a streamlined algorithmic flow and enhances computational efficiency.
The main contributions of this study can be summarized as follows.
  • In this paper, a binary variant of the Restructuring Particle Swarm Optimization (RPSO) algorithm is proposed, aiming to extend the capabilities of RPSO towards addressing binary optimization problems.
  • The proposed approach eliminates the incorporation of both the velocity updating formula and the transfer function commonly utilized in traditional BPSO. This removal serves to eliminate the influence of the transfer function and simplify the algorithmic flow.
  • The introduction of a novel perturbation term in BRPSO presents an advancement that not only reduces the number of required parameters but also satisfies diverse requirements across different stages of evolution.
The remainder of this paper is structured as follows: Section 2 introduces the related work of this paper which includes feature selection and RPSO. The proposed binary RPSO is presented in Section 3. In Section 4, the experimental results and discussion are given. Lastly, concluding remarks are provided in Section 5.

2. Related Work

In this section, the development of BPSO will be reviewed first. Following this, the concept of feature selection is introduced as the test problem. Furthermore, the Restructuring Particle Swarm Optimization algorithm for continuous optimization problems is introduced at the end.

2.1. Review of BPSO

The improvement of BPSO can be broadly categorized into two avenues: the modification of the transfer function and the improved updating strategies in velocity.

2.1.1. Reviews of Transfer Function

The original BPSO algorithm employs the sigmoid function as the transfer function for velocity transformation. However, this choice of transfer function does not align with the fundamental characteristics of PSO. In PSO, a large absolute value of velocity means the particle should take greater movement to search the optimal solution. Conversely, a small absolute velocity value indicates that particle has approached the optimal solution and needs small movement. Nevertheless, in the original BPSO algorithm, a negative large velocity is likely to result in a next position of 0, while a positive large velocity is prone to yield a next position of 1. Otherwise, the next position takes 0 or 1 with a probability 0.5 when the velocity takes 0.
In 2008, a probability binary PSO (PBPSO) algorithm was proposed by Wang et al. [8]. PBPSO incorporates a novel linear transfer function and a set of rules to generate new positions. Specifically, it utilizes the continuous search space position to update the binary position. The detailed updating rule is provided below.
T F x i d ( t + 1 ) = x i d ( t + 1 ) R min R max R min
x b i d ( t + 1 ) = 1 if rand ( ) TF x i d ( t + 1 ) 0 if rand ( ) > TF x i d ( t + 1 )
where [Rmin, Rmax] is a predefined range to transfer continuous position into the probability value. However, the linear transfer function also faced the same drawbacks as S-shape transfer function. To overcome the shortcoming of inconsistency between S-shape transfer function and PSO characteristics, Nezamabadi-pour et al. [9] proposed a new discrete binary PSO (NBPSO) which used a new transfer function described as follows.
T F V i d = tanh α v i d
x i d ( t + 1 ) = exchange x i d ( t ) if T F v i d ( t + 1 ) > rand ( ) x i d ( t ) if T F v i d ( t + 1 ) rand ( ) .
where α is a constant value. When the next position will change when the transfer value is greater than a random number in the range [0, 1]. On the contrary, it remains the same. The use of V-shape transfer function and new updating rules avoid the conflicts between the transfer function and PSO. However, the global exploration ability of NBPSO is still insufficient. When the position is a local optimal solution and the velocity approaches zero, the algorithm hardly escapes the local optimal space when using V-shape transfer function.
Mirjalili and Lewis [10] proposed six modified transfer functions and divided them into two categories: S-shape and V-shape. Through the comparison on 25 benchmark functions among six modified transfer functions, the results that the proposed V-shape transfer function performs better than the S-shape transfer function are obtained.
In order to address the limitations of existing transfer functions in providing adequate exploration and exploitation capabilities for BPSO, Md. Jakirul Islam et al. [11] conducted research on the exploration and exploitation stages of BPSO, as well as analyzed the behavior of existing S-shape, V-shape, and linear transfer functions. They raised a time-varying transfer function which provided a high flipping probability of particles in the early stages to ensure the exploration of search space and a low flipping probability in the final stages to exploit the solution space. However, it faced the same drawbacks as the traditional S-shape transfer function.
Different shape transfer function gives BPSO distinct search ability, so Mirjalili et al. [12] proposed a novel U-shape transfer function with multiple alterable parameters, and applied it in BPSO. Through the experiments in a set of benchmark functions and 0/1 knapsack problems, the following U-shape transfer function had been proved performances better than S-shape and V-shape transfer functions.
T F V i d = α | x β |
where α and β are two parameters that control the slope and width of the transfer function. However, U-shape transfer function has the similar disadvantage as V-shape transfer function.
In order to address the issue of BPSO algorithms being prone to local optima, Guo et al. [13] proposed an asymmetric mapping function named Z-shape transfer function to map the continuous search space to a binary space. Nine typical benchmark functions are adopted to test the effectiveness of Z-shape, S-shape, and V-shape transfer functions. Furthermore, the raised Z-shape transfer function achieved the best performance. The Z-shape transfer function is defined as follows.
T F v i d ( t ) = 1 a v i d ( t )
where α is the parameter to control the slope of the transfer function. Because BPSO easily stuck into the local optimal solution or converged at a slower speed when using the exitied transfer function, Beheshti [14] raised a novel X-shape transfer function to enhance the exploration and exploitation ability of BPSO (XBPSO) in the discrete binary search space.
T F 1 v i d ( t + 1 ) = v i d ( t + 1 ) 1 + v i d ( t + 1 ) × 0.5 + 0.5
y i d = 1 if rand 1 > T F 1 v i d ( t + 1 ) 0 if rand 1 T F 1 v i d ( t + 1 )
T F 2 v i d ( t + 1 ) = v i d ( t + 1 ) 1 1 + v i d ( t + 1 ) 1 0.5 + 0.5
z i d = 1 if rand 2 < T F 2 v i d ( t + 1 ) 0 if rand 2 T F 2 v i d ( t + 1 )
P i ( t + 1 ) = y i if f y i is better than f z i z i if f z i is better than f y i
In XBPSO, two positions y i and z i were first generated by two transfer functions. Furthermore, the better one is selected as p i ( t + 1 ) . If the fitness value of p i ( t + 1 ) is better than the x i ( t ) , x i ( t + 1 ) will be p i ( t + 1 ) . Otherwise, XBPSO will generate two new positions by crossing operate pi(t + 1) and x i ( t ) , and select the better one to be x i ( t + 1 ) . Through the experiments, the effectiveness of the XBPSO had been demonstrated. However, it should be noted that XBPSO used at least twice the number of function evaluations in a generation. Furthermore, the computation cost of XBPSO is large. In the same year, Beheshti [15] proposed an upgrade transfer function (UTF) and applied it in BPSO. UTF also generates at least twice the number of solutions in each generation. Multiple fitness function evaluation times will be used in one generation in both X-shape and UTF.
Various new transfer functions were proposed in recent years, but they all exhibit different shortcomings. S-shape and linear-shape transfer functions do not fit the characteristics of the algorithm well. V-shape and U-shape transfer functions do not help the algorithm escape from the local optimal solution when the velocity approaches toward 0. X-shape and UTF require several fitness evaluations in one iteration.

2.1.2. Reviews of Improvement in Updating Strategies of BPSO

The improvement in particle updating strategies focuses on parameter control and the modified learning models.
Liu et al. [16] used a mathematical model to analyze the role inertial weight was played in BPSO. Furthermore, obtained the conclusion that, when two acceleration factors are set to constant, a small inertia weight provides better exploration capability and a large inertia weight enhances exploitation. Furthermore, they proposed linearly increasing inertia weight BPSO(LIWBPSO), the setting of inertial weight is described as follows.
w = w + π × ( w ¯ w ̲ ) ρ × π ¯ , if π ρ × π ¯ w ¯ if ρ × π ¯ < π π ¯
where π and π ¯ are the number of current iteration and the maximal number of iterations, respectively. w and w ¯ stand for the lower and upper bounds of w. ρ is a parameter that controls the increasing rate of w. The experiments adopted on 0–1 knapsack problem demonstrate the effectiveness of the raised increasing inertia weight. However, this research improved BPSO by only considering the influence of inertial weight.
Aiming to cope with premature convergence of the BPSO, Vieira et al. [17] put forward a modified BPSO (MBPSO) which used the mutation mechanism to avoid premature convergence of BPSO. Furthermore, Mingo López et al. [18] raised a hybrid genetic inspired BPSO (HPSOGO) which introduced crossover and mutation operations into BPSO. The personal historical best position and the global best position are used to crossover/mutation with the current position to generate the next position in HPSOGO.
Based on bare bones particle swarm optimization [19], Zhang et al. [20] first proposed the binary bare bones particle swarm optimization (BBPSO) and used it to solve the feature selection problem. BBPSO used a reinforced memory strategy to update the personal best historical position archive and the Uniform combination was adopted to balance the exploration and exploitation of BBPSO. In order to enable BPSO to better solve feature selection problems, Song et al. [21] raised a novel BBPSO with mutual information (MIBBPSO). In MIBBPSO, the correlation between features and class labels is used to initial the position to improve the convergence rate of algorithms. The supplementary operator and the deletion operator are designed to enhance the local search ability. An adaptive flip mutation operator is developed to balance the global search and local development in MIBBPSO.
A hybrid Taguchi binary particle swarm optimization (HTBPSO) was proposed by Jia et al. [22]. HTBPSO used the Taguchi method to enhance the local exploitation of BPSO and used catfish effect operation in the last 10% worse particle positions to avoid premature convergence.
Ji et al. [23] introduced three novel factors to improve BPSO (IBPSO) based on Lévy flight, weighting inertia coefficient, and mutation mechanism, respectively. New factors improve the performance of the algorithm in exploitation, exploration, and maintaining population diversity.
Hu et al. [24] raised a multi-surrogate assisted BPSO(MSABPSO) which used the multi-surrogate multi-swarm model to improve the convergence ability of binary PSO. Furthermore, a new dynamic S-shape transfer function was adopted to balance the abilities of exportation and exploitation for the MSABPSO. The experimental results on benchmark functions and feature selection problems showed that MSABPSO significantly improved prediction accuracy. However, it incurred greater costs associated with computational resources.
Among the literature mentioned above, various strategies were employed to enhance some capabilities of BPSO. However, there are some aspects they unconsidered. Some improvement strategies only considered the effect of a single variable. Some strategies combined with the other evolutional algorithms, increased the algorithm complexity and computation cost. Moreover, the use of some problem-specific improvement strategies limited the scalability of the algorithm to solve other problems.

2.2. Feature Selection

Feature selection is a data pre-processing strategy. It is a technique for extracting key features from data where it performs a series of complex calculations on a dataset to achieve the effect of eliminating irrelevant, redundant, and misleading features. Feature selection aims to decrease the number of features, which consequently results in a reduction in the duration of model training and computational expenses. Furthermore, it can enhance the performance of the model. Feature selection is widely used in pattern recognition and machine learning [25]. Feature selection can be roughly categorized into three groups based on the strategies employed: filter, wrapper and embedded [26]. The embedded methods embed feature selection into the learner training process. The filter techniques select relevant features purely based on the inherent relationships among the features, without requiring the integration of any learning method [27,28]. The Wrapper methods combine a search strategy with a learning method to identify the optimal subset of features for classification performance by utilizing a subset of features obtained through the search strategy to train a classifier [29,30]. The use of classifiers to train and evaluate each feature subset results in high computational overhead for the wrapper method. However, in the wrapper method, the classifier performance is directly used to evaluate the importance of features, which allows the wrapper method to achieve better classification performance than the filter technique. Due to the simplified implementation and low computation cost, BPSO is popularly used in the wrapper method. BPSO-based wrapper achieved significant performance in feature selection problems [31,32,33,34].

2.3. Restructuring Particle Swarm Optimization

Based on the theory analysis results of PSO which used linear system theory, Zhu et al. [24] proposed a novel structure of PSO named Restructuring Particle Swarm Optimization algorithm (RPSO). RPSO used only one formula to update the particles and took fewer parameters which simplified the algorithm. The specific formula is described as Equation (17)
x i t = r × p b e s t i + ( 1 r ) × gbest + c × rand × ( 0.8 rand + w ) ( 1 t / T ) + c × rand × ( 0.8 rand + w ) ( 1 t / T )
where x represents the position of particle i in iteration t. r takes random number between 0 and 1. w linear decreasing from 0.8 to 0.2 with iterations. t is current iteration and T marks the maximum iteration.
Compared to the canonical PSO, RPSO has a simpler structure and fewer parameters. However, RPSO is proposed to solve the continuous optimization problem. There is no literature to use the BPSO to solve problems with binary variables. The setting of perturbation also takes many parameters, which seems complex.

3. Proposed Approach

In this section, the motivation behind our study is presented, followed by a detailed description of our approach and its corresponding flowchart. Finally, an explanation of the perturbation item’s configuration is provided.

3.1. Motivation

In recent years, numerous improved BPSO algorithms have been proposed, with a prevalent approach being the utilization of transfer functions to map velocity values onto the [0, 1] interval. Previous studies employed probability-based particle updates, guided by the mapped velocity values. The use of the transfer function complicated the algorithm flow. Moreover, the design of the transfer function plays a crucial role in determining the algorithm’s performance. A well-designed transfer function, aligned with the inherent characteristics of the algorithm, can enhance its performance. Conversely, an ill-suited transfer function can significantly impair the algorithm’s effectiveness. The transfer functions currently in use exhibit certain shortcomings, as discussed in Section 2.1.1. Furthermore, the improvement strategies presented in Section 2.1.2 encounter challenges related to the increased complexity of the algorithm.
In order to mitigate the impact of the transfer function and streamline the algorithmic process, this study introduces a binary variant of RPSO (BRPSO) that eliminates the use of a transfer function. The primary objective of BRPSO is to address binary optimization problems.

3.2. Binary Restructuring Particle Swarm Optimization

While RPSO is designed to tackle continuous optimization problems, binary optimization problems possess distinct solution spaces. To address this disparity, the Binary Restructuring Particle Swarm Optimization (BRPSO) is specifically developed for binary optimization problems. In this subsection, BRPSO is presented. It is a proposed method for addressing binary optimization problems.
In RPSO, the particle updating process is governed by a single formula influenced by the cumulative weights of the personal best ( p b e s t ) and global best ( g b e s t ) solutions, along with a perturbation term that diminishes with each iteration. Building upon these findings, the position update equation in BRPSO is expressed as Equation (18).
x i = r 1 × p b e s t i + 1 r 1 × gbest + p
x i = 1 , if x i > rand x i = 0 , if x i rand
where r 1 takes a random number in [0, 1] interval and changes with iteration. p b e s t i is i-th particle own best historical position and g b e s t is the global best position. p represents a decreasing perturbation term that eventually converges to 0. The detailed formula of p will propose in the next subsection. With the iteration process, the value obtained by Equation (18) will converge from [−1, 2] interval to [0, 1] interval which has the same expected value of [0, 1] interval. So, the final value of x i is “0” or “1” in each iteration is determined by comparing it with a random number in [0, 1] interval. Above updating formula can good substitutes the role of the transfer function.
The pseudocode of BRPSO framework is provided in Algorithm 1 and the flowchart is shown in Figure 1.
Algorithm 1 BRPSO
Require: maximum iteration, population size, dimension
1:
for each particle i do
2:
    Initialize position x i for particle i;
3:
    Evaluate f ( x i ) of particle i and set p i = x i ;
4:
end for
5:
p g = m i n ( p i )
6:
while iteration < Maximum iteration do
7:
    for  i = 1 to N do
8:
        Update position of particle i with Equations (18) and (19)
9:
        Evaluate particle i
10:
        if  f ( x i ) < f ( p i )  then
11:
            p i = x i
12:
        end if
13:
        if  f ( x i ) < f i t ( p g )  then
14:
            p g = p i
15:
        end if
16:
    end for
17:
     i t e r a t i o n = i t e r a t i o n + 1
18:
end while
Ensure:  p g

3.3. Perturbation p

According to the theoretical idea of RPSO, appropriate perturbation can effectively improve the search ability of the algorithm. The addition of perturbation terms can significantly enhance the randomness of the algorithm in the early evolutionary stage which can better search the solution space. The existence of perturbation can avoid particles from falling into the local optimal solution and converging prematurely in the early stage. As the iteration proceeds, the perturbation item p gradually tends to 0 to ensure the convergence of the algorithm at the later stage. The detailed formula of p is described as follows.
p = r 2 × ( ( cos ( t / T × π ) + 1 ) / 2
where r 2 takes a random number in the interval [−1, 1] and changes with iteration to ensure both forward and reverse perturbation exists. t is current iteration and T is the maximum iteration. The trend of p-value changes with iteration is shown in Figure 2.
Figure 2 illustrates the perturbation value p will converge to 0 with the iteration process, which meets the requirements of the setting rule of perturbation. During the iteration process, p takes positive or negative randomly in Figure 2 which improves the search ability of the algorithm.

4. Experimental Section

In this section, a series of tests involving feature selection problems is conducted to evaluate the performance the proposed approach. The problem definition of feature selection, along with the corresponding fitness function employed, is initially introduced in this section. Subsequently, the experimental setup and a comprehensive overview of the obtained results are presented. Finally, a detailed analysis of the convergence and stability of our proposed approach is conducted.

4.1. Problem Definition

Feature selection is a typical NP-hard combinatorial optimization problem. It is a process to choose the important feature subset in the original set of features [35]. Assuming that the dataset has n features, it may exist 2 n potential feature subset solutions. Furthermore, feature selection is to select the important subset from 2 n feature subset which can provide high classification accuracy. However, the number of selected features is also important.

4.2. Fitness Function

The fitness function plays important an role in PSO-based wrapper technique for feature selection. Classification accuracy and the number of selected features are two important valuation indicators. Therefore, the fitness function used in this paper includes the classification error rate and the number of selected features. The specific definition of the function is described as follows.
fitness = α × error + ( 1 α ) S F T F
where α controls the importance of classification quality and takes a number in [0, 1] interval. error represents the classification error rate obtained by the used classifier. SF and T F are the numbers of selected features and total features, respectively.

4.3. Experimental Set

A total of 15 datasets from the UCI machine learning repository [36] are used to test the effectiveness of BRPSO. Details about these datasets are shown in Table 1. The k-nearest-neighbor (kNN) classifier with k = 5 [37] is adopted as the classification model to test the quality of the feature subset. Furthermore, the n-fold cross-validation (with n = 10) is used on training data. α in fitness function is set to 0.99 to ensure classification accuracy.
The original BPSO [3], NBPSO [8], LIWBPSO [15], and BGWOPSO [38] are selected to benchmark with BRPSO to test the effectiveness of BRPSO. Referring to the existing literature [39], the population size and the maximum number of iterations are set to 20 and 100, respectively. The problem dimension is equal to the number of features in the dataset. Each algorithm independently runs 20 times to test the stability of algorithms. All peer algorithms are implemented by MATLAB-R2021b and tested on a computer with an AMD Ryzen 7 4800H CPU @ 2.90 GHz and 16.0G GB of RAM. The operating system running on this computer is Windows 10 (×64).

4.4. Experiment Results

4.4.1. Numerical Comparison Results

The performance results of all algorithms were statistically tested in this work, utilizing T-test and Friedman-test to test the performance of each method, respectively. The T-test confidence interval is set to 95%. The symbol “+” indicates that BRPSO outperforms other algorithms, “−” indicates that BRPSO performs worse than other algorithms, and “=” indicates that BRPSO and the other algorithms perform comparably.
The results obtained by BRPSO on several benchmark datasets were compared with the average classification results of peer algorithms, as shown in Table 2. From Table 2, it can be seen that the results produced by BRPSO can achieve the best average classification accuracy among the 5 algorithms on 11 datasets, thus showing the superiority of BRPSO in feature selection for accurate classification. In addition, T-test results further clearly demonstrate the performance comparison results of BRPSO with other algorithms. It can be clearly seen that BRPSO achieves better results than BGWOPSO on all datasets. Furthermore, BRPSO achieves better classification accuracy than NBPSO, LIWPSO, and BPSO on more than half of the datasets.
In addition, we also compared the number of features selected by all algorithms, as shown in Table 3. From Table 3, it can be seen that BRPSO can achieve better classification accuracy with the smallest number of features that can be selected on 13 datasets. On two datasets with more than 200 features, namely LSVT Voice Rehabilitation, and arrhythmia, the other algorithms need to select more than 100 features to achieve better accuracy. In contrast, BRPSO only needs to select less than 100 features to achieve better results. This indicates that BRPSO is able to select features that are more conducive to classification accuracy. In addition, on datasets with more classification categories, such as arrhythmia and LibrasMovemen, which have 16 and 15 categories, respectively, BRPSO can select the smallest number of features to achieve the best classification accuracy. This indicates the superiority of BRPSO in dealing with complex classification problems.
To compare the comprehensive performance of all algorithms, the fitness of all algorithms was also calculated according to Equation (21). The results are shown in Table 4. From Table 4, it can be seen that BRPSO achieved the best comprehensive performance on 11 data sets. Although BRPSO did not perform as well as LIWPSO on BreastCancer, SPECT, Tic-Tac-Toe Endgame, and BreastEW, BRPSO was able to obtain better classification accuracy or select fewer features than LIWPSO on these datasets. From the t-test results, BRPSO obtained better performance than BGWOPSO on all datasets. In addition, better fitness values than BPSO, NBPSO, and LIWPSO are also obtained on more than half of the datasets.

4.4.2. Statistical Test Results

In addition, we also summarize the statistical tests for all the experimental results. A summary of the results of T-test results for each strategy in Table 2, Table 3 and Table 4 is done to make an overall comparison of the performance of all algorithms. The comprehensive performance values in Table 5 are “+” minus “−” for each metric. Table 6 shows the results of the Friedman test for all algorithms on error, select, and fitness. Where error is 1 minus accuracy.
The results in Table 5 show that all the CP results of BRPSO outperform the other algorithms. In particular, the CP values compared with BGWOPSO show that BGWOPSO did not perform better than BPSO on any of the datasets. BRPSO also clearly performs better on more evaluation metrics on multiple datasets compared to the other three algorithms.
In this part, the Friedman test is utilized to demonstrate the comprehensive performance of the 5 peer algorithms. The Friedman test necessitates rating K algorithms on each dataset based on their absolute value. The best performing algorithm is ranked 1, while the worst performing algorithm is ranked K. As can be seen from Table 6, BRPSO achieves the optimal performance on all three evaluation metrics, further demonstrating the superior performance of BRPSO in solving the feature selection problem.

4.5. Convergence and Stability Analysis

The convergence figures among five peer algorithms in eight datasets are shown in Figure 3. It can be found that BRPSO exhibits fast convergence speed on Heart, Ionosphere and Sonar datasets. In other five datasets, BRPSO displayed strong capacity in jumping out the local optimal solution and reached the best fitness value. The comprehensive capacity of BRPSO in convergence achieved favorite performance as shown in Figure 3.
The boxplot is used in this subsection to reveal the stability of five test algorithms. The boxplot which includes part of the results among five benchmark algorithms in 20 running times is displayed in Figure 4. The results shown in Figure 4 demonstrate that BRPSO obtained the high-quality best solution sets. As seen from Figure 4, the results gained by BRPSO in 20 running times are approaching which approved the stability of BRPSO.

5. Conclusions

This paper introduces a binary variant of Restructuring Particle Swarm Optimization (RPSO) specifically designed for discrete binary optimization problems. The proposed approach, known as Binary RPSO (BRPSO), eliminates the velocity updating formula and transfer function employed in Binary Particle Swarm Optimization (BPSO), resulting in a reduction in parameters and computational cost. BRPSO leverages historical experience from individual particles and the swarm to ensure its convergence capabilities. Furthermore, the integration of a decreasing perturbation factor effectively balances the exploration and exploitation abilities of BRPSO. To assess the performance of BRPSO, a comprehensive set of comparison experiments is conducted to solve feature selection problems. The experimental results, obtained by evaluating BRPSO against four benchmark algorithms across fifteen datasets, demonstrate its superior performance.
It is important to note that while our approach serves as a general algorithm for binary optimization problems, its performance may not outperform task-specific algorithms in certain scenarios. Nonetheless, this limitation also ensures the scalability of BRPSO for larger or more complex scenarios when combined with specific strategies. In future research, we anticipate potential modifications to the learning model and disturbance settings to address the unique challenges presented by real-world problems. Particular attention will be given to striking the right balance between exploration and exploitation capabilities, aligning with the demands of diverse problem domains.

Author Contributions

Conceptualization, J.Z. and J.L.; methodology, J.Z. and J.L.; software, J.Z.; validation, J.Z., Y.C. and J.L.; formal analysis, J.Z. and J.L.; investigation, J.Z.; resources, J.Z.; data curation, J.Z.; writing—original draft preparation, J.Z.; writing—review and editing, J.Z. and J.L.; visualization, J.Z.; supervision, J.L.; funding acquisition, J.L., X.X. and S.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Natural Science Foundation of China, grant number 62172095 and Fujian University of Technology Development Fund, grant number GY-Z20046.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kennedy, J.; Eberhart, R. Particle swarm optimization. In Proceedings of the ICNN’95—International Conference on Neural Networks, Perth, WA, Australia, 27 November–1 December 1995. [Google Scholar]
  2. Shi, Y.; Eberhart, R. A modified particle swarm optimizer. In Proceedings of the 1998 IEEE International Conference on Evolutionary Computation Proceedings—IEEE World Congress on Computational Intelligence, Anchorage, AK, USA, 4–9 May 1998. [Google Scholar]
  3. Kennedy, J.; Eberhart, R.C. A discrete binary version of the particle swarm algorithm. In Proceedings of the 1997 IEEE International Conference on Systems, Man, and Cybernetics, Computational Cybernetics and Simulation, Orlando, FL, USA, 12–15 October 1997. [Google Scholar]
  4. Bansal, J.C.; Deep, K. A Modified Binary Particle Swarm Optimization for Knapsack Problems. Appl. Math. Comput. 2012, 218, 11042–11061. [Google Scholar] [CrossRef]
  5. Lin, G.; Guan, J.; Li, Z.Y.; Feng, H.B. A hybrid binary particle swarm optimization with tabu search for the set-union knapsack problem. Expert Syst. Appl. 2019, 135, 201–211. [Google Scholar] [CrossRef]
  6. Lin, J.C.W.; Yang, L.; Fournier-Viger, P.; Hong, T.-P.; Voznak, M. A binary PSO approach to mine high-utility itemsets. Soft Comput. 2017, 21, 5103–5121. [Google Scholar] [CrossRef]
  7. Zhu, J.; Liu, J.H.; Wang, Z.H.; Chen, Y.X. Restructuring Particle Swarm Optimization algorithm based on linear system theory. In Proceedings of the 2022 IEEE Congress on Evolutionary Computation, Padua, Italy, 18–23 July 2022. [Google Scholar]
  8. Wang, L.; Wang, X.; Jingqi, F.; Zhen, L. A novel probability binary particle swarm optimization algorithm and its application. J. Softw. 2008, 3, 28–35. [Google Scholar] [CrossRef]
  9. Nezamabadi-pour, H.; Maghfoori-Farsangi, M. Binary particle swarm optimization: Challenges and new solutions. J. Comput. Sci. Eng. 2008, 6, 21–32. [Google Scholar]
  10. Seyedali, M.; Andrew, L. S-shaped versus V-shaped transfer functions for binary Particle Swarm Optimization. Swarm Evol. Comput. 2013, 9, 1–14. [Google Scholar]
  11. Jakirul Islam, M.; Li, X.; Mei, Y. A time-varying transfer function for balancing the exploration and exploitation ability of a binary PSO. Appl. Soft Comput. 2017, 59, 182–196. [Google Scholar] [CrossRef]
  12. Mirjalili, S.; Zhang, H.Y.; Mirjalili, S.; Chalup, S.; Noman, N. A Novel U-Shaped Transfer Function for Binary Particle Swarm Optimisation. In Proceedings of the Soft Computing for Problem Solving 2019, Liverpool, UK, 2–4 September 2019. [Google Scholar]
  13. Guo, S.; Wang, J.; Guo, M. Z-shaped transfer functions for binary particle swarm optimization algorithm. Comput. Intell. Neurosci. 2020, 59, 21. [Google Scholar] [CrossRef]
  14. Beheshti, Z. A novel x-shaped binary particle swarm optimization. Soft Comput. 2021, 25, 3013–3042. [Google Scholar] [CrossRef]
  15. Zahra, B. UTF: Upgrade transfer function for binary meta-heuristic algorithms. Appl. Soft Comput. 2021, 106, 107346. [Google Scholar]
  16. Liu, J.H.; Mei, Y.; Li, X.D. An Analysis of the Inertia Weight Parameter for Binary Particle Swarm Optimization. IEEE Trans. Evol. Comput. 2016, 20, 666–681. [Google Scholar] [CrossRef]
  17. Susana, M.V.; Luis, F.M.; Gonçalo, J.F.; Joao, M.C.S. Modified binary PSO for feature selection using SVM applied to mortality prediction of septic patients. Appl. Soft Comput. 2013, 13, 3494–3504. [Google Scholar]
  18. Mingo, L.L.F.; Gomez Blas, N.; Arteta, A. Multidimensional knapsack problem optimization using a binary particle swarm model with genetic operations. Soft Comput. 2018, 22, 2567–2582. [Google Scholar] [CrossRef]
  19. Kennedy, J. Bare bones particle swarms. In Proceedings of the 2003 IEEE Swarm Intelligence Symposium—SIS’03, Indianapolis, IN, USA, 5 June 2003. [Google Scholar]
  20. Zhang, Y.; Gong, D.W.; Hu, Y.; Zhang, W.Q. Feature selection algorithm based on bare bones particle swarm optimization. Neurocomputing 2015, 148, 150–157. [Google Scholar] [CrossRef]
  21. Song, X.F.; Zhang, Y.; Gong, D.W.; Sun, X.Y. Feature selection using bare-bones particle swarm optimization with mutual information. Pattern Recognit. Lett. 2021, 112, 107804. [Google Scholar] [CrossRef]
  22. Jia, X.N.; Lu, G.Z. A Hybrid Taguchi Binary Particle Swarm Optimization for Antenna Designs. IEEE Antennas Wirel. Propag. Lett. 2019, 18, 1581–1585. [Google Scholar] [CrossRef]
  23. Ji, B.; Lu, X.Z.; Sun, G.; Zhang, W.; Li, J.H.; Xiao, Y.Z. Bio-Inspired Feature Selection: An Improved Binary Particle Swarm Optimization Approach. IEEE Access 2020, 8, 85989–86002. [Google Scholar] [CrossRef]
  24. Hu, P.; Pan, J.S.; Chu, S.C.; Sun, C.L. Multi-surrogate assisted binary particle swarm optimization algorithm and its application for feature selection. Appl. Soft Comput. 2022, 121, 108736. [Google Scholar] [CrossRef]
  25. Cai, J.; Luo, J.W.; Wang, S.L.; Yang, S. Feature selection in machine learning: A new perspective. Neurocomputing 2018, 300, 70–79. [Google Scholar] [CrossRef]
  26. Rama, K.E.; Nagaraju, D. Binary Golden Eagle Optimizer with Time-Varying Flight Length for feature selection. Knowl. Syst. 2022, 247, 108771. [Google Scholar]
  27. Guha, R.; Ghosh, M.; Mutsuddi, S.; Sarkar, R.; Mirjalili, S. Embedded chaotic whale survival algorithm for filter–wrapper feature selection. Soft Comput. 2020, 24, 12821–12843. [Google Scholar] [CrossRef]
  28. Qaraad, M.; Amjad, S.; Hussein, N.K.; Elhosseini, M.A. Large scale salp-based grey wolf optimization for feature selection and global optimization. Neural Comput. Appl. 2022, 34, 8989–9014. [Google Scholar] [CrossRef]
  29. Qaraad, M.; Amjad, S.; Hussein, N.K.; Elhosseini, M.A. Addressing constrained engineering problems and feature selection with a time-based leadership salp-based algorithm with competitive learning. J. Comput. Des. Eng. 2022, 9, 2235–2270. [Google Scholar] [CrossRef]
  30. Ma, B.J.; Liu, S.; Heidari, A.A. Multi-strategy ensemble binary hunger games search for feature selection. Knowl. Syst. 2022, 248, 108787. [Google Scholar] [CrossRef]
  31. Xue, B.; Nauyen, S.; Zhang, M.J. A New Binary Particle Swarm Optimisation Algorithm for Feature Selection. In Proceedings of the Applications of Evolutionary Computation, Granada, Spain, 23–25 April 2014. [Google Scholar]
  32. Shang, L.; Zhou, Z.; Liu, X. Particle swarm optimization-based feature selection in sentiment classification. Soft Comput. 2016, 20, 3821–3834. [Google Scholar] [CrossRef]
  33. Chen, K.; Zhou, F.Y.; Yuan, X.F. Hybrid particle swarm optimization with spiral-shaped mechanism for feature selection. Expert Syst. Appl. 2019, 128, 140–156. [Google Scholar] [CrossRef]
  34. Hu, Y.; Zhang, Y.; Gong, D.W. Multiobjective Particle Swarm Optimization for Feature Selection With Fuzzy Cost. IEEE Trans. Cybern. 2021, 51, 874–888. [Google Scholar] [CrossRef]
  35. Liu, H.; Yu, L. Toward integrating feature selection algorithms for classification and clustering. IEEE Trans. Knowl. Data Eng. 2005, 17, 491–502. [Google Scholar]
  36. Dua, D.; Graff, C. Machine Learning Repository; University of California, Irvine, School of Information and Computer Sciences: Irvine, CA, USA, 2017. [Google Scholar]
  37. Aha, D.W.; Kibler, D.; Albert, M.K. Instance-based learning algorithms. Mach. Learn. 1991, 6, 37–66. [Google Scholar] [CrossRef] [Green Version]
  38. Al-Tashi, Q.; Abdul, K.; Said, J.; Rais, H.M.; Mirjalili, S.; Alhussian, H. Binary Optimization Using Hybrid Grey Wolf Optimization for Feature Selection. IEEE Access 2019, 7, 39496–39508. [Google Scholar] [CrossRef]
  39. Thaher, T.; Chantar, H.; Too, J.W.; Mafarja, M.; Turabieh, H.; Houssein, E.H. Boolean Particle Swarm Optimization with various Evolutionary Population Dynamics approaches for feature selection problems. Expert Syst. Appl. 2022, 195, 116550. [Google Scholar] [CrossRef]
Figure 1. The flowchart of BRPSO.
Figure 1. The flowchart of BRPSO.
Biomimetics 08 00266 g001
Figure 2. The change trend of p.
Figure 2. The change trend of p.
Biomimetics 08 00266 g002
Figure 3. Convergence graph of each algorithm.
Figure 3. Convergence graph of each algorithm.
Biomimetics 08 00266 g003aBiomimetics 08 00266 g003b
Figure 4. Boxplot of each algorithm.
Figure 4. Boxplot of each algorithm.
Biomimetics 08 00266 g004aBiomimetics 08 00266 g004b
Table 1. Datasets description.
Table 1. Datasets description.
DatasetNo. AttributesNo. InstanceNo. Class
BreastCancer96992
Heart142702
ionosphere343512
Sonar602082
SPECT222672
Waveform415003
Tic-Tac-Toe Endgame99582
Hill10012122
Urban1476759
Vehicle188464
Musk version 11684762
LSVT Voice Rehabilitation3091262
BreastEW305962
arrhythmia27945216
LibrasMovemen9036015
Table 2. Results on accuracy.
Table 2. Results on accuracy.
BRPSOBPSONBPSOLIWPSOBGWOPSO
BreastCancer98.798.8(=)98.7(=)98.8(−)98.5(+)
Heart95.294.9(=)94.1(+)94.9(=)93.1(+)
ionosphere98.597.2(+)97.5(+)97.9(+)95.8(+)
Sonar10099.6(+)99.6(+)99.7(+)98.6(+)
SPECT92.592.2(=)91.9(+)92.8(−)90.7(+)
Waveform86.785.1(+)85.5(+)85.7(+)85.9(+)
Tic-Tac-Toe Endgame84.284.6(−)83.2(+)84.6(−)83.4(+)
Hill70.864.5(+)65.9(+)65.7(+)67.6(+)
Urban92.890.1(+)90.7(+)90.8(+)90.4(+)
Vehicle82.181.7(=)81.3(+)82.0(=)80.3(+)
Musk version 198.796.7(+)97.2(+)97.5(+)96.7(+)
LSVT Voice Rehabilitation97.496.5(+)96.5(+)97.0(=)95.3(+)
BreastEW99.699.5(=)99.5(=)99.7(=)99.2(+)
arrythmia76.572.6(+)73.6(+)73.8(+)73.6(+)
LibrasMovemen90.988.1(+)88.5(+)88.9(+)87.1(+)
Table 3. Results on select.
Table 3. Results on select.
BRPSOBPSONBPSOLIWPSOBGWOPSO
BreastCancer2.9753.175(+)3.16(+)3.18(+)3.615(+)
Heart4.8154.8(=)4.865(=)4.78(=)5.795(+)
ionosphere4.4257.935(+)6.39(+)5.985(+)9.15(+)
Sonar9.6818.985(+)16.5(+)16.625(+)18.75(+)
SPECT7.047.555(+)7.365(+)7.235(=)8.605(+)
Waveform21.1921.745(+)21.79(+)21.565(=)27.63(+)
Tic-Tac-Toe Endgame7.5157.695(=)6.76(−)7.645(=)7.31(=)
Hill39.15544.72(+)43.55(+)43.465(+)57.86(+)
Urban48.3465.23(+)62.14(+)62.595(+)67.59(+)
Vehicle8.518.805(=)8.685(+)8.555(=)10.45(+)
Musk version148.28573.22(+)69.075(+)69.245(+)66.78(+)
LSVT Voice Rehabilitation64.98119.315(+)107.325(+)108.115(+)107.91(+)
BreastEW4.1456.36(+)5.58(+)5.205(+)7.575(+)
arrhythmia94.925125.975(+)120.635(+)123.41(+)122.575(+)
LibrasMovemen21.834.79(+)31.245(+)31.9(+)31.84(+)
Table 4. Results on fitness.
Table 4. Results on fitness.
BRPSOBPSONBPSOLIWPSOBGWOPSO
BreastCancer1.58  × 10 2 1.51  × 10 2 (=)1.63  × 10 2 (=)1.50  × 10 2 (=)1.86  × 10 2 (+)
Heart5.08  × 10 2 5.47  × 10 2 (=)6.17  × 10 2 (+)5.45  × 10 2 (=)7.25  × 10 2 (+)
ionosphere1.64  × 10 2 3.00  × 10 2 (+)2.71  × 10 2 (+)2.24  × 10 2 (+)4.42  × 10 2 (+)
Sonar2.08  × 10 3 7.42  × 10 3 (+)6.32  × 10 3 (+)5.36  × 10 3 (+)1.67  × 10 2 (+)
SPECT7.76  × 10 2 8.11  × 10 2 (=)8.34  × 10 2 (+)7.42  × 10 2 (=)9.63  × 10 2 (+)
Waveform1.37  × 10 1 1.53  × 10 1 (+)1.49  × 10 1 (+)1.47  × 10 1 (+)1.47  × 10 1 (+)
Tic-Tac-Toe Endgame1.65  × 10 1 1.61  × 10 1 (−)1.74  × 10 1 (+)1.61  × 10 1 (−)1.73  × 10 1 (+)
Hill2.93  × 10 1 3.56  × 10 1 (+)3.42  × 10 1 (+)3.44  × 10 1 (+)3.26  × 10 1 (+)
Urban7.48  × 10 2 1.03  × 10 1 (+)9.62  × 10 2 (+)9.53  × 10 2 (+)1.00  × 10 1 (+)
Vehicle1.82  × 10 1 1.86  × 10 1 (+)1.90  × 10 1 (+)1.83  × 10 1 (=)2.00  × 10 1 (+)
Musk version11.62  × 10 2 3.68  × 10 2 (+)3.15  × 10 2 (+)2.88  × 10 2 (+)3.71  × 10 2 (+)
LSVT Voice Rehabilitation2.79  × 10 2 3.81  × 10 2 (+)3.81  × 10 2 (+)3.35  × 10 2 (+)4.98  × 10 2 (+)
BreastEW5.47  × 10 3 6.64  × 10 3 (+)6.82  × 10 3 (+)5.12  × 10 3 (=)1.01  × 10 2 (+)
arrhythmia2.36  × 10 1 2.76  × 10 1 (+)2.66  × 10 1 (+)2.63  × 10 1 1(+)2.66  × 10 1 (+)
LibrasMovemen9.28  × 10 2 1.22  × 10 1 (+)1.18  × 10 1 (+)1.14  × 10 1 (+)1.32  × 10 1 (+)
Table 5. Statistical results of T-test between BRPSO and other 4 competitors.
Table 5. Statistical results of T-test between BRPSO and other 4 competitors.
BRPSO vs. BPSONBPSOLIWPSOBGWOPSO
(#)+accuracy913815
select12131014
fitness11141015
(#)−accuracy1030
select0100
fitness1010
CP 30392444
Table 6. Statistical results of Friedman test between BRPSO and other 4 competitors.
Table 6. Statistical results of Friedman test between BRPSO and other 4 competitors.
RankErrorSelectFitness
1BRSPO1.40BRSPO1.27BRSPO1.40
2LIWPSO1.93NBPSO2.6LIWPSO1.93
3BPSO3.67LIWPSO2.87NBPSO3.60
4NBPSO3.67BPSO4.13BPSO3.73
5BGWOPSO4.33BGWOPSO4.13BGWOPSO4.33
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhu, J.; Liu, J.; Chen, Y.; Xue, X.; Sun, S. Binary Restructuring Particle Swarm Optimization and Its Application. Biomimetics 2023, 8, 266. https://doi.org/10.3390/biomimetics8020266

AMA Style

Zhu J, Liu J, Chen Y, Xue X, Sun S. Binary Restructuring Particle Swarm Optimization and Its Application. Biomimetics. 2023; 8(2):266. https://doi.org/10.3390/biomimetics8020266

Chicago/Turabian Style

Zhu, Jian, Jianhua Liu, Yuxiang Chen, Xingsi Xue, and Shuihua Sun. 2023. "Binary Restructuring Particle Swarm Optimization and Its Application" Biomimetics 8, no. 2: 266. https://doi.org/10.3390/biomimetics8020266

APA Style

Zhu, J., Liu, J., Chen, Y., Xue, X., & Sun, S. (2023). Binary Restructuring Particle Swarm Optimization and Its Application. Biomimetics, 8(2), 266. https://doi.org/10.3390/biomimetics8020266

Article Metrics

Back to TopTop