Next Article in Journal
Multiple-Attribute Decision Making ELECTRE II Method under Bipolar Fuzzy Model
Next Article in Special Issue
Solving the Urban Transit Routing Problem Using a Cat Swarm Optimization-Based Algorithm
Previous Article in Journal
A QUBO Model for the Traveling Salesman Problem with Time Windows
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhancing Backtracking Search Algorithm using Reflection Mutation Strategy Based on Sine Cosine

1
School of Information Engineering, Hebei GEO University, Shijiazhuang 050031, China
2
Laboratory of Artificial Intelligence and Machine Learning, Hebei GEO University, Shijiazhuang 050031, China
3
College of Resources and Environment, Beibu Gulf University, Qinzhou 535011, China
*
Author to whom correspondence should be addressed.
Algorithms 2019, 12(11), 225; https://doi.org/10.3390/a12110225
Submission received: 17 September 2019 / Revised: 24 October 2019 / Accepted: 24 October 2019 / Published: 28 October 2019
(This article belongs to the Special Issue Hybrid Intelligent Algorithms)

Abstract

:
Backtracking Search Algorithm (BSA) is a younger population-based evolutionary algorithm and widely researched. Due to the introduction of historical population and no guidance toward to the best individual, BSA does not adequately use the information in the current population, which leads to a slow convergence speed and poor exploitation ability of BSA. To address these drawbacks, a novel backtracking search algorithm with reflection mutation based on sine cosine is proposed, named RSCBSA. The best individual found so far is employed to improve convergence speed, while sine and cosine math models are introduced to enhance population diversity. To sufficiently use the information in the historical population and current population, four individuals are selected from the historical or current population randomly to construct an unit simplex, and the center of the unit simplex can enhance exploitation ability of RSCBSA. Comprehensive experimental results and analyses show that RSCBSA is competitive enough with other state-of-the-art meta-heuristic algorithms.

1. Introduction

There are many global optimization problems in the real world. These problems are characterized by complexity, multimodality, strong-nonlinearity, dynamic change, and non-differentiality. The traditional optimization algorithms do not show satisfactory performance on such optimization problems. Therefore, many scholars have begun developing new methods to effectively solve the optimization problems. Evolutionary algorithms (EA), which are a kind of population-based global optimization algorithm, have been widely used to solve such problems, like function optimization [1,2,3], combinatorial optimization [4], neural network training [5,6], and image processing [7]. At present, many EAs have been proposed, like Genetic Algorithm (GA) [8], Particle Swarm Optimization (PSO) [9], Differential Evolution (DE) [10], Artificial Bee Colony algorithm (ABC) [1], Grey Wolf Optimizer (GWO) [11], Whale Optimization Algorithm (WOA) [12] and Cultural Algorithms [13].
Backtracking search algorithm (BSA) is a new population-based heuristic method proposed by Civicioglu for solving real-valued numerical optimization problems [14]. BSA has a simple structure and only one control parameter needs to be set, and its performance is insensitive to the initial value of the control parameter [15,16]. BSA uses a historical population which records individuals of the previous generation as the search direction matrix. Due to the historical population, BSA can simultaneously use the current population and previous population to improve the diversity of the population. Meanwhile, BSA owns a mutation operator with strong randomness and double crossover mechanism with probability, which leads to its strong global search ability. Hence, BSA is researched widely by many scholars and many variants are proposed. These variants can be divided roughly into four categories. The first category focuses on improving its initialization. Kolawol et al. [17] proposed chaotic-based BSA, in which random generation of the initial population is replaced with chaotic mapping to enhance the disorder of the initial population. To adequately use the search space, Yuan et al. [18] adopted an orthogonal design method rather than random generation. To enhance the diversity of the initial population, Lin [19] and Xu [7] et al. proposed different BSA variants based on opposition learning, in which the initial population and oppositional initial population are produced simultaneously to select the best candidate solution and the jump-out strategy in the opposition learning is added to update population. The main idea of the second category is that crossover and mutation strategy are improved. Zhao et al. [20] proposed an improved BSA algorithm mixing three mutation strategies, to handle the constrained optimization problem. For slow convergence speed at the late stage and poor exploration ability, Wang et al. [21] proposed a BSA variant embedding mutation strategy of DE, which improves convergence speed without adding complexity of the algorithm. Zhao et al. [22] designed an optimal solution-guided BSA, where the algorithm introduces an experience parameter to divide the overall iteration into two stages. At the earlier stage of the iteration, the algorithm adopts the mutation operator of the original BSA, while at the latter of the iteration, the algorithm employs the optimal solution-guided new mutation operator. Tian et al. [23] developed a new mutation strategy, which does not consider the optimal solution-guided information. This algorithm improves mutation operation though decreasing perturbation to strengthen convergence speed of the algorithm. For the drawbacks of slow convergence speed and falling into local optimum, Wang et al. [24] proposed an improved BSA fusing optimal solution-guided mutation strategy and niche technology. Chen et al. [25] proposed an ensemble learning, niche technology, mutation perturbation-based hybrid BSA and it is successfully applied to neural network training. The third category is improved parameter-based variants. BSA has a crossover probability parameter m i x r a t e and an adaptive mutation control parameter F. In [26], Wang et al. proposed an adaptive parameter F strategy which obeys Maxwell Boltzmann distribution since parameter F value influences the convergence speed of the algorithm. Duan et al. [27] introduced a fitness distance-based adaptive parameter F and m i x r a t e with experience parameters to accelerate convergence speed. Nama et al. [28] used a fitness information-based feedback mechanism to design a more available adaptive parameter F which can vary in [0.45, 1.99] with the change of fitness. Meanwhile, m i x r a t e randomly varied in [0,1] is designed to improve search efficiency of the algorithm. Chen et al. [29] proposed a new adaptive parameter F with experience parameter, which can adaptively reduce with iteration increase of the algorithm, to effectively balance exploration and exploitation of the BSA. Askarzadeh et al. [30] employed burger chaotic mapping to adjust parameter F, but two experience parameters are introduced into the strategy. Shaheen et al. [31] analyzed five different distributions F to impact algorithm performance and select the best available design of F. The fourth category is with the local search mechanisms embedded. To improve local exploitation of BSA and convergence speed, some search mechanisms are embedded into the framework of BSA. A chaotic mapping-based local search is embedded into BSA and combines with another algorithm to solve the hydrothermal generator set problem. Ali et al. [32] added a random walk guided local exploration strategy into BSA after selection-II of BSA, enhancing exploration performance of the algorithm. To solve the displacement flow shop scheduling problem, Lin et al. [33] proposed three different strategies to improve BSA. A new initial generator is constructed to improve the quality of the initial solution. In addition, the selection mechanism accepting a worse solution in the Simulated Annealing (SA) is employed to avoid trapping in local optimum. A local search mechanism randomly inserted is introduced to enhance exploitation ability of BSA. Moreover, due to its flexibility and efficiency, BSA and its variants have been widely applied to a wide range of real-world optimization problems, such as economic dispatch problems [34], optimal power flow [35], parameter identification [36,37], feature selection [38], artificial neural network [25], flow shop scheduling [39], and nonlinear optimal control problem [40].
However, BSA still has some drawbacks and its performance needs to be further improved. The first is that only the historical population information is adopted to guide the search and the information in current population is not used sufficiently, which cannot maintain the population diversity efficiently and thus the exploration ability of BSA is weak. The second is that there is no guidance regarding the approach to the current best individual during the evolution process, which leads to slow convergence speed and poor exploitation ability of BSA.
Based on the above discussion, in this paper, an enhancing backtracking search optimization algorithm with reflection mutation strategy based on sine cosine, named RSCBSA, is proposed. In RSCBSA, inspired by reflection operation in Nelder–Mead method [41] and Sine Cosine Algorithm (SCA) [42], a new reflection mutation strategy based on sine cosine is developed to address the above-mentioned drawbacks, in which the best solution and sine and cosine math models are introduced to balance exploration and exploitation ability of BSA.
The contributions of the paper are listed as follows:
  • A new reflection mutation strategy based on sine cosine is proposed to balance exploration and exploitation ability of BSA. To improve exploration ability, the best global individual is used to guide search direction, while sine and cosine math functions are used to enhance exploitation ability of BSA. Based on the strategy, a novel backtracking search algorithm with reflection mutation strategy based on sine cosine (RSCBSA) is proposed to solve global optimization problems.
  • In the above strategy, the center of a unit simplex constructed by three individuals selected randomly is employed to enhance diversity of population, since it considers more information of individuals. In addition, in RSCBSA, crossover operator of BSA is replaced with that of DE.
  • A comprehensive experiment is designed to verity the effectiveness of the proposed RSCBSA. In addition, a new parameter in RSCBSA is analyzed to set suitable values so that the performance of RSCBSA is the best.
The rest of this paper is organized as follows. Section 2 reviews the classic BSA. Section 3 presents the detail of RSCBSA, including its constituent RSCBSA variants and its framework. The experimental setting and results are reported and analyzed in Section 4 and Section 5. Finally, Section 6 concludes this paper.

2. Backtracking Search Optimization Algorithm

BSA is a young EA which is designed to be a global optimizer. It employs two populations: current population and historical population to search for the optimal solution. As done in other EAs, BSA has simple structure and can be divided into five processes by its functions: initialization, selection-I, mutation, crossover, and selection-II.
Initialization: In the stage, BSA randomly produces the initial population P using Equation (1).
P i , j = l b j + r a n d · u b j l b j
where i = 1 , 2 , 3 , , N P and N P is the number of individuals in population. j = 1 , 2 , 3 , , D and D is the number of dimension of variables. l o w j and u p j are the lower and upper boundaries on the jth variable, respectively. r a n d is a random number within [0,1].
Selection-I: In the BSA’s selection-I stage, the initial historical population o l d P is generated randomly using Equation (2).
o l d P i , j = l b j + r a n d · u b j l b j
Afterwards, at the beginning of each iteration, o l d P is redefined by Equation (3). In addition, Equation (4) is used to randomly update the order of the members in o l d P .
o l d P = P , if a < b o l d P , otherwise
o l d P = p e r m u t i n g o l d P
where p is a random number within [0,1]. p e r m u t i n g function is a random shuffling function. Through Equation (3), BSA assigns a population randomly selected previous generation as the historical population and saves the historical population until it is changed.
Mutation and Crossover: Mutation and crossover operators aim at producing a new individual. Equation (5) involving historical population and current population is used to generate a trail individual. Then, crossover operator is performed by Equation (6). From Equation (6), a binary integer-value map is employed to guide the crossover direction.
M = P + F · ( oldP P )
V i , j = P i , j , if m a p i , j = 1 M i , j , if m a p i , j = 0
where F is a scale factor and it generally takes value of 3 · r a n d , where r a n d is a random number between 0 and 1. V i , j is the value of the jth variable for the ith trial individual.
Selection-II: In the selection-II stage, BSA adopts a greedy selection mechanism to pick out a new solution P i n e w . For the minimum problem, if the fitness value of the trial individual V is better than that of the current individual P i , V is selected, as shown in Equation (7);
P i n e w = V i , if f V i f P i P i , otherwise
In the BSA, four above-mentioned processes 2–5 are repeatedly executed until it meets the termination criteria.

3. The Proposed Algorithm

In this section, the proposed algorithm is described in detail.

3.1. Initialization

The initialization stage is similar to that of the origin BSA. The initial population is produced.
P i , j = l b j + r a n d · u b j l b j
where i is the index of individual in population and j is the number of dimension of variables. R a n d is a uniformly distributed random number between 0 and 1. U b and l b are the upper and low boundary of each jth variable, respectively.

3.2. Reflection Mutation Strategy Based on Sine Cosine

In the proposed algorithm, reflection mutation strategy based on sine cosine is a key operator. Inspired by Nelder–Mead method and SCA, the strategy is proposed. After the operator, a new trial individual is generated to guide the research direction. It is similar to reflection mutation strategy in [43], but the difference is that sine and cosine math functions are introduced into the proposed strategy. First, four individuals are selected from the current population or historical population randomly to construct a unit simplex. The center of the unit simplex is computed using Equations (9) and (10).
X o = ω 1 · X a + ω 2 · X b + ω 3 · X c
ω 1 = f X a f X a + f X b + f X c ω 2 = f X b f X a + f X b + f X c ω 3 = f X c f X a + f X b + f X c
where a , b and c are the index of three different individuals randomly selected from population to construct a unit simplex. ω 1 , ω 2 and ω 3 are the weights of three vertexes in the unit simplex. X a , X b and X c are three vertexes of the unit simplex. X o is the center of the unit simplex.
Based on the center, the best individual found so far, and sine cosine are introduced into the strategy to balance exploration and exploitation. The best individual provides the better search direction to improve exploration ability of the algorithm, while sine and cosine math models are used to enhance diversity ability and the center of the unit simplex also provides more diversity information. The formulation of the strategy is listed as follows:
V i , j = X b e s t , j + η · sin r 1 r 2 X o , j X m , j if r 3 < 0.5 X b e s t , j + η · cos r 1 r 2 X o , j X m , j otherwise
η = a · 1 t M a x G e n
where r 1 and r 3 are random number in [0, 1]. r 2 is a random number in [0, 2]. X b e s t , j indicates the jth dimension of the best individual. X m , j is the jth dimension of the m individual randomly selected from population. t is the current iteration and M a x G e n is the maximum iteration. a is a perturbation parameter and is usually set to 2.0.

3.3. Crossover Operator

To produce better trial individuals, crossover operator is executed. Like the DE, the crossover operator uses two individuals V i ( t ) and X ( t ) to generate offspring individual randomly, where V i ( t ) is generated by the above mutation strategy and X ( t ) is current individual. The formula of crossover operator is showed as follows:
u i , j = v i , j , if rand C R j = l x i , j , otherwise
where j = 1 , 2 , , D , r a n d is in [0,1], l is a random integer in [1, D]. C R is crossover probability and is usually set to 0.9.

3.4. The Framework of The Proposed Algorithm

Based on above description, the framework of the proposed RSCBSA is similar to that of BSA, but the difference is that mutation and crossover stages are replaced with reflection mutation based on sine cosine and crossover operator. The framework of RSCBSA is presented in Algorithm 1.
Algorithm 1: Framework of the Proposed Algorithm
Algorithms 12 00225 i001

3.5. Complex Analysis of The Proposed Algorithm

The proposed algorithm is based on the basic BSA framework. Five processes in the algorithm are implemented: initialization, selection-I, reflection mutation strategy based on sine cosine, crossover operator, and selection-II, where reflection mutation strategy based on sine cosine and crossover operator are the main steps. In the reflection mutation stage, N new individuals are produced, which leads to O ( N D ) time complexity, where N is population size and D is the dimension of test problem. Crossover operator needs O ( N D ) time complexity for each individual in the trial population. In brief, the time complexity of the proposed algorithm is O ( N D ) .

4. Experimental Simulations

In this section, the experimental simulation is listed. In Section 4.1, benchmark test suit used in the experiment is described. Then, the parameter setting is stated in Section 4.2.

4.1. Benchmark Test Suit

The 23 classical benchmark functions used widely by many researchers are used in the experiment to verify performance of the proposed algorithm. These benchmark functions F 1 F 23 can be found in [12]. Typically, these functions can be divided into three categories: (1) unimodal benchmark functions, (2) multimodal benchmark functions and (3) fixed-dimension multimodal functions. F 1 F 7 belong to unimodal benchmark functions, whose dimension are set to 30 to test the convergence performance of search algorithms. F 8 F 13 are multimodal benchmark functions with several local optimum, whose dimension also are set to 30 to verify the convergence performance and avoidance premature convergence of the proposed algorithm, while the other functions with low-dimension are used to deal with small number of local optima. Their function name, expression, dimension, search range and theoretical optimal value can be found in [12]. All of the algorithms are programmed in “Matlab 2014” and implemented on “Windows 10 64bit” environment on a computer with “Intel Core i5” processor and 8 GB memory.

4.2. Parameter Setting

Common parameters of all considered algorithms are set to same size. The maximum number of generations ( M a x G e n ) is equal to 3000. Population size is set to 30. All algorithms are run 30 times independently. The other parameters used in different algorithm are set as follows:
F in DE: random number in [0.2, 0.8];
C R in DE and RSCBSA: 0.9;
a in RSCBSA: 2.0;
Personal learning coefficient c 1 in PSO: 1.5;
Global learning coefficient c 2 in PSO: 2.0;
l i m i t in ABC: 100;
Acceptance Ratio p A c c e p t in CA: 0.35.

5. Experimental Results

In this section, five state-of-the-art algorithms are compared with the proposed algorithm, including DE, PSO, ABC, CA and original BSA, to verify the performance of the proposed algorithm. Statistical results on the three categories test problems are reported in Tables 1, 3, and 5, respectively, in which the best value, mean value, the worst value and standard deviation obtained by the six algorithms are listed. To have statistically sound conclusions, the Friedman test with Bonferroni-Dunn’s procedure is employed to achieve the final ranking of different algorithms on the different type benchmark functions. Moreover, the Wilcoxon’s rank-sum test for independent sample at a 0.05 significance level [44], which is a nonparametric statistical test method, is conducted to judge the significance of the results between two algorithms. Signs “+“, “−“ and “=“ indicate that the corresponding comparative algorithm is worse than, better than, and similar to RSCBSA, respectively. Then, convergence behavior of RSCBSA is analyzed and showed. Finally, parameter a in the reflection mutation based on sine cosine is analyzed by taking different value.

5.1. Compared with State-of-the-Art Algorithms

Unimodal Benchmark Problems ( F 1 F 7 ): From Table 1, it is seen that the proposed algorithm is superior to other algorithms considered on the unimodal benchmark functions. The PSO obtains the best performance on 1 ( F 6 ) test function. ABC obtains the best solution on 1 ( F 5 ) test function. However, RSCBSA can perform the best than other algorithms except 2 ( F 5 and F 6 ) test functions. The reason maybe that the best global individual is employed to strength the exploration ability which leads to that the algorithm can explore more promising area to find the better solution. In addition, the Friedman rank is implemented based on KEEL software [45], and the results are reported in Table 2. As shown in Table 2, RSCBSA obtains the best place.
Multimodal Benchmark Problems ( F 8 F 13 ): Experiment results are listed in Table 3 on the multimodal test functions. From Table 3, DE is superior to other five compared algorithms on the 4 ( F 8 and F 11 F 13 ) benchmark functions. RSCBSA can obtain the best solution on the 3 ( F 9 F 11 ) benchmark functions, but on the other functions it does not perform well. ABC and BSA outperform RSCBSA on the 3 ( F 8 and F 12 F 13 ) benchmark functions and CA cannot beat BSA on the any benchmark functions except F 8 . Although RSCBSA can get the best on three benchmark functions, accounting for 50%, RSCBSA only gets the third rank in terms of the Friedman rank, as shown in Table 4. DE performs the best than other peer competitors and gets the first rank.
Fixed-dimension multimodal benchmark functions ( F 14 F 23 ): For the category of benchmark function, the proposed RSCBSA can obtain the best performance on 7 test functions ( F 14 , F 16 F 19 and F 21 F 22 ) from Table 5. BSA can perform the best than other competitors on the all test functions. The performance of PSO, DE, CA is inferior to that of the BSA and RSCBSA. From Table 6, it is observed that RSCBSA can rank the second place according to the Friedman rank.
To further detect the significant differences between the proposed RSCBSA and the five competitors, the Wilcoxon’s rank-sum test is executed. The statistical results are reported in Table 7. It is seen that RSCBSA outperforms 9, 14, 9, 12 and 17 benchmark functions than BSA, PSO, ABC, DE and CA, respectively.
In summary, the proposed RSCBSA can exhibit very competitive performance than those of the other well-known algorithms, which is able to generate the high-quality solutions and accelerate the convergence speed.

5.2. Convergence Analysis

To analyze the convergence of the proposed RSCBSA, convergence curves of RSCBSA, PSO, DE, ABC, CA and BSA are shown in Figure 1 as iteration increases. In the Figure 1, X axial represents for the total number of iteration and Y axial stands for the logarithm of the function optimal value. It is observed that RSCBSA tends to extensively explore promising areas of design space and exploit the best one. The convergence curve changes abruptly in the early stages of the optimization process and then gradually converge. According to Berg et al. [46], such a behavior can guarantee that a population-based algorithm eventually convergences to a point in a search space. On the unimodal benchmark function, RSCBSA have an obvious advantage than other competitors. However, although RSCBSA does not perform better than other algorithms on the multimodal benchmark function, RSCBSA and other algorithms can converge the global optimal on the almost multimodal benchmark functions. It can be seen that RSCBSA is enough competitive with other state-of-the-art meta-heuristic algorithms.

5.3. Parameter Sensitivity Analysis

To investigate the effect of parameter a in the mutation strategy for the proposed algorithm, an experiment is designed. In the experiment, parameter a is set from 1 up to 6 and experiment results are plotted in Figure 2, where X axial represents for the total number of iteration and Y axial stands for the logarithm of the function optimal value. As shown in Figure 2, the performance of the proposed RSCBSA leads to the optimal on the almost test functions when parameter a is set to 2.0.

5.4. Runtime Analysis

To present execution times, RSCBSA, PSO and DE are selected to run 20 times individually and record their runtimes. The average time is computed and showed in the Figure 3. From Figure 3, it is seen that RSCBSA obtains the best execution speed on the five benchmark problems, while DE and PSO rank the second and third place, respectively.

5.5. Remarks

Based on the above results, some insights can be summarized: (1) RSCBSA can obtain the best performance on the almost unimodal benchmark functions, since in the reflection mutation strategy based on sine cosine, the best individual found so far is employed to improve the exploration ability of the algorithm. In addition, sine cosine math models are introduced to enhance diversity. However, RSCBSA perform badly on the multimodal test function. The reason may be that the crossover operator is weak for enhancing diversity of population. (2) CA is the worst among the six algorithms, but CA can perform the best on the fixed-dimension multimodal benchmark functions from Figure 1, due to strong performance of jump-out local optimal. (3) DE has obvious advantages on the multimodal test functions.

6. Conclusions

This study presented a new backtracking search algorithm, in which a novel reflection mutation strategy based on sine cosine is proposed to balance exploration and exploitation ability of BSA. The proposed algorithm (called RSCBSA) replaces mutation strategy with the proposed reflection mutation strategy based on sine cosine. In the strategy, the best individual found so far is employed to improve convergence speed, while sine cosine and the center of a unit simplex constructed by three individuals selected from population or historical population are used to enhance exploitation ability. Reflection mutation is similar to that in the Nelder–Mead method. To enhance a new trail individual produced by mutation strategy, crossover operator of BSA is replaced with that of DE. In addition, selection-II stage is saved to avoid the loss of the best individual. To verify the performance of RSCBSA, a comprehensive experiment is designed. Experimental results show that the proposed RSCBSA can obtain the best performance on the almost benchmark functions.
In the future, some work should be done to further improve performance of the algorithm. A parameter adaptation strategy can be designed to enhance search ability of RSCBSA. RSCBSA can be tested on some more difficult functions [47,48], to improve the effectiveness of their algorithm. RSCBSA can be extended to discrete/combinatorial spaces using some algebraic-based strategies [49,50,51]. In addition, several other well-studied evolutionary algorithms (EAs), i.e., CMA-ES [52] and WOA [12] can be combined so that the resultant algorithms can effectively complement one another.

Author Contributions

Conceptualization, C.Z. (Chong Zhou) and S.L.; methodology, C.Z. (Chong Zhou); software, S.L.; data curation, S.L. and Y.Z.; writing–original draft preparation, C.Z. (Chong Zhou); writing–review and editing, C.Z. (Cuijun Zhang); project administration, C.Z. (Chong Zhou); funding acquisition, C.Z. (Chong Zhou); resources, Zhikun Chen; computing resources and other analysis tools, C.Z. (Chong Zhou).

Funding

This research was partially funded by Natural Science Youth Foundation of Hebei Province under Grant No. F2019403207, the High-Level Talents Research Projects of Beibu Gulf University under Grant No.2019KYQD27 and Research Fund for the Doctoral Startup Program of Hebei GEO University.

Acknowledgments

The authors thank the anonymous reviewers for their constructive comments. Chong Zhou, Shengjie Li and Cuijun Zhang are equivalent contribution authors.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Karaboga, D.; Basturk, B. A powerful and efficient algorithm for numerical function optimization: Artificial bee colony (ABC) algorithm. J. Glob. Optim. 2007, 39, 459–471. [Google Scholar] [CrossRef]
  2. Pan, Q.K.; Sang, H.Y.; Duan, J.H.; Gao, L. An improved fruit fly optimization algorithm for continuous function optimization problems. Knowl. Based Syst. 2014, 62, 69–83. [Google Scholar] [CrossRef]
  3. Soleimanpour-Moghadam, M.; Nezamabadi-Pour, H.; Farsangi, M.M. A quantum inspired gravitational search algorithm for numerical function optimization. Inf. Sci. 2014, 267, 83–100. [Google Scholar] [CrossRef]
  4. Baykasoğlu, A.; Hamzadayi, A.; Köse, S.Y. Testing the performance of teaching–learning based optimization (TLBO) algorithm on combinatorial problems: Flow shop and job shop scheduling cases. Inf. Sci. 2014, 276, 204–218. [Google Scholar] [CrossRef]
  5. Yaghini, M.; Khoshraftar, M.M.; Fallahi, M. A hybrid algorithm for artificial neural network training. Eng. Appl. Artif. Intell. 2013, 26, 293–301. [Google Scholar] [CrossRef]
  6. Eberhart, R.C.; Kennedy, J. A new optimizer using particle swarm theory. In Proceedings of the Sixth International Symposium on Micro Machine and Human Science, Nagoya, Japan, 4–6 October 1995; pp. 39–43. [Google Scholar]
  7. Xu, L.; Jia, H.; Lang, C.; Peng, X.; Sun, K. A novel method for multilevel color image segmentation based on dragonfly algorithm and differential evolution. IEEE Access 2019, 7, 19502–19538. [Google Scholar] [CrossRef]
  8. Goldberg, D. Genetic Algorithms in Search, Optimization, and Machine Learning; Addison-Wesley Longman Publishing Co., Inc.: Boston, MA, USA, 1989. [Google Scholar]
  9. Kennedy, J.; Eberhart, R. Particle swarm optimization. In Proceedings of the ICNN’95—International Conference on Neural Networks, Perth, WA, Australia, 27 November–1 December 1995. [Google Scholar]
  10. Storn, R.; Price, K. Differential evolution—A simple and efficient heuristic for global optimization over continuous spaces. J. Glob. Optim. 1997, 11, 341–359. [Google Scholar] [CrossRef]
  11. Mirjalili, S.; Mirjalili, S.M.; Lewis, A. Grey wolf optimizer. Adv. Eng. Softw. 2014, 69, 46–61. [Google Scholar] [CrossRef]
  12. Mirjalili, S.; Lewis, A. The whale optimization algorithm. Adv. Eng. Softw. 2016, 95, 51–67. [Google Scholar] [CrossRef]
  13. Reynolds, R.G. An introduction to cultural algorithms. In Evolutionary Programming—Proceedings of the Third Annual Conference; World Scientific: Singapore, 1994; pp. 131–139. [Google Scholar]
  14. Civicioglu, P. Backtracking search optimization algorithm for numerical optimization problems. Appl. Math. Comput. 2013, 219, 8121–8144. [Google Scholar] [CrossRef]
  15. Madasu, S.D.; Kumar, M.S.; Singh, A.K. Comparable investigation of backtracking search algorithm in automatic generation control for two area reheat interconnected thermal power system. Appl. Soft Comput. 2017, 55, 197–210. [Google Scholar] [CrossRef]
  16. Islam, N.N.; Hannan, M.; Shareef, H.; Mohamed, A. An application of backtracking search algorithm in designing power system stabilizers for large multi-machine system. Neurocomputing 2017, 237, 175–184. [Google Scholar] [CrossRef]
  17. Kolawole, S.O.; Duan, H. Backtracking search algorithm for non-aligned thrust optimization for satellite formation. In Proceedings of the 11th IEEE International Conference on Control & Automation (ICCA), Taichung, Taiwan, 18–20 June 2014; pp. 738–743. [Google Scholar]
  18. Yuan, X.; Wu, X.; Tian, H.; Yuan, Y.; Adnan, R.M. Parameter identification of nonlinear Muskingum model with backtracking search algorithm. Water Resour. Manag. 2016, 30, 2767–2783. [Google Scholar] [CrossRef]
  19. Lin, J. Oppositional backtracking search optimization algorithm for parameter identification of hyperchaotic systems. Nonlinear Dyn. 2015, 80, 209–219. [Google Scholar] [CrossRef]
  20. Zhao, W.; Wang, L.; Yin, Y.; Wang, B.; Yi, W.; Yin, Y. An improved backtracking search algorithm for constrained optimization problems. In International Conference on Knowledge Science; Springer: Cham, Switzerland, 2014. [Google Scholar]
  21. Wang, L.; Zhong, Y.; Yin, Y.; Zhao, W.; Wang, B.; Xu, Y. A hybrid backtracking search optimization algorithm with differential evolution. Math. Probl. Eng. 2015, 2015, 1–16. [Google Scholar] [CrossRef]
  22. Zhao, W.; Wang, L.; Wang, B.; Yin, Y. Best guided backtracking search algorithm for numerical optimization problems. In Knowledge Science, Engineering and Management; Lehner, F., Fteimi, N., Eds.; Springer International Publishing: Cham, Switzerland, 2016; pp. 414–425. [Google Scholar]
  23. Tian, K. Effective self-learning backtracking search optimization algorithm. Electron. Sci. Tech. 2015, 28, 41. [Google Scholar]
  24. Wang, P.; Chen, D.; Zou, F.; Li, Z. Guidance and niching backtracking search optimization algorithm. CEA 2017, 53, 126–131. [Google Scholar]
  25. Chen, D.; Lu, R.; Feng, Z.; Li, S.; Peng, W. A learning and niching based backtracking search optimisation algorithm and its applications in global optimisation and ANN training. Neurocomputing 2017, 266, 579–594. [Google Scholar] [CrossRef]
  26. Wang, X.; Liu, S.; Tian, W. Improved backtracking search optimization algorithm with new effective mutation scale factor and greedy crossover strategy. J. Comput. Appl. 2014, 34, 2543. [Google Scholar]
  27. Duan, H.; Luo, Q. Adaptive backtracking search algorithm for induction magnetometer optimization. IEEE Trans. Magn. 2014, 50, 1–6. [Google Scholar] [CrossRef]
  28. Nama, S.; Saha, A.K.; Ghosh, S. Improved backtracking search algorithm for pseudo dynamic active earth pressure on retaining wall supporting c-Φ backfill. Appl. Soft Comput. 2016, 52, 885–897. [Google Scholar] [CrossRef]
  29. Chen, X.; Liu, S.; Wang, Y. Emergency resources scheduling based on improved backtracking search optimization algorithm. Comput. Appl. Softw. 2015, 32, 235–238. [Google Scholar]
  30. Askarzadeh, A.; Coelho, L.D.S. A backtracking search algorithm combined with Burger’s chaotic map for parameter estimation of PEMFC electrochemical model. Int. J. Hydrog. Energy 2014, 39, 11165–11174. [Google Scholar] [CrossRef]
  31. Shaheen, A.M.; El-Sehiemy, R.A.; Farrag, S.M. Integrated strategies of backtracking search optimizer for solving reactive power dispatch problem. IEEE Syst. J. 2016, PP, 1–10. [Google Scholar] [CrossRef]
  32. Ali, A.F. A memetic backtracking search optimization algorithm for economic dispatch problem. Egypt. Comput. Sci. J. 2015, 39, 56–71. [Google Scholar]
  33. Lin, Q.; Liang, G.; Li, X.; Zhang, C. A hybrid backtracking search algorithm for permutation flow-shop scheduling problem. Comput. Ind. Eng. 2015, 85, 437–446. [Google Scholar] [CrossRef]
  34. Modiri-Delshad, M.; Kaboli, S.H.A.; Taslimi-Renani, E.; Rahim, N.A. Backtracking search algorithm for solving economic dispatch problems with valve-point effects and multiple fuel options. Energy 2016, 116, 637–649. [Google Scholar] [CrossRef]
  35. Ayan, K.; Kilic, U. Optimal power flow of two-terminal HVDC systems using backtracking search algorithm. Int. J. Electr. Power Energy Syst. 2016, 78, 326–335. [Google Scholar] [CrossRef]
  36. Ahandani, M.A.; Ghiasi, A.R.; Kharrati, H. Parameter identification of chaotic systems using a shuffled backtracking search optimization algorithm. Soft Comput. 2017, 1–23. [Google Scholar] [CrossRef]
  37. Yu, K.; Liang, J.; Qu, B.Y.; Zhiping, C.; Heshan, W. Multiple learning backtracking search algorithm for estimating parameters of photovoltaic models. Appl. Energy 2018, 226, 408–422. [Google Scholar] [CrossRef]
  38. Chu, Z.; Zhou, J.; Li, C.; Fu, W.; Tian, P. A compound structure of ELM based on feature selection and parameter optimization using hybrid backtracking search algorithm for wind speed forecasting. Energy Convers. Manag. 2017, 143, 360–376. [Google Scholar]
  39. Lin, J.; Wang, Z.J.; Li, X. A backtracking search hyper-heuristic for the distributed assembly flow-shop scheduling problem. Swarm Evol. Comput. 2017, 36, S2210650216305028. [Google Scholar] [CrossRef]
  40. Su, Z.; Wang, H.; Peng, Y. A hybrid backtracking search optimization algorithm for nonlinear optimal control problems with complex dynamic constraints. Neurocomputing 2016, 186, 182–194. [Google Scholar] [CrossRef]
  41. Nelder, J.A.; Mead, R. A simplex method for function minimization. Comput. J. 1965, 7, 308–313. [Google Scholar] [CrossRef]
  42. Mirjalili, S. SCA: A sine cosine algorithm for solving optimization problems. Knowl. Based Syst. 2016, 96, 120–133. [Google Scholar] [CrossRef]
  43. Qian, W.; Chai, J.; Zhang, Z. Adaptive differential evolution algorithm based on reflective mutation strategy. Comput. Eng. Appl. 2018, 54, 166–173. [Google Scholar]
  44. Derrac, J.; Garcia, S.; Molina, D.; Herrera, F. A practical tutorial on the use of nonparametric statistical tests as a methodology for comparing evolutionary and swarm intelligence algorithms. Swarm Evol. Comput. 2011, 1, 3–18. [Google Scholar] [CrossRef]
  45. Introduction to KEEL Software Suite. Available online: https://sci2s.ugr.es/keel/development.php (accessed on 26 October 2019).
  46. Bergh, F.V.D.; Engelbrecht, A.P. A study of particle swarm optimization particle trajectories. Inf. Sci. 2006, 176, 937–971. [Google Scholar]
  47. García-Martínez, C.; Gutiérrez, P.D.; Molina, D.; Lozano, M.; Herrera, F. Since CEC 2005 competition on real-parameter optimisation: a decade of research, progress and comparative analysis’s weakness. Soft Comput. 2017, 21, 1–11. [Google Scholar] [CrossRef]
  48. Suganthan, P.N.; Hansen, N.; Liang, J.J.; Deb, K.; Chen, Y.-P.; Auger, A.; Tiwari, S. Problem Definitions and Evaluation Criteria for the CEC 2005 Special Session on Real-Parameter Optimization; Technical Report; Nanyang Technological University: Singapore, 2005. [Google Scholar]
  49. Baioletti, M.; Milani, A.; Santucci, V. Automatic Algebraic Evolutionary Algorithms; Springer: Cham, Switzerland, 2017. [Google Scholar]
  50. Baioletti, M.; Milani, A.; Santucci, V. Algebraic Particle Swarm Optimization for the permutations search space. In Proceedings of the 2017 IEEE Congress on Evolutionary Computation (CEC), San Sebastian, Spain, 5–8 June 2017. [Google Scholar]
  51. He, Y.; Wang, X. Group theory-based optimization algorithm for solving knapsack problems. Knowl. Based Syst. 2018. [Google Scholar] [CrossRef]
  52. Hansen, N. Covariance matrix adaptation evolution strategy. In Proceedings of the 10th International Conference on Parallel Problem Solving from Nature, Dortmund, Germany, 13–17 September 2008; Springer: Cham, Switzerland, 2008. [Google Scholar]
Figure 1. Convergence figures on test functions F1–F23, where (aw) indicate the convergence curves on the functions F1–F23 respectively.
Figure 1. Convergence figures on test functions F1–F23, where (aw) indicate the convergence curves on the functions F1–F23 respectively.
Algorithms 12 00225 g001aAlgorithms 12 00225 g001bAlgorithms 12 00225 g001c
Figure 2. Convergence curves on test functions F1–F23 for different a values, where (aw) indicate the convergence curves on the functions F1–F23 respectively.
Figure 2. Convergence curves on test functions F1–F23 for different a values, where (aw) indicate the convergence curves on the functions F1–F23 respectively.
Algorithms 12 00225 g002aAlgorithms 12 00225 g002bAlgorithms 12 00225 g002c
Figure 3. Runtime(s) of the three algorithms on five test problems.
Figure 3. Runtime(s) of the three algorithms on five test problems.
Algorithms 12 00225 g003
Table 1. Results of unimodal benchmark functions.
Table 1. Results of unimodal benchmark functions.
BenchmarkFunctionPSOABCDECABSARSCBSA
F1best 2.6078 × 10 21 5.3796 × 10 16 2.5816 × 10 46 2.5562 × 10 2 1.4976 × 10 17 0.0000 × 10 0
mean 3.9736 × 10 7 7.5685 × 10 16 5.4588 × 10 45 1.6712 × 10 3 2.4454 × 10 15 0.0000 × 10 0
worst3.6505 × 10 6 5.4117 × 10 15 2.4629 × 10 44 4.1270 × 10 3 2.8521× 10 14 0.0000 × 10 0
std1.3868 × 10 12 6.8910 × 10 31 2.6423 × 10 89 9.5577 × 10 5 3.0966 × 10 29 0.0000 × 10 0
F2best5.4694 × 10 28 1.6002 × 10 15 4.1335 × 10 28 5.8462 × 10 0 5.0408 × 10 10 4.3459× 10 218
mean1.3657 × 10 12 1.8228 × 10 15 2.0667 × 10 27 2.2100 × 10 1 3.2572 × 10 9 3.9314 × 10 214
worst1.5274 × 10 11 3.3803 × 10 15 5.6259 × 10 27 6.9183 × 10 1 1.0338 × 10 8 3.1240 × 10 213
std1.5008 × 10 23 1.6937 × 10 31 1.2307 × 10 54 2.3890 × 10 2 5.5714 × 10 18 0.0000 × 10 0
F3best3.9653 × 10 103 3.5792 × 10 3 7.3893 × 10 3 8.3334 × 10 3 3.0345 × 10 1 8.1950 × 10 204
mean6.4404 × 10 93 5.9449 × 10 3 1.1857 × 10 4 2.0627 × 10 4 1.7266 × 10 2 7.7379 × 10 188
worst1.5655 × 10 91 1.1140 × 10 4 1.9801 × 10 4 4.0887 × 10 4 3.7642 × 10 2 2.3203 × 10 186
std8.1160 × 10 184 3.2828 × 10 6 6.0971 × 10 6 7.8083× 10 7 6.6997 × 10 3 0.0000 × 10 0
F4best7.8193 × 10 83 2.8317 × 10 1 5.1221 × 10 4 3.0161 × 10 1 9.0018 × 10 1 8.1338 × 10 152
mean1.5871 × 10 72 3.9440 × 10 1 1.0355 × 10 3 4.6255 × 10 1 2.1297 × 10 0 4.7338 × 10 144
worst2.4649 × 10 71 5.0921 × 10 1 2.2349 × 10 3 7.9715 × 10 1 3.9292 × 10 0 8.8191 × 10 143
std2.1488 × 10 143 3.7018 × 10 1 1.5905 × 10 7 1.4438 × 10 2 4.0870 × 10 1 2.6901 × 10 286
F5best7.9093 × 10 2 8.6115 × 10 3 2.3793 × 10 1 1.0168 × 10 5 8.6456 × 10 1 2.3440× 10 1
mean2.2878 × 10 0 5.3359 × 10 1 3.3665 × 10 1 1.2073 × 10 6 5.5375 × 10 1 2.4658 × 10 1
worst7.0303 × 10 0 1.8079 × 10 1 8.8254 × 10 1 4.4121 × 10 6 8.6258 × 10 1 2.6975 × 10 1
std3.1431 × 10 0 1.3289 × 10 1 4.0340 × 10 2 1.0354 × 10 12 9.4312 × 10 2 6.2485 × 10 1
F6best0.0000 × 10 0 5.5033 × 10 16 0.0000 × 10 0 4.3707 × 10 2 2.2988 × 10 17 8.6469 × 10 7
mean1.9105 × 10 32 7.4696 × 10 16 0.0000 × 10 0 1.8850 × 10 3 4.8673 × 10 16 1.4989 × 10 1
worst1.1093 × 10 31 2.5076 × 10 15 0.0000 × 10 0 3.1993 × 10 3 1.9639× 10 15 5.1162× 10 1
std8.1244× 10 64 1.3274 × 10 31 0.0000 × 10 0 5.4730 × 10 5 2.3180 × 10 31 2.7930 × 10 2
F7best1.1152 × 10 4 9.4208 × 10 2 3.1487 × 10 3 3.5096 × 10 1 4.3966 × 10 3 3.3201 × 10 5
mean7.4359 × 10 4 2.0345 × 10 1 8.2541 × 10 3 1.4979 × 10 0 1.4448 × 10 2 1.7506 × 10 4
worst2.1471 × 10 3 3.5513 × 10 1 1.2403 × 10 2 3.4948 × 10 0 2.1994 × 10 2 4.4544 × 10 4
std3.1481 × 10 7 3.7637 × 10 3 4.3611 × 10 6 8.3832 × 10 1 1.6176× 10 5 1.1159 × 10 8
Table 2. Average Rankings of the algorithms (Friedman) on unimodal benchmark functions.
Table 2. Average Rankings of the algorithms (Friedman) on unimodal benchmark functions.
AlgorithmRanking
PSO2.3571
ABC3
DE3.4286
CA6
BSA4.0714
RSCBSA2.1429
Table 3. Results of multimodal benchmark functions.
Table 3. Results of multimodal benchmark functions.
BenchmarkFunctionPSOABCDECABSARSCBSA
F8best−3.2818 × 10 3 −1.2570 × 10 4 −1.2570 × 10 4 −1.2049 × 10 4 −1.2569 × 10 4 −1.0278 × 10 4
mean−2.4289 × 10 3 −1.2541 × 10 4 −1.2980 × 10 4 −9.8465 × 10 3 −1.2569 × 10 4 −8.8889 × 10 3
worst−1.6827 × 10 3 −1.2209 × 10 4 −1.2214 × 10 4 −6.9449 × 10 3 −1.2569 × 10 4 −8.0066 × 10 3
std9.9518 × 10 4 8.0666 × 10 3 −1.2442 × 10 4 3.3642 × 10 6 2.2234 × 10 1 2.3757 × 10 5
F9best5.9698 × 10 0 1.1369 × 10 13 6.8781 × 10 12 7.7160 × 10 1 1.0854 × 10 1 0.0000 × 10 0
mean1.1840 × 10 1 1.4061 × 10 8 5.0229 × 10 0 1.4651 × 10 2 3.3603 × 10 0 0.0000 × 10 0
worst2.0894 × 10 1 9.9496 × 10 1 2.6971 × 10 1 2.5216 × 10 2 6.6977 × 10 0 0.0000 × 10 0
std1.5334 × 10 1 6.2573 × 10 2 6.5267 × 10 1 1.8300 × 10 3 3.1008 × 10 0 0.0000 × 10 0
F10best4.4409 × 10 15 4.7074 × 10 14 7.9936 × 10 15 5.5878 × 10 0 2.1721 × 10 9 8.8818 × 10 16
mean3.8505 × 10 2 6.0574 × 10 14 7.9936 × 10 15 1.0163 × 10 1 2.5030 × 10 8 8.8818 × 10 16
worst1.1552 × 10 0 1.5721 × 10 13 7.9936 × 10 15 1.3951× 10 1 7.7180 × 10 8 8.8818 × 10 16
std4.2996× 10 2 5.4819 × 10 28 2.2403 × 10 59 3.9939 × 10 0 5.5106 × 10 16 3.8894 × 10 62
F11best3.9319 × 10 2 1.1102 × 10 15 0.0000 × 10 0 5.5845 × 10 0 0.0000 × 10 0 0.0000 × 10 0
mean9.3162 × 10 2 3.5826 × 10 4 0.0000 × 10 0 1.5086 × 10 1 1.6533 × 10 11 0.0000 × 10 0
worst2.2875 × 10 1 3.5406 × 10 2 0.0000 × 10 0 4.6938 × 10 1 4.9135 × 10 10 0.0000× 10 0
std1.9610 × 10 3 7.4801 × 10 5 0.0000 × 10 0 8.2619 × 10 1 7.7747 × 10 21 0.0000× 10 0
F12best4.7116 × 10 32 6.1436 × 10 16 1.5705 × 10 32 6.2196 × 10 1 2.8984 × 10 19 6.0657 × 10 7
mean5.1561 × 10 32 7.8009 × 10 16 1.5705 × 10 32 2.7468 × 10 5 2.8011 × 10 17 8.8653 × 10 3
worst1.2553 × 10 31 5.2760 × 10 13 1.5705 × 10 32 1.7374 × 10 6 4.4505 × 10 16 2.3351 × 10 2
std2.1613 × 10 64 8.9331 × 10 27 2.9963 × 10 95 1.7490 × 10 11 6.7812 × 10 33 4.6445 × 10 5
F13best1.0987 × 10 2 5.3680 × 10 16 1.3498 × 10 32 6.4179 × 10 3 2.2895 × 10 18 1.1010 × 10 2
mean1.0987× 10 3 8.3837 × 10 16 1.3498 × 10 32 2.6594 × 10 6 3.7103 × 10 16 5.0903 × 10 1
worst1.0987 × 10 2 2.2538 × 10 15 1.3498 × 10 32 1.2544 × 10 7 2.9765 × 10 15 1.4244 × 10 0
std1.0865 × 10 5 1.0659 × 10 31 0.0000 × 10 0 1.0542 × 10 13 4.7169 × 10 31 1.0055 × 10 1
Table 4. Average Rankings of the algorithms (Friedman) on multimodal benchmark functions.
Table 4. Average Rankings of the algorithms (Friedman) on multimodal benchmark functions.
AlgorithmRanking
PSO4.5833
ABC2.5833
DE2.25
CA5.6667
BSA2.5833
RSCBSA3.3333
Table 5. Results of fixed-dimension multimodal benchmark functions.
Table 5. Results of fixed-dimension multimodal benchmark functions.
BenchmarkFunctionPSOABCDECABSARSCBSA
F14best9.9800 × 10 1 9.9800 × 10 1 9.9800 × 10 1 9.9800 × 10 1 9.9800 × 10 1 9.9800 × 10 1
mean3.4567× 10 0 9.9800 × 10 1 1.1624 × 10 0 6.6697 × 10 0 9.9800 × 10 1 9.9800 × 10 1
worst1.2671 × 10 1 9.9800 × 10 1 5.9289 × 10 0 1.6441 × 10 1 9.9800 × 10 1 9.9800 × 10 1
std9.3179 × 10 0 4.4373 × 10 31 7.8343 × 10 1 1.9194 × 10 1 1.9722 × 10 31 1.9722 × 10 31
F15best3.0749 × 10 4 4.1171 × 10 4 3.0749 × 10 4 4.8171 × 10 4 3.0749 × 10 4 3.0749 × 10 4
mean3.3567 × 10 3 5.5933 × 10 4 4.7100 × 10 4 2.2035 × 10 3 3.0749 × 10 4 3.5056 × 10 4
worst2.0363 × 10 2 1.0451 × 10 3 7.8431 × 10 4 1.4641 × 10 2 3.0749 × 10 4 1.2232 × 10 3
std4.4754 × 10 5 1.3891 × 10 8 2.6811 × 10 8 6.8205 × 10 6 1.1755 × 10 38 2.8670 × 10 8
F16best−1.0316 × 10 0 −1.0316 × 10 0 −1.0316 × 10 0 −1.0316 × 10 0 −1.0316 × 10 0 −1.0316 × 10 0
mean−1.0316 × 10 0 −1.0316 × 10 0 −1.0316 × 10 0 −1.0316 × 10 0 −1.0316 × 10 0 −1.0316 × 10 0
worst−1.0316 × 10 0 −1.0316 × 10 0 −1.0316 × 10 0 −1.0316 × 10 0 −1.0316 × 10 0 −1.0316 × 10 0
std0.0000 × 10 0 0.0000 × 10 0 0.0000 × 10 0 0.0000 × 10 0 1.9722 × 10 31 1.9722 × 10 31
F17best3.9789 × 10 1 3.9789 × 10 1 3.9789 × 10 1 3.9789 × 10 1 3.9789 × 10 1 3.9789 × 10 1
mean3.9789 × 10 1 3.9789 × 10 1 3.9789 × 10 1 3.9789 × 10 1 3.9789 × 10 1 3.9789 × 10 1
worst3.9789 × 10 1 3.9789 × 10 1 3.9789 × 10 1 3.9789 × 10 1 3.9789 × 10 1 3.9789 × 10 1
std2.7733 × 10 32 2.7733 × 10 32 2.7733 × 10 32 2.7733 × 10 32 3.0815 × 10 33 1.9600 × 10 13
F18best3.0000 × 10 0 3.0000 × 10 0 3.0000 × 10 0 3.0000 × 10 0 3.0000 × 10 0 3.0000 × 10 0
mean3.0000 × 10 0 3.0002 × 10 0 3.0000 × 10 0 3.0000 × 10 0 3.0000 × 10 0 3.0000 × 10 0
worst3.0000 × 10 0 3.0039 × 10 0 3.0000 × 10 0 3.0000 × 10 0 3.0000 × 10 0 3.0000 × 10 0
std0.0000 × 10 0 5.0468 × 10 7 0.0000 × 10 0 0.0000 × 10 0 0.0000 × 10 0 2.0556 × 10 13
F19best−3.8628 × 10 0 −3.8628 × 10 0 −3.8628 × 10 0 −3.8628 × 10 0 −3.8628 × 10 0 -3.8628 × 10 0
mean−3.8370 × 10 0 −3.8628 × 10 0 −3.8628 × 10 0 −3.8628 × 10 0 −3.8628 × 10 0 −3.8628 × 10 0
worst−3.0898 × 10 0 -3.8628 × 10 0 −3.8628 × 10 0 −3.8628 × 10 0 −3.8628 × 10 0 −3.8628 × 10 0
std1.9255 × 10 2 1.9722 × 10 31 1.9722 × 10 31 1.9722 × 10 31 3.1554 × 10 30 3.1554 × 10 30
F20best−3.3220 × 10 0 −3.3220 × 10 0 −3.3220 × 10 0 −3.3220 × 10 0 −3.3220 × 10 0 −3.3220 × 10 0
mean−3.2863 × 10 0 −3.3220 × 10 0 −3.3212 × 10 0 −3.2809 × 10 0 −3.3220 × 10 0 −3.2863 × 10 0
worst−3.2031 × 10 0 −3.3220 × 10 0 −3.2974 × 10 0 −3.1993 × 10 0 −3.3220 × 10 0 −3.2031 × 10 0
std2.9688 × 10 3 3.1554 × 10 30 1.9578 × 10 5 3.1319 × 10 3 7.8886 × 10 31 2.9685 × 10 3
F21best−1.0153 × 10 1 −1.0153 × 10 1 −1.0153 × 10 1 −1.0153 × 10 1 −1.0153 × 10 1 −1.0153 × 10 1
mean−4.7418 × 10 0 −1.0153 × 10 1 −9.7358 × 10 0 −6.4744 × 10 0 −1.0153 × 10 1 −1.0153 × 10 1
worst−2.6305 × 10 0 −1.0153 × 10 1 −2.6829 × 10 0 −2.6305 × 10 0 −1.0153 × 10 1 −1.0153 × 10 1
std1.0831 × 10 1 1.2622 × 10 29 2.5369 × 10 0 1.1149 × 10 1 1.2622 × 10 29 4.2873 × 10 8
F22best−1.0403 × 10 1 −1.0403 × 10 1 −1.0403 × 10 1 −1.0403 × 10 1 −1.0403 × 10 1 −1.0403 × 10 1
mean−6.1650 × 10 0 −1.0403 × 10 1 −-1.0227 × 10 1 −6.7240 × 10 0 −1.0403 × 10 1 −1.0403 × 10 1
worst−1.8376 × 10 0 −1.0403 × 10 1 −5.1288 × 10 0 −2.7519 × 10 0 −1.0403 × 10 1 −1.0403 × 10 1
std1.2604 × 10 1 5.0487 × 10 29 8.9629 × 10 1 1.2263 × 10 1 0.0000 × 10 0 6.5010 × 10 9
F23best−1.0536 × 10 1 −1.0536 × 10 1 −1.0536 × 10 1 −1.0536 × 10 1 −1.0536 × 10 1 −1.0536 × 10 1
mean−6.7622 × 10 0 −1.0536 × 10 1 −1.0536 × 10 1 −6.2481 × 10 0 −1.0536 × 10 1 −1.0133 × 10 1
worst−2.4217E × 10 0 −1.0512 × 10 1 −1.0536 × 10 1 −2.4217 × 10 0 −1.0536 × 10-3.8354 × 10 0
std1.4389 × 10 1 1.9658 × 10 5 7.8886 × 10 29 1.4417 × 10 1 2.8399 × 10 29 2.3087 × 10 0
Table 6. Average Rankings of the algorithms (Friedman) on fixed multimodal benchmark functions.
Table 6. Average Rankings of the algorithms (Friedman) on fixed multimodal benchmark functions.
AlgorithmRanking
PSO4.85
ABC2.95
DE3.3
CA4.6
BSA2.35
RSCBSA2.95
Table 7. Test statistical results of Wilcoxon rank-sum test.
Table 7. Test statistical results of Wilcoxon rank-sum test.
BenchmarkRSCBSA vs. BSARSCBSA vs. PSORSCBSA vs. ABCRSCBSA vs. DERSCBSA vs. CA
Hp-ValueWinnerHp-ValueWinnerHp-ValueWinnerHp-ValueWinnerHp-ValueWinner
F111.2118 × 10 12 +11.2118 × 10 12 +11.2118 × 10 12 +11.2118 × 10 12 +11.2118 × 10 12 +
F213.0199 × 10 11 +13.0199 × 10 11 +13.0199 × 10 11 +13.0199 × 10 11 +13.0199 × 10 11 +
F313.0199 × 10 11 +13.0199 × 10 11 +13.0199 × 10 11 +13.0199 × 10 11 +13.0199 × 10 11 +
F413.0199 × 10 11 +13.0199 × 10 11 +13.0199 × 10 11 +13.0199 × 10 11 +13.0199 × 10 11 +
F511.0763 × 10 2 +13.0199 × 10 11 13.0199 × 10 11 11.2493 × 10 5 +13.0199 × 10 11 +
F613.0199 × 10 11 12.3692 × 10 11 13.0199 × 10 11 11.2118 × 10 12 13.0199 × 10 11 +
F713.0199 × 10 11 +12.0283 × 10 7 +13.0199 × 10 11 +13.0199 × 10 11 +13.0199 × 10 11 +
F812.2076 × 10 11 13.0104 × 10 11 +11.4248 × 10 11 12.5416 × 10 11 13.8481 × 10 3
F911.2118 × 10 12 +11.1661 × 10 12 +11.0566 × 10 12 +11.2118 × 10 12 +11.2118 × 10 12 +
F1011.2118 × 10 12 +11.5702 × 10 13 +19.7992 × 10 13 +11.6853 × 10 14 +11.2118 × 10 12 +
F1114.5700 × 10 12 +11.2118 × 10 12 +11.2078 × 10 12 +0NaN=11.2118 × 10 12 +
F1213.0199 × 10 11 12.4291 × 10 11 13.0199 × 10 11 11.2118 × 10 12 13.0199 × 10 11 +
F1313.0199 × 10 11 12.6537 × 10 11 13.0199 × 10 11 11.2118 × 10 12 13.0199 × 10 11 +
F140NaN=19.7829 × 10 13 +11.6853 × 10 14 =12.7085 × 10 14 +11.1642 × 10 12 +
F1512.1633 × 10 11 09.7028 × 10 1 =14.1804 × 10 9 +18.8803 × 10 6 +11.2050 × 10 10 +
F160NaN=11.6853 × 10 14 =11.6853 × 10 14 =11.6853 × 10 14 =11.6853 × 10 14 =
F1716.6113 × 10 4 =13.8943 × 10 13 =13.8943 × 10 13 =13.8943 × 10 13 =13.8943 × 10 13 =
F1814.1865 × 10 2 =14.1865 × 10 2 =04.6889 × 10 1 =14.1865 × 10 2 =14.1865 × 10 2 =
F190NaN=12.7085 × 10 14 +11.6853 × 10 14 =11.6853 × 10 14 =11.6853 × 10 14 =
F2016.2958 × 10 4 15.4952 × 10 3 =13.5049 × 10 13 17.8511 × 10 11 09.6974 × 10 1 =
F2112.1150 × 10 6 =16.7082 × 10 5 +12.1150 × 10 6 =11.9600 × 10 4 +12.5975 × 10 2 +
F2211.4331 × 10 4 =11.2791 × 10 8 +11.4992 × 10 5 =11.1131 × 10 5 +13.3861 × 10 8 +
F2313.1216 × 10 4 12.0775 × 10 6 +12.7674 × 10 3 12.7674 × 10 3 16.7273 × 10 7 +
+/−/=9/7/714/4/59/7/712/6/517/1/5

Share and Cite

MDPI and ACS Style

Zhou, C.; Li, S.; Zhang, Y.; Chen, Z.; Zhang, C. Enhancing Backtracking Search Algorithm using Reflection Mutation Strategy Based on Sine Cosine. Algorithms 2019, 12, 225. https://doi.org/10.3390/a12110225

AMA Style

Zhou C, Li S, Zhang Y, Chen Z, Zhang C. Enhancing Backtracking Search Algorithm using Reflection Mutation Strategy Based on Sine Cosine. Algorithms. 2019; 12(11):225. https://doi.org/10.3390/a12110225

Chicago/Turabian Style

Zhou, Chong, Shengjie Li, Yuhe Zhang, Zhikun Chen, and Cuijun Zhang. 2019. "Enhancing Backtracking Search Algorithm using Reflection Mutation Strategy Based on Sine Cosine" Algorithms 12, no. 11: 225. https://doi.org/10.3390/a12110225

APA Style

Zhou, C., Li, S., Zhang, Y., Chen, Z., & Zhang, C. (2019). Enhancing Backtracking Search Algorithm using Reflection Mutation Strategy Based on Sine Cosine. Algorithms, 12(11), 225. https://doi.org/10.3390/a12110225

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop