1. Introduction
Combinatorial optimization is devoted to the mathematical process of searching for the optimal solution (maxima or minima) of an objective function with a discrete domain of decision variables. The possible number of solutions for a combinatorial optimization problem is equal to
, where
is the discrete design domain vector and
represents the number of design variables [
1]. Therefore, the optimization problem becomes more computationally difficult to be solved when the number of design variables increases. Accordingly, many combinatorial optimization problems are hard to solve within deterministic polynomial time (or NP-hard). A Travelling Salesman (or TSP) is a typical example of this type of optimization problem where the number of cities to be visited is given and the shortest path needs to be determined [
2]. As the number of cities increases, the number of possible solutions increases too and this leads to the computational complexity of the problem, where it is not possible to enumerate all these solution possibilities with limited computation resources, such as memory size or processor speed. Furthermore, many structural design problems fall under this category of optimization problems because of the discrete nature of their design domain. It makes the objective function a multimodal (non-convex) function. As a consequence, the design space will have more than one optimal solution which then turns the structure design into an NP-hard design optimization problem. Trusses, beams, and composite structures are good examples of real-life combinatorial optimization problems [
3]. In recent years, several studies have been published regarding the optimization of composite laminated structures [
4]. This interest of the research community has been driven by the shared property of high-specific strength (or strength/weight ratio), thermal stability, corrosion resistance, and the high impact strength of the composite materials [
5]. Hence, to solve such problems, many optimization techniques have been developed.
The Ant Colony Optimization (ACO) algorithm demonstrated a significant performance improvement in solving NP-hard combinatorial optimization problems. The Traveling Salesman Problem (TSP) is a good example of such problems and it is solved using an early version of ACO [
6]. The improvements in subsequent ACO algorithms focused on enhancing the algorithm variants to yield better searching and computational performance. As a result of the improved algorithm performance, many new applications of ACO appeared, such as the probabilistic TSP using estimation based ACO in Weiler and Biesinger’s [
7] work. Gambardella and Dorigo [
8] used a hybrid ACO with a novel local search tool to solve the Sequential Ordering Problem (SOP). Zheng and Zecchin [
9] introduced a novel approach of ACO to optimize a water distribution system design. Furthermore, the optimization of the space truss design, using improved ACO, was demonstrated by Kaveh and Talatahari [
10].
In the field of structural design problems, a well-known optimization problem of Stacking Sequence Design (SSD) of composite laminate has been solved. It is done by using ACO to determine the optimal stacking sequence design for maximum critical buckling load factor or natural frequency [
11]. Aymerich and Serra [
12] examined the ACO performance in solving the Stacking Sequence Design problem and he compared a modified standard ACO performance with Genetic Algorithm (GA) and Tabu Search (TS). He found that ACO performed much better than GA and TS in terms of solution cost and quality. Rama Mohan Rao [
13] presented a hybrid ACO–TS algorithm to optimize the stacking sequence of a composite laminate subjected to bidirectional compression loading. He concluded that ACO is an effective optimization technique if combined with an appropriate local searching tool. Bloomfield and Herencia [
14] conducted a comparison study of three meta-heuristics of GA, ACO, and Particle Swarm Optimization (PSO) to determine the optimal stacking sequence composite laminate. Based on the results of this comparison study, ACO was found to outperform GA and PSO algorithms in the field of Stacking Sequence Design (SSD). This remarkable performance of ACO in solving such NP-hard combinatorial optimization is expected where it was designed to solve discrete optimization problems [
2].
The literature of SDD optimization is full of valuable contributions that advanced our knowledge in this specific area. This subject of research still evaluates debates on various issues, not only how we could achieve the desired maximum buckling load, but also how we could make this happen efficiently at the lowest possible computational cost. In this regard, selecting the optimization algorithm to solve a SDD problem was not addressed clearly, i.e., most published studies do not explain why a certain algorithm has been selected. Therefore, the selection of an efficient optimization algorithm needs to develop a substantial compromise criterion to determine which MH becomes a cost-effective solution for a SDD optimization problem. Well-defined selection criteria will help to make the appropriate selection in this regard. Additionally, investigating possible improvements of candidate MH exploration and exploitation features could lead to finding a new cost-effective solution that defeats the solution obtained by the original MH.
In the current study, a comprehensive assessment criterion has been developed to assess the performance of the meta-heuristic (MH) solutions in the domain of structural design. Thereafter, the proposed criterion was employed to compare five different variants of Ant Colony Optimization (ACO). This was done based on a well-known structural optimization problem of laminate Stacking Sequence Design (SSD) [
1].
The Enhanced HCF-ACO (EHCFACO) variant presented here is as novel optimizer for structural design optimization problems. The standard HCF-ACO variant was never examined before as a solver of any structural design optimization problems, to the best of our knowledge, even though it has been used successfully to solve other NP-hard combinatorial optimization problems. Furthermore, a comprehensive performance assessment criterion for MHs was used to solve structural design problems. The new criterion introduced new performance measures such as a fitness–distance correlation factor, performance rate, and solution quality. Both the proposed MH and the comprehensive assessment criterion presented here will help engineers optimize their designs more efficiently.
This paper is structured as follows: firstly, it presents a review of standard ACO followed by a brief description of other considered ACO variants; secondly, the new optimization approach is explained in detail; then, the performance assessment criterion is introduced, followed by the numerical experiments section; next, the results of the performance survey are discussed; finally, the paper concludes with a summary of the study’s research contributions, limitations, and prospective directions for future research.
2. Ant Colony Optimization Algorithms (ACOs)
Dorigo [
6] developed the basic Ant Colony Optimization (ACO) algorithm, or Ant System (AS), which is a metaheuristic approach inspired by the collaborative work of ants in finding the source of food. The ants cooperate to find the best possible path from their colony to the food source,
Figure 1. Ants search cooperatively using pheromone trails. During their searching tour, the ants communicate by depositing a certain amount of a substance, called the pheromone, on their way to the food site. The following group of ants tends to follow the paths with higher pheromone concentration. Over time, the less selected paths gradually lose information due to the pheromone evaporation process.
The virtual ants traveling and selecting paths can be interpreted as a probabilistic selection of certain nodes, of which they are part of the solution, in the path based on the pheromone value. The ACO general procedure is illustrated in Algorithm 1.
Algorithm 1 Ant Colony Optimization procedure. |
1: | Initialization |
2: | While (termination criteria not satisfied) Do |
3: | Construct Solutions Table by Ants |
4: | Local Search (optional) |
5: | Global Pheromones Updating |
6: | End ACO algorithm |
To understand the mathematical interpretation of ACO, there is a need to go through each step of the ACO procedure shown in Algorithm 1. ACO starts with initial values of the pheromone trail
, set to a small value for all ant trails as this gives all nodes
of the design variable
, an equal probability of selection. Next, each ant starts to construct its own solution by applying the rule of selection, which has the following general form:
represents the probability of selecting the path
for the
ant,
is the updated pheromone trail,
denotes the value of heuristic information for each feasible solution
,
indicates the neighborhood nodes of the
ant, when it is located at node
and
are the amplification parameters for pheromone trials and the influence of heuristic information on the algorithm behavior, respectively [
15]. At the end of each tour, all the pheromone trails are updated through two steps of pheromone evaporation and depositing, according to the following formula:
where
is the evaporation rate,
, and
is the amount of deposited pheromone by ant
that could be determined as:
where
is a constant and
represents the distance traveled by ant
. Equation (3) is the basic form of the pheromone trail updating which is used to solve the TSP optimization problem and it could be implemented in more general form:
where
,
are the worst and the best values of the objective function
obtained by
ants in tour
and
is the global pheromone scaling factor [
16]. Eventually, the ACO loop continues until one of the termination conditions is met.
In a SDD optimization problem, the thickness of each ply (equivalent to the distance between the cites in TSP) is assumed to be constant, so the heuristic information value,
, will be constant all over the ant tours
t which simplifies the probability of selection, in Equation (1), into:
Local search
The procedure of the ACO algorithm includes the option of improving the intensification feature of the ACO algorithm by adding some local search algorithms or movements that could improve the search of the solution neighborhood [
2].
2.1. Elitist Ant Colony (EACO)
Gambardella and Dorigo [
8] introduced an improved version of the ACO algorithm that uses the elitism strategy. The idea behind this strategy is a reinforcement of the best solution path found once the algorithm is initialized. The rule of pheromone updating for EACO is written as follows:
where the reinforcement of selection probability of the best path
occurs by adding the value of
where
e is the weighting parameter and it represents the number of elitist ants [
13].
2.2. The Rank-Based Ant Colony Optimization (RBACO)
Bullnheimer and Hartl [
17] proposed a new extension of the ACO that enhances the performance of the original EACO by ranking the ants based on their path length. The deposited value of pheromone decreases according to its rank index,
. Moreover, only the best ants,
, will be updated, which prevents the over-concentration of pheromones on local optima paths chosen by other ants. Hence, the pheromone updating rule of RBACO is:
2.3. Max-Min Ant Colony (MMACO)
Previous ACO algorithms used the strategy of reinforcing only the best-found paths. This strategy could cause the excessive increase of pheromone values on optimal local paths causing all other ants to follow this path. To overcome this drawback, Stützle and Hoos [
18] proposed a modified version of ACO that limits the pheromone values to a specific interval,
. In addition, the initialization of pheromone value is set to the upper limit of the pheromone interval, with a small evaporation rate to increase the algorithm search diversification. The pheromone rule is:
and
.
where
values are determined by the following formulas:
where
denotes the probability of the best solution, it has a value greater than
, while
represents the number of ants.
2.4. Best-Worst Ant Colony (BWACO)
Zhang and Wang [
19] presented BWACO as an extension of MMACO, where the algorithm exploitation capability benefits from both best and worst solutions. During the search tour, the pheromone trail update uses the positive return of the best solution and the negative one generated by the worst solution. The pheromone updating rule can be written as:
where
is a coefficient that has value within
interval and it could be noticed that BWACO became MMACO if
.
2.5. Hyper-Cube Framework ACO (HCFACO)
The different algorithms of ACO build a limited hyperspace of the pheromone values. The Hyper-Cube Framework of ACO algorithms, proposed by Blum in 2001, generates a binary convex hull hyperspace from pheromone values for the feasible solutions. In other words, the values of the pheromone vector,
, are limited to the interval [0, 1], and this is carried out by changing the pheromone update rule. The following formula expresses the rule of pheromone updating in HCFACO:
where:
, and is the number of ants follow the same best path.
HCFACO algorithms overcome the undesirable problem of different behavior of standard ACO algorithms when the same objective function is scaled, which affects the algorithm robustness. Moreover, it reduces the search effort and improves the algorithm search diversification [
2]. Lastly, it is worth mentioning that the HCF update rule is not limited to standard ACO algorithm (or Ant System (AS)) as it can also be used with MMACO, where the maximum and minimum limits of MMACO pheromone trail are set to be 0 and
, respectively [
20].
3. Enhanced Hyper-Cube ACO Algorithms
Dorigo experimentally observed that using local search techniques can improve the overall performance of the ACO [
2,
21]. Local search can be carried out by hybridizing the ACO with local search algorithms such as Tabu search or using permutation operators to explore the solution neighborhood [
12,
22]. The commonly used operators in SSD optimization problem are two-points permutation and swap. Two-points permutation means selecting two bits in the solution string and reversing the order of the bits in between [
1]. The swap operator is used to switch the position of two randomly selected bits of the solution string [
23].
The HCFACO algorithms presented here adopted two other permutation operators to perform the algorithm enhancement. The first operator is called a bit flip (also known as single point mutation), which is used successfully with Permutation Genetic Algorithm [
1]. The second operator is inspired by using one of the Tabu Search movements named the insertion [
2]. The proposed Enhanced HCFACO procedure is listed in Algorithm 2.
Algorithm 2 Enhanced HCFACO procedure |
| Input: | length of the design domain vector, maximum number of iterations, () ) maximum convergence rate. |
| Output: | ) number of optimal solution iterations. |
| Initialization: |
1: | initialize the EHCFACO main loop counter. |
2: | , generate initial population. |
3: | , evaluate the initial population. |
4: | , update global pheromone trial. |
5: | While (termination criterion not satisfied) Do: |
6: | update loop counter. |
7: | , construct the new solution table. |
8: | evaluate the objective function of the new solution. |
9: | , determine the current optimal solution. |
| - -
Local search:
|
10: | | , insertion movement. |
11: | | , bit flip movement. |
12: | | , update the current optimal solution. |
| - -
Update global optimal solution:
|
13: | | |
| - -
Global pheromone updating:
|
14: | | |
| - -
Convergence check:
|
15: | | , number of global optimal solution occurrences. |
16: | | , imposed convergence rate has been achieved. , all artificial ants followed the same path. |
17: | End while |
18: | Print. |
19: | End |
The Enhanced HCFACO Algorithms starts by defining the standard ACO parameters such as the maximum number of iterations , number of ants , number of design variables , the initial pheromone trail (), and evaporation rate (). In addition, the solution convergence rate counter is imposed and its value determines whether the convergence rate is slow or fast. When the ACO loop starts, all solution edges have the same deposited pheromone trail , which gives all nodes the same probability of selection to be a part of the feasible solution. The artificial ants, , start building the solution table, , by randomly choosing a node di on their way to build the solution vector . Next, the evaluation of the solutions table is carried out by calling the objective function, and the obtained values are stored in matrix. The best solution of the stacking sequence design has the maximum value of the objective function listed in matrix of the current iteration. The best solution of each iteration is stored in the best solution matrix ). Thereafter, the global pheromone trail update is performed as described in the Hyper-Cube Framework of ACO in Equation (11).
The local search actions are enforced as soon as the best solution of the current tour is determined. Following this, a comparison of the generated solutions with the best solution obtained so far is made. The best solution matrix is then updated if any improvement is detected. Finally, the HCFACO loop continues until the termination criteria are met. The global optimal solution is determined as the best solution matrix member with the maximum value of the objective function.
4. Performance Evaluation
The time required by an algorithm to find the global optima is widely used to evaluate its performance [
24]. However, a single performance measure cannot reflect the effectiveness of the algorithm in exploring the design space or determining solution quality. In the current study, three different groups of performance measures have been applied to ensure a fair evaluation of the proposed algorithm.
4.1. Computational Effort
In addition to the elapsed time, literature has shown that other measures can be used to measure computational effort. The first is the price
, which is defined as the number of objective function evaluations within a search run and reflects the computational cost of the search process. The second measure is practical reliability
and it is defined as the percentage of runs that achieve practical optima (
), at a specific run. Practical optima is defined as the solution with
error in the best possible solution [
1]. The last is the normalized price
, which is defined as the ratio of price and practical reliability [
1,
25,
26]. Finally, the performance rate measure, Prate, is also considered to link the computation effort with the number of function evaluations [
24].
4.2. Solution Quality
The solution quality of an algorithm can be measured by determining the absolute error between the current solution and the best-known global solution [
1,
25,
26].
4.3. Fitness Landscape Analysis
The design space of a combinatorial optimization problem can significantly affect the search performance of an algorithm. The notion of fitness landscape appeared in literature as an answer to the question of “what does the design space look like?”. The fitness landscape is defined by the feasible solutions set, the objective function (fitness), and the structure of the solution neighborhood. To find the connection between the fitness landscape and the problem hardness, Jones and Forrest [
27] introduced a Fitness Landscape–Distance Correlation (FDC) to determine the hardness of optimization problems to be solved using Genetic Algorithm (GA). The distance mentioned here is defined as the number of movements that should be imposed on a Solution
to eliminate dissimilarity with the optimal solution
. The proposed correlation by Jones is computed using the correlation factor,
:
where
indicates the
and
are the standard deviation of
and
, respectively. The values of the correlation coefficient r are limited to interval
where negative values are desirable for maximization and indicate better searching performance. Finally, using the scattering of fitness versus the distance to the global optima can reveal valuable information about
of the optimization problem solved by an algorithm [
18,
26].
5. Numerical Experiments
To demonstrate the performance of the new approach, we selected a well-known NP-hard combinatorial optimization problem in the field of composite laminated structures. The optimization objective is maximizing the critical buckling load of composite laminated plate exposed to bidirectional compression loading. The decision variables are the fiber orientation of each composite layer (lamina) which form the optimal stacking sequence of the laminate (a group of layers). To employ ACO as an optimization algorithm for a SSD optimization problem, there is a need to understand specific problem characteristics such as solution representation, constraints, and objective function formulation. In meta-heuristic algorithms, the solution (stacking sequence) takes the form of a bit string that consists of a combination of plies with the available angle fiber orientations (e.g., and ). The different solutions have integer coding with 1, 2, and 3 numbers, which represent the three possible fiber orientations, respectively. For instance, the laminate with stacking sequence describes the laminate of fiber orientations.
The simplicity of using an integer representation along with significant performance gains made it the most widely used method in meta-heuristic optimization algorithms for composite laminated design. The buckling load factor lb for simply supported rectangular laminated plate subjected to bi-axial loading is determined as follows:
where
denotes the bending stiffness,
is the axial loading in x-direction,
is the axial loading in y-direction,
and
are the number of half waves in
directions. The critical buckling load factor
is defined as the minimum obtained value of
. The critical values of
and
are linked to different factors such as laminate material, the number of plies, loading conditions, and the plate aspect ratio. In uniaxial loading and a simply supported plate, the critical buckling load occurs when
whereas in biaxial the critical buckling loads, it needs to be determined as the minimum value of
[
1,
13]. Finally, the constraints in stacking sequence optimization with constant laminate thickness t could be imposed as follow:
- -
Symmetry constraint is enforced by optimizing half of the laminate.
- -
Balancing constraint is enforced by selecting for the standard fiber orientation set of and 90.
- -
Only ply orientations are needed to describe laminate as a result of balancing constraints.
- -
Contiguity constraint is handled by using the penalty parameter .
- -
The critical buckling load factor objective function
could be formulated as:
To compare the performance of the proposed algorithm alongside the other ACO algorithms, we implemented all the algorithms presented here using MATLAB R2019b software. The benchmarking problem from the literature of Stacking Sequence Design optimization is accredited to Le Riche and has been used by previous studies [
1,
23]. The original problem describes a simply supported plate subjected to an in-plane biaxial loading as shown in
Figure 2.
The thickness of each ply
is assumed constant, and the ply orientations are limited to
and
sets of angles. The number of plies
is constant. The required properties, dimensions, and loading conditions are listed in
Table 1 and
Table 2. The objective function is set to maximize the critical buckling load. The constraints are integrated into the solution (e.g., balanced laminate, symmetrical, etc.). The implemented ACO algorithms were executed on the same computer for the same number of experiments;
. This number was used to overcome the stochastic behaviour of meta-heuristic algorithms [
1]. Furthermore, this number of experiments was conducted over ten different random generating seeds of
. Then, the average of the performance measures values was used in the comparison of different ACO algorithms.
Lastly, all ACO algorithms were examined at two different levels of convergence rate, slow and fast. The slow rate forces the algorithm searching loop to stop after 56 iteration without improvement, while the fast rate needs just 10 iterations to be terminated [
1].
5.1. ACO Parameters Setting
To ensure a fair assessment of the ACO algorithms performance, the following standard ACO parameters were assumed for all implemented algorithms: number of Ants , the maximum number of iterations , evaporation rate , the parameter of the pheromone trail relative importance , and initial one trail (except for MMACO and BWACO algorithms were ). Best solution probability was for MMACO and BWACO algorithms and lastly the coefficient of worst solution pheromone trail was for BWACO algorithm only.
5.2. Termination Criteria
All algorithms stop as soon as one of the following conditions are satisfied:
- -
If there is no improvement in the solution after 10 or 56 iterations.
- -
If the number of iterations exceeds 150 and the best solution is equal to the worst solution (means all artificial ants following the same path).
- -
If a maximum number of iterations has been generated.
6. Results and Discussion
The case study described in the previous section was optimized using nine different algorithms: standard ACOA, EACO, RBACO, MMACO, BWACO, HCF/EHCF for both ACO and MMACO algorithms. Analysis of the algorithm’s performance will be divided into two parts. First, the performance of ACO algorithms with Hyper-Cube Framework will be assessed. The second part is dedicated to the comparison of EHCFACO algorithm with the rest of ACO algorithms.
6.1. Hyper-Cube Framework ACO Algorithms Results Analysis
Referring to
Section 2.5, the Hyper-Cube Framework (HCF) can be applied for both versions of the standard ACOA and MMACO. Hence, this part of the analysis is devoted to determining which version of both ACO algorithms, with HCF and EHCF, could exhibit a better performance. The performance measures for the original ACOA, MMACO, HCFACO, HCFMMACO, EHCFACO, and EHCFMMACO are listed in
Table 3. The performance values listed in
Table 3 reveal that applying HCF to the ACOA positively affected the overall performance of ACO. The average practical reliability increased by
and the normalized price declined from 51.17 to
for the fast convergence rate and from
to
for the slow one. The performance rate doubled at the slow rate while remaining the same for the fast. The FDC correlation coefficient
decreased slightly for both levels of convergence.
Further improvement of HCFACO performance was acquired when the proposed local search movements were imposed. The average practical reliability became more than twice of the standard ACO, and the normalized price decreased more to hold at
instead of
and
instead of 181.2 for both convergence rates. The performance rate was slightly increased, and the FDC correlation coefficient r was partially improved. On the contrary, HCFMMACO performed poorly compared to the original MMACO. However, the performance of MMACO improved when the Enhanced HCF was applied, but the computational effort became more costly. Based on these results, we conclude that EHCFACO delivers an inexpensive solution with significant performance. Eventually, the solution convergence of the algorithms mentioned above was plotted,
Figure 3: solution convergence of ACO-MMACO and their Hyper-Cube Framework variants, for the same selected experiment (seeds = 75).
6.2. Other ACO Algorithms Results Analysis
All ACO algorithms successfully found the best-known value of the maximum critical buckling load factor,
. A different sample of optimal SSD obtained by different ACO algorithms is listed in
Table 4. The solution convergence of EACO, RBACO, and BWACO for a selected experiment (seeds = 75) is graphically illustrated in
Figure 4.
It is observed from
Table 4 that the optimal stacking sequence design followed the same pattern of switching between two groups of
and
fiber orientations which confirm the results of previous studies [
1,
23].
On the other hand, the solution convergence plot in
Figure 4 illustrates that both RBACO and BWACO algorithms develop gradual search trends on their way to the optima whereas EACO and EHCFACO algorithms smoothly converge to the global optima. Further, the numerical experiments confirm the fluctuation of ACO algorithms in finding the global optimal solution due to their stochastic nature, as illustrated in
Figure 5.
According to the introduced performance assessment criteria in
Section 5, the average values of different performance measures of reliability, performance rate, solution quality, normalized price, and searching effort coefficient were determined for EACO, RBACO, BWACO at fast and slow convergence rates. These results, alongside the EHCFACO results, are plotted in
Figure 6,
Figure 7,
Figure 8,
Figure 9,
Figure 10,
Figure 11 and
Figure 12 to provide a sensible comparison of the performance evaluation of the proposed algorithm with other ACO algorithms. The average practical reliability of the algorithms is introduced in
Figure 6. Both EACO and RBACO algorithms show low practical reliability values, with
and
, respectively, while BWACO presented better value at the slow rate of convergence, but it poorly performed at the fast rate. The EHCFACO algorithm exhibited significant reliability values of 89.6–98.95%. Furthermore, it demonstrated the highest performance rate measure (0.012–0.035), see
Figure 9. The solution quality results of the algorithms are summarized and depicted in
Figure 10, which reveals that all ACO algorithms produce an excellent solution quality for this particular case study.
In terms of solution cost, the normalized price results plotted on the scatter chart are illustrated in
Figure 7. As mentioned before, the normalized price measure reflects the balance between the solution cost and the reliability. Thus, it is quite clear that EHCFACO outperformed other ACO algorithms. BWACO comes second at a slow convergence rate, whereas RBACO and EACO deliver a costly solution.
The numerical experiments results show that ACO variants with high reliability need more computational time to find the global optimal solution while the number of iterations required to find it is low. The Fitness-Distance Correlation values demonstrated the complexity of such a NP-hard optimization to be solved where ACOs with high reliabilities have low r values. Eventually, applying the Hyper-Cube Framework (HCF) to standard ACO has a significant influence on the overall performance of ACO. In addition, imposing local search movements, as an enhancement of exploitation effort, helped HCFACO to deliver a cost-effective solution.
7. Conclusions
A comprehensive MHs performance assessment criterion was developed in the current study. In addition to the computation time and statistical matrices, several new measures such as the practical reliability, price (computational cost), normalized price, performance rate, solution quality, and fitness landscape analysis were included. Additionally, two different convergence rates were imposed on examining the MHs at both the slow and the fast rates. The reproducibility of the numerical experiments’ results was also considered within the procedure of MH assessment. Thereafter, the proposed criterion was employed to compare five different variants of Ant Colony Optimization (ACO). The proposed measures demonstrated a comprehensive assessment of the compared ACOs performance. The initial results of the comparison study revealed that the Hyper Cube Framework (HCF) ACO variant outperforms the others. Consequently, an investigation of further improvement led to introduce an enhanced version of HCFACO (or EHCFACO). The new variant has advanced intensification features that use both the insertion and bit-flip movements to enhance the local search effort. Eventually, the EHCFACO variant was compared with other ACO variants and it exhibited a significant performance.
Based on the findings of the proposed assessment scheme presented through this research work, it was noticed that MHs with high-reliability solutions need more time to find the global optima but need a relatively small number of computation iterations. This observation led to another important remark, i.e., the MHs that show good performance are hardly exploring the design space. It is the case where their Fitness-Distance Correlation, r, and figures become lower than the other MHs. The difference in values was noticed, but it does not negatively impact the overall performance of the designated MHs, except in increasing the computational time. Furthermore, it was observed that applying the Hyper-Cube Framework (HCF) to standard ACO has a significant influence on the overall performance of ACO. Moreover, imposing the local search movements, as an enhancement of exploitation effort, helped HCFACO to deliver a cost-effective solution. These improvements, in ACO performance, are in line with the suggestions made by the previous studies.
The MHs literature is full of studies that approve the noticeable impact of the parameters’ fine-tuning on the MHs performance; for all examined MHs, the parameters have been assumed constant. This perceived ignorance of such an aspect refers to the associated computation cost of this repetitive investigation, and such an exhaustive practice would diminish the scope of this paper significantly.
The current study allows engineers the access to an efficient variant of ACO MH to optimize their structure designs at low computational cost. Furthermore, the comprehensive assessment criterion makes it easier for the engineer to decide between any other obtainable MHs. Eventually, the obtained data of the numerical experiments conducted in this paper are available to other researchers to assess and reproduce the results obtained in
Supplementary Materials.