1. Introduction
With tremendous advances in computational sciences, information technology, and artificial intelligence, design optimization becomes increasingly popular in many engineering subjects, such as mechanical, civil, structural, aerospace, automotive, and energy engineering. It helps to shorten the design-cycle time and to identify creative designs that are not only feasible but also progressively optimal, given predetermined design criteria.
At the outset of design optimization, running a gradient-based algorithm with a multi-start process proves to be very successful in finding the global optimum of simple problems when gradient information is available [
1]. While under the pressure of being faced with increasingly complex optimization problems in which derivative information is unreliable or unavailable, researchers gradually focus on the development of derivative-free optimization methods [
2] and metaheuristic methods to address this issue. Followed by Glover’s convention [
3], modern metaheuristic algorithms such as simulated annealing (SA) [
4], genetic algorithms (GA) [
5,
6], particle swarm optimization (PSO) [
7], and ant colony optimization (ACO) [
8] have been applied with good success in solving complex nonlinear optimization problems [
9,
10]. The popularity of these nature-inspired algorithms lies in their ease of implementation and the capability to obtain a solution close to the global optimum. However, for many real-life design problems, more than thousands of calls for high-fidelity simulations (for example, computational fluid dynamics simulation) may be executed to seek a near-optimal solution. This is the overwhelming part of the total run time required in the design cycle. Thus, it is desirable to retain the appeal of metaheuristic algorithms on a global search while replacing as many as possible calls to the solver with evaluations on metamodels for the purpose of less computational cost [
11].
The typical techniques for metamodel building include Kriging [
12], polynomial response surface (PRS) [
13], radial basis function (RBF) [
14], artificial network (ANN) [
15], etc. Among them, PRS and ANN are regression methods that have advantages in dealing with convex problems; Kriging and RBF belong to interpolation methods that are more appropriate for nonconvex or multi-modal problems [
16]. Therefore, metamodels have been successfully employed to assist evolutionary optimizations [
17,
18,
19] and the PSO method. For example, Tang et al. [
20] proposed a hybrid surrogate model formed from a quadratic polynomial and a RBF model to develop a surrogate-based PSO method and applied it to solve mostly low-dimensional test problems and engineering design problems. Regis [
21] used RBF surrogates on PSO to identify the most promising trail position surrounding the current overall global best position for solving a 36-dimensional bioremediation problem. However, the inherent nature of the PSO method leads to an extremely large number of calls for function evaluations, which might be prohibitive in simulation-based optimization.
In this paper, an adaptively integrated swarm intelligence-metamodelling technique (ASIM) is proposed, which combines multi-level search and model management during the entire optimization process. It orients the solution of the approximate model to the global optimum with a smaller number of iterations of analyses and achieves a higher level of efficiency than conventional approximation methods. Meanwhile, model management in the optimization process has been established, which integrates an adaptive trust-region strategy with a space reduction scheme implemented in the multipoint approximation method (MAM) framework. The model management has been able to facilitate the optimization process and to improve robustness during iterations. It especially has allowed a small perturbation to be assigned to the current position in case of no updates for the optimal position. The developed ASIM makes full use of the global-exploration potential of PSO and local-exploitation advantage of MAM to efficiently and accurately seek the global optimal solution with low computational cost. In comparison to the results by other algorithms such as conventional MAM, particle swarm optimization [
22], hybrid cuckoo search [
23], water cycle algorithm [
24], etc., the superiority of ASIM has been demonstrated in terms of computational expense and accuracy throughout three case studies.
2. Brief Review of the Multipoint Approximation Method (MAM)
The MAM [
25,
26] was proposed to tackle black-box optimization problems and has gained continuous development in recent years, e.g., Polynkin [
27] enhanced MAM to solve large-scale optimization problems, one of which is the optimization of transonic axial compressor rotor blades; Liu [
28] implemented discrete capability into MAM. Recently, Caloni [
29] has applied MAM to solve a multi-objective problem. Based on a response surface methodology, multipoint approximation method (MAM) aims to construct mid-range approximations and is suitable to solve complex optimization problems owing to (1) producing better-quality approximations that are sufficiently accurate in a current trust region and (2) affordability in terms of computational costs required for their building. These approximation functions have a relatively small number (
where
N is number of design variables) of regression coefficients to be determinedm and the corresponding least squares problem can be solved easily [
25].
In general, a black-box optimization problem can be formulated as follows:
where
refers to the vector of design variables;
and
are the given lower and upper bounds of the design variable
;
N is the total number of design variables;
is the objective function;
is the
jth constraint function and
M is the total number of the constraint functions.
In order to represent the detailed physical model using the response functions and to reduce the number of calls for the response function evaluations, the MAM replaces the optimization problem with a sequence of approximate optimization problems as follows:
where
and
are the functions which approximate the functions
and
defined in Equation (
1);
and
are the side constraints of a trust sub-region; and
k is the iteration number.
Compared with the time spent by evaluation of the actual response functions
, the selected form of approximate functions
remarkably reduces the computational expense and adequately improves the accuracy in a current trust region. This is achieved by appropriate planning of numerical experiments and use of the trust region defined by the side constraints
and
. Once the current suboptimization problem is solved, the suboptimal solution becomes the starting point for the next step. Meanwhile, the move limits are modified and the trust region is resized [
25,
26]. Based on this information, the metamodel is updated in the next iteration until eventually the optimum is reached.
The process of metamodel building in MAM can be described as an assembly of multiple surrogates into one single metamodel using linear regression. Therefore, there are two stages of metamodel building.
In the first stage, the parameter
of an individual surrogate
is determined by solving a weighted least squares problem involving
n fitting points as
where
denotes the weighting parameters and
F is the original function that needs to be approximated. Here, the selection of weighting factors
should reflect the quality of the objective function and the location of a design point with respect to the border between the feasible and the infeasible design subspace [
30], which are defined as
where
are user-defined constants, where, here,
and
are used;
is the starting point in the
kth iteration; and
is the
ith design point in the fitting points. With this definition, a point with a larger objective function has a smaller weighting coefficient component
. For a constraint function
, a point that is much closer to the boundary of the feasible region of
is given a larger weighting coefficient component
. For building a surrogate of the objective function
, the weighting coefficient
only considers the component
. However, for building a surrogate of the constraint function
, the weighting coefficient
will also take the constraint component
into consideration.
It should be noted here that, in MAM, both the objective and constraint functions will be approximated by Equation (
3). The simplest case of
is the first-order polynomial metamodel, and more complex ones are intrinsically linear functions (ILF) that have been successfully applied for solving various design optimization problems [
25,
28,
29]. ILFs are nonlinear but they can be led to linear ones by simple transformations. Currently, five functions are considered in the regressor pool
:
In the second stage, for each function (
or
), different surrogates are assembled into one metamodel:
where
is the number of surrogates applied in the model bank
and
is the regression coefficient corresponding to each surrogate
, which reflects the quality of the individual
on the set of validation points. Similar to Equation (
3),
can be determined in the same manner:
It should be noted that, in the process of metamodel building, the design of experiments (DOE) is fixed, i.e., remains unchanged across the aforementioned stages.
Figure 1 illustrates the main steps in MAM. Note that, once the metamodels for the objective and constraint functions have been built, the constrained optimization subproblem formulated in the trust region (Equation (
2)) could be solved by any existing optimizers. In this paper, the sequential quadratic programming (SQP) method [
31] is applied to solve the constrained optimization subproblem for the optimal solution. Since numerical optimization solvers such as SQP are deterministic, the quality of the obtained solution is highly sensitive to the initial point. In other words, MAM could not perform the global search very well. To address this issue, the ASIM framework in
Section 4 has been proposed to integrate the stochastic nature with the exploratory search ability of PSO for the global optimal solution.
3. Brief Review of Particle Swarm Optimization (PSO)
Particle swarm optimization (PSO), inspired from swarm behaviors in nature such as fish and bird schooling, was developed by Kennedy and Eberhart [
32]. Since then, PSO has attracted a lot of attention and been developed as a main representative form of swarm intelligence. PSO has been applied to many areas, such as image and video analysis applications, engineering designs and scheduling applications, classification and data mining, etc. [
33]. There are at least twenty PSO variants as well as hybrid algorithms obtained by combining PSO with other existing algorithms, which are also becoming increasingly popular [
34,
35,
36].
To integrate PSO with MAM to find the global optimum, adaptive multi-level search is proposed in this paper. PSO is employed for the global-level exploration in the first step. A number of particles are first placed in the search space of the optimization problem with initial positions and velocities. However, the particles can fly over the entire design space not only determined by the individual and collective knowledge of positions from the global-level search but also based on the “local” information of each particle. Here, the “local” information means the local-level exploitation in the second step. In the neighborhood of each particle, an adaptive metamodel is constructed using MAM in
Section 2, which replaces the original optimization problem by a sequence of mathematical approximations that use much simpler objective and constraint functions. Hence, the critical information about individual constraint functions is kept and this leads to the improved accuracy of metamodels. During the process of metamodel building, each particle is endowed with a horizon in the surrounding region and then is refined with the current individual position to boost the possibility of finding an optimal position. Eventually, the swarm as a whole, similar to a flock of birds collectively foraging for food while each bird brilliantly and directly finds the most tasty food within the limited horizon, has the ability to move toward to a global optimum.
Each particle in PSO represents a point in the design space of an optimization problem with an associated velocity vector. In each iteration of PSO, the velocity vector is updated by using a linear combination of three terms shown in Equation (
10). The first term called inertia or momentum reflects a memory of the previous flight direction and prevents the particle from changing directions thoroughly. The second term, called the cognitive component, describes the tendency of particles returning to the previously found best positions. The last term, called the social component, quantifies the group norm or standard that should be attained. In other words, each particle tends to move toward the position of the current global best
and the location of the individual best
while moving randomly [
33]. The aim is to find the global best among all the current best solutions until the objective no longer improves or a certain number of iterations are reached. The standard iteration procedure of PSO is formulated as follows:
where
is the parameter called inertial weight,
t is the current iteration number,
and
are parameters called acceleration coefficients, and
and
are two homogeneously distributed random vectors generated within the interval
, respectively. If the values of
,
, and
are properly chosen (
and
), it has been proven that PSO could converge to an optimum [
37].
Even if PSO has been used in a variety of industry applications, it should be noted that the standard PSO suffers the disadvantages of information loss in the penalty function and high computational cost, especially in solving constrained optimization problems. Therefore, the proposed ASIM framework in the following section takes the advantage of PSO in global searching and reduces the burden on computation by introducing the metamodel building technique, model management, and trust region strategy.