Next Article in Journal
Capturing Causality for Fault Diagnosis Based on Multi-Valued Alarm Series Using Transfer Entropy
Next Article in Special Issue
Particle Swarm Optimization and Uncertainty Assessment in Inverse Problems
Previous Article in Journal
Entropic Updating of Probabilities and Density Matrices
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Improved Chaotic Optimization Algorithm Applied to a DC Electrical Motor Modeling

1
Dipartimento di Ingegneria dell’Informazione, Università Politecnica delle Marche, 60131 Ancona, Italy
2
Master’s Program Systems and Control, Technische Universiteit Eindhoven, 5612 Eindhoven, The Netherlands
*
Author to whom correspondence should be addressed.
Entropy 2017, 19(12), 665; https://doi.org/10.3390/e19120665
Submission received: 1 October 2017 / Revised: 16 November 2017 / Accepted: 24 November 2017 / Published: 4 December 2017
(This article belongs to the Special Issue Probabilistic Methods for Inverse Problems)

Abstract

:
The chaos-based optimization algorithm (COA) is a method to optimize possibly nonlinear complex functions of several variables by chaos search. The main innovation behind the chaos-based optimization algorithm is to generate chaotic trajectories by means of nonlinear, discrete-time dynamical systems to explore the search space while looking for the global minimum of a complex criterion function. The aim of the present research is to investigate the numerical properties of the COA, both on complex optimization test-functions from the literature and on a real-world problem, to contribute to the understanding of its global-search features. In addition, the present research suggests a refinement of the original COA algorithm to improve its optimization performances. In particular, the real-world optimization problem tackled within the paper is the estimation of six electro-mechanical parameters of a model of a direct-current (DC) electrical motor. A large number of test results prove that the algorithm achieves an excellent numerical precision at a little expense in the computational complexity, which appears as extremely limited, compared to the complexity of other benchmark optimization algorithms, namely, the genetic algorithm and the simulated annealing algorithm.

1. Introduction

A chaotic systems is a complex system that shows sensitivity to initial conditions, such as an economy, a stock market, the earth’s weather system, the behavior of water boiling on a stove, migratory patterns of birds, the spread of vegetation across a continent, an astronomical or a social system. In such systems, any uncertainty—no matter how small—in the beginning will produce rapidly escalating errors in the prediction of the system’s future behavior. To make an accurate prediction of the long-term behavior of such systems, the initial conditions must be known in their entirety and to an infinite level of accuracy. Even very simple or small systems can exhibit very complex behaviors. Typical features of chaotic systems include [1]:
  • Nonlinearity: A linear system cannot be chaotic; therefore, only nonlinear systems may evolve in time forming chaotic trajectories.
  • Determinism: A chaotic system is deterministic, as it does not include any random or stochastic element. Given an initial state for the system, it evolves according to deterministic rules. Chaos theory is the study of nonlinear dynamics, in which seemingly random events are actually predictable from deterministic equations.
  • Sensitivity to initial conditions: Small changes in the initial state of a chaotic system may lead to radically different trajectories (this is commonly referred to as the ”butterfly effect”).
  • Irregularity: Hidden order including a large or infinite number of unstable periodic trajectories forms the infrastructure of irregular chaotic systems. Chaos refers to an apparent lack of order in a system that nevertheless obeys particular laws or rules.
  • Long-term unpredictability: Long-term prediction of a trajectory of a chaotic system is extremely hard due to its sensitivity to initial conditions, which can be known only up to a finite degree of precision.
Historically, the study of chaos started in mathematics and physics and expanded into engineering, as well as into information and social sciences. Information entropy is tightly related to chaoticity in real-world discrete-time sequences, as underlined, for example, in Hagmair et al. [2]. There exist different kinds of potential commercial and industrial applications of chaotic systems, which, for simplicity, are classified into stabilization, synthesis, and analysis [3]. A typical application of chaotic systems is found in the secure transmission of information, where a meaningful message is superimposed to a chaotic signal that makes it non-understandable to someone who neither knows the system that produces the chaotic trajectory nor its initial state. To decode the received signal, it is necessary to synchronize the chaotic oscillators at the transmitter and at the receiver [4].
In the present research study, chaotic systems are applied to constrained global optimization of possibly non-smooth, nonlinear complex functions of several variables. Although non-conventional optimization techniques based on stochastic methods have a long hystory (see, for example, [5]), optimization techniques based on chaos theory have been less investigated.
The present study took its moves from the seminal contribution [6] that proposed how to optimize complex functions by chaos search. The main idea behind the resulting chaos-based optimization algorithm (COA) [6] is to generate chaotic trajectories by means of nonlinear dynamical systems to explore the search space while looking for the global minimum of a complex criterion function. The main reasons to choose a chaotic search over a deterministic search may be summarized as follows:
  • The criterion function to optimize may be non-smooth and non-differentiable, hence, standard search methods such as gradient steepest descent may not be applied in its optimization.
  • A chaotic system may produces complex, non-repeating trajectories that hardly visit the same state more than once, hence a chaos-based search algorithm is able to visit the whole space efficiently while looking for the optimal configuration.
  • A multivariate criterion function may include a large number of variables to optimize simultaneously, therefore, a simple mechanism to generate multivariate candidate solutions over a high-dimensional search space makes the search algorithm fast and easy to implement.
Despite its success in the optimization of complex functions, to the best of our knowledge, the potentiality of the chaos-based optimization algorithm has only been marginally investigated in the literature [7,8]. The aim of the present research study is to investigate the numerical properties of the COA, both on complex optimization test-functions from the literature and on a real-world problem, in order to contribute to the understanding of its global-search features. Through a large number of tests and numerical evaluations, it was found that the algorithm may be further fine-tuned in order to improve its numerical precision at a little expense in computational complexity, which is nevertheless extremely limited, compared to the complexity of other optimization algorithms, namely, the genetic algorithm and the simulated annealing algorithm.
The present paper is organized as follows. Section 2 summarizes the chaos-based optimization algorithm and tests a number of possible chaotic maps over a large number of test functions drawn from the specialized literature. Section 2 also suggests a fine-tuning of the original COA algorithm and compares the modified algorithm with the genetic algorithm and the simulated-annealing based optimization algorithm. Section 3 illustrates the performances of the refined COA algorithm in the modeling of a direct-current electrical motor. Section 4 concludes the paper.

2. The Chaos-Based Optimization Algorithm and Its Fine-Tuning

Traditional algorithms are unable to solve some optimization problems since these algorithms can get trapped into local minima or need too much search time. The properties of chaotic systems allow the development of an efficient chaos-based optimization algorithm. Unlike traditional stochastic optimization algorithms that escape from local minima by accepting some sub-optimal solutions according to a certain probability, the COA’s search takes advantage of chaotic motions’ regularity. A chaotic movement can go through every state in a definite area thanks to its own regularity and every state is obtained only once.
This Section illustrates the chaos-based optimization algorithm in Section 2.1 and describes in Section 2.2 several chaotic maps, which are indispensable for the development and the implementation of the COA; it also shows in Section 2.3 the performances of the COA in the optimization of a large set of benchmark functions. The results obtained from the large set of test experiments suggest a refinement of the original chaos-based optimization method, which is explained and numerically tested in Section 2.4. Section 2.5 presents a comparison between the refined COA and both the genetic optimization algorithm and the simulated-annealing-based optimization algorithm.

2.1. The Chaos-Based Optimization Algorithm

If a given optimization problem is described by a criterion f : R n R and the constraints that it is subjected to prescribe its variables x i to belong to specified intervals, the optimization problem can be described as min f ( x i ) , i = 1 , , n , subjected to a i x i b i , with a i , b i R . In the COA, thanks to the carrier wave method, the optimization variables follow the trajectories of specific chaotic functions whose ergodic areas are sized according to the constraints of the problem. The COA is outlined as follows:
  • Generate n chaotic variables x i , k from the same nonlinear chaotic oscillator, where the index i denotes the variable’s index and k denotes a discrete-time index. There will be n initial states x i , 0 that are chosen very close one to another (for instance 0.1 , 0.1001 , 0.1002 ). These initial states, because of chaotic systems properties, will give rise to completely different trajectories.
  • Launch the first carrier wave: This method allows us to obtain the optimization variables with the amplification of the ergodic areas of the n chaotic variables by the equation x i , k = c i + d i x i , k , where c i and d i denote amplification constants. Perform the rough search: Let x ¯ i = x i , 0 , calculate the value of the objective function f at the initial state and save its value f ¯ = f ( x i , 0 ) . Then start to iterate over the time k: Increase the time-index (iteration counter) k, generate a new attempted solution x i , k , calculate the value of the objective function f = f ( x i , k ) and evaluate the solution: If f f ¯ then set f ¯ = f and x ¯ i = x i , k . Continue with the iteration until f ¯ does not improve in K 1 search steps, where the integer constant K 1 determines the complexity of the search method. The first carrier wave will produce a rough solution ( x ¯ 1 , x ¯ 2 , , x ¯ n ) .
  • Launch the second carrier wave: The second carrier wave explores the search space in a narrower interval around the rough solution, in fact, the trajectories of the optimization variables are generated according to the rule x i , k = x ¯ i + e i + f i x i , k , where the x i , k denote again the states of n discrete-time chaotic systems, while e i and f i are the new amplification constants to perform the search within small ergodic areas around x ¯ i . Perform the fine search: Start to iterate over the time k: Increase the time-index k, generate a new attempted solution x i , k , calculate the value of the objective function f = f ( x i , k ) and evaluate the solution: If f f ¯ then set f ¯ = f and x ¯ i = x i , k . Continue with the iteration until f ¯ does not improve in K 2 search iterations.
  • The process ends and the values of parameters ( x ¯ 1 , x ¯ 2 , , x ¯ n ) , together with the criterion value f ¯ , will be the best approximation of the global solution of the given optimization problem as found by the chaos-based optimization algorithm.

2.2. Chaotic Maps

To generate the carrier waves, the COA makes use of a chaotic map that controls the features of the chaotic system used to generate the variables that are essential for the optimization.
There are some known universal properties of the chaotic maps, for instance:
  • They may present different kinds of attractors: An attractor is a set of numerical values towards which a dynamical system evolves after a sufficiently long time. A set of numerical values can be defined as an attractor if the trajectories that are related to them stay close to each other even if they are slightly perturbed.
  • They may present different values of the Lyapunov coefficient: This coefficient measures how much the system’s orbits are dependent from the initial state. Specifically, the Lyapunov coefficient measures the average parting speed of two orbits that are initially close and then drift apart.
Four chaotic maps have been surveyed. The characteristics that are more interesting for the optimization problem are the state’s trajectory variation (assuming that the initial states x i , 0 are very close to each other) and the ergodic area of each chaotic system, which is measured by pooling numerically the distribution of its states.
  • Logistic map. The “logistic map” is a second order polynomial function, described by:
    x k + 1 = r x k ( 1 x k ) , k = 0 , 1 , 2 , 3 , ,
    where r is the control parameter. It is supposed that 0 x 0 1 and that 0 r 4 . The Equation (1) represents a discrete-time, deterministic dynamical system without any stochastic interference. The long-term behavior of said system cannot be predicted because it changes completely as the control parameter r varies. In particular, when r = 4 , the system becomes chaotic. As we can see in Figure 1, from k > 10 , two waves originating from very close initial states start to behave in a different way and the system’s initial information is completely lost. It may be observed that the values generated by the logistic map are condensed around its ergodic area’s bounds, which are 0 and 1.
  • Quadratic map. The “quadratic map” arose as the real-valued version of the Mandelbrot set complex map and is described by:
    x k + 1 = x k 2 c , k = 0 , 1 , 2 , 3 , ,
    where the constant c is the control parameter. For c = 2 , the system acquires interesting complicated dynamics. The ergodic area is included between 2 and 2, as may be readily observed in Figure 2.
  • Tent map. The “tent map” is defined by the following dynamical system:
    x k + 1 = { μ x k if x k < 1 2 , μ ( 1 x k ) if x k 1 2 ,
    where μ is the control parameter and k = 0 , 1 , 2 , 3 , . For μ 1.5 , the waves start to show bifurcations. The value that has been used in the tests is μ = 1.9999 . Unlike the two previous maps, the tent map has an uniform distribution in its ergodic area [ 0 , 1 ] , as it may be readily appreciated in Figure 3.
  • Sine map. The “sine map” gives rise to the following discrete-time dynamical system:
    x k + 1 = λ sin ( π x k ) , k = 0 , 1 , 2 , 3 , ,
    where λ denotes the control parameter. This map generates a unique bifurcations diagram, symmetrical under both plane’s axis λ , x . The value that has been used in the tests is λ = 1 . Figure 4 shows the behavior of the sine map and that its ergodic area is again [ 0 , 1 ] .

2.3. Optimization of Test Functions

Seven benchmark functions have been taken into consideration to test and to further develop the COA. These functions are usually used in the literature [9] because they show a wide spectrum of characteristics, namely, they give rise to continuous/discontinuous, convex/nonconvex, unimodal/multimodal, quadratic/nonquadratic optimization problems.
  • F1: Generalized Rosenbrock function. This function of n variables is defined as
    f 1 ( x ) = i = 1 n 1 100 ( x i 2 x i + 1 ) 2 + ( 1 x i ) 2 .
    It was proposed by Rosenbrock in 1960 and is commonly used to test optimization algorithms since, for n = 2 , the search for the minimum is problematic in its parabolic area x 2 = x 1 2 ; furthermore, it becomes multimodal for n > 3 . The constraints and the global minimum of the generalized Rosenbrock function are recalled in the Table 1 and its graphical rendering is given in the Figure 5.
  • F2: Parabolic function. This function is defined as
    f 2 ( x ) = i = 1 n x i 2 .
    It was introduced to test optimization by genetic adaptive systems. The global minimum is not difficult to find because it is the only minimum of the function, but it can provide some information about the precision of the algorithm. The constraints that this optimization problem is subjected to are 5.12 x i 5.12 .
  • F3: Goldstein–Price function. This function is defined as
    f 3 ( x ) = ( 1 + ( x 1 + x 2 + 1 ) 2 ( 19 14 x 1 + 3 x 1 2 14 x 2 + 6 x 1 x 2 + 3 x 2 2 ) ) × ( ( 30 + ( 2 x 1 3 x 2 ) 2 ) ( 18 32 x 1 + 12 x 1 2 + 48 x 2 36 x 1 x 2 + 27 x 2 2 ) ) .
    The Goldstein–Price Function only has two variables but it exhibits several local minima. The constraints and the global minimum of this function within the considered domain are recalled in the Table 2, while its graphical rendering is given in the Figure 6.
  • F4: Schaffer function. This function is defined as
    f 4 ( x ) = 0.5 sin 2 x 1 2 + x 2 2 0.5 1 + 0.001 ( x 1 2 + x 2 2 ) 2 .
    The Schaffer function (number 4) is quite interesting because its global minimum is not unique, but it is given by all the points situated on a circumference with a well-defined radius r * . The constraints and the global minimum (in terms of radius) of this function are recalled in the Table 3 and its graphical rendering is given in the Figure 7.
  • F5: Step function. This function is defined as
    f 5 ( x ) = i = 1 n x i ,
    where the symbol · denotes rounding down to the nearest integer. This function tests an optimization algorithm’s ability to overcome discontinuities. The constraints and the global minimum of this function are recalled in Table 4 and its graphical rendering is given in Figure 8. Since there is not any unique global minimum, an optimization algorithm is supposed to stop when x i 5 .
  • F6: Normalized Schwefel function. This function is defined as
    f 6 ( x ) = 1 n i = 1 n x i sin | x i | .
    The Schwefel Function has its global minimum far away from the best local minimum: For this reason, optimization algorithms could converge on a wrong direction. The constraints and the global minimum of this function are recalled in the Table 5 and its graphical rendering is given in Figure 9.
  • F7: Rastring function. This function of n independent variables is defined as
    f 7 ( x ) = 10 n + i = 1 n x i 2 10 cos ( 2 π x i ) .
    The Rastring function is an example of a nonlinear, multimodal function. Its optimization is considered as a complicated problem because of its extended search space and its several local minima. The constraints and the global minimum of this function are recalled in Table 6 and its graphical rendering is given in the Figure 10.
The COA algorithm has been tested over the above seven benchmark functions, initially with only the first carrier wave and subsequently with the full algorithm. All the recalled four types of chaotic maps have been used to generate the optimization variables trajectories. The parameters of the algorithm are recalled below for the convenience of the reader:
  • K 1 : Stop criterion for the “rough search”.
  • K 2 : Stop criterion for the “fine search”.
  • c i , d i : Amplifiers for the first carrier wave.
  • e i , f i : Amplifiers for the second carrier wave.
We recall that the amplifiers c i , d i are constants that allow the chaotic maps’ ergodic areas to adapt to the specific search intervals (constraints) of the test function s. In the following tables, the figure Total Loops stands for the number of iterations taken to run and Error is the percentage distance (compared to the search interval) from the coordinates of the global minimum. The rough search has been run several times for all the functions and for all the chaotic maps with K 1 = 100 , 200 , 400 , 600 , 800 , 1000 . Only the results that correspond to an error under 10% and a limited number of Total Loops have been reported.
Table 7 illustrates the results obtained by applying a rough search to the benchmark functions F1 and F2 with n = 2 variables.
Only the quadratic map achieves a reasonable value but with too many iterations. With regard to the function F2, the algorithm, during the first 100 loops, could not find any better results than the initial states (that have been set at x 1 , 0 = 0.1 and x 2 , 0 = 0.1001 ).
Table 8 illustrates the results obtained by applying a rough search to the benchmark functions F3 and F4 with n = 2 variables. F3 function’s error is overall low, the logistic map is the one that exhibits better results. The optimization of the F4 function could be resolved by an infinite number of points that belongs to the circumference of radius r * = 1.5692 . Even in this case, the quadratic map needs the highest number of iterations.
The algorithm could easily overcome the discontinuity of the F5 function and it could find immediately values lower than 5 , therefore it is not necessary to show the relative table.
Table 9 illustrates the results obtained by applying a rough search to the benchmark functions F6 and F7 with n = 2 variables. The results about the optimization of the functions F6 and F7 do not provide any further information about the algorithm’s behavior but they confirm its excellent optimization ability.
In summary, the logistic map is the chaotic map that allows the first carrier wave to obtain better outcomes, namely, lower number of Total Loops, stop criterion to about 200 iterations and a final value close to any global minimum (in the worst case, with an Error of 8%). Moreover, the tent map provide good results. The quadratic map proves to be the worst in almost all tests (the cause may be its ergodic area included between 2 and 2). As it can be seen, the rough search does not need a high value of K 1 ; in fact, since the initial search area is rather wide, increasing the number of Total Loops does not always imply a proportional Error’s reduction.
It is now worth recalling that the amplifiers e i , f i of the second carrier wave are the algorithm’s most important parameters because they influence substantially its precision. These parameters are problem-dependent: specific constants were used, that could modify the ergodic area of the chaotic waves. In some cases, this area was changed to be 3% of the total search area, in others it was change to 2% or 8%. Thanks to its good performance during the “rough search”, the logistic map was used to guide the second carrier wave. The amplifiers were calculated in the following way:
e i = ε | α i | , f i = 2 ε | β i α i | ,
where α i and β i are, respectively, the lower bound and the upper bound of the constraints that were associated to the variable x i , ε is the area’s shrinkage percentage (for example, 2% or 3% or 8%).
Several tests were effected by changing at every test the stopping parameter K 2 . The obtained results are displayed in Table 10. In the table, x ¯ 1 and x ¯ 2 denote the values found by the COA, whereas P is a precision index that shows the ratio between the errors’ average on the individual variables and a normalization constant proportional to the width of the search area, namely:
P = i = 1 n | x ¯ i x i * | i = 1 n | β i α i | .
In all experiments, n = 2 variables were chosen.
The results pertaining to some selected values of the parameters K 2 are included in the table to point out the precision that can be obtained with either a low or a high number of iterations. The substantial Error of the “rough search” affects the performance of the “fine search” applied to the F1 function, which is the most troublesome for both stopping criteria. Therefore, it is important to use, for the purpose of optimization, a limited neighborhood for the second search. The values of the performance index P lay between 10 5 and 10 4 in most cases, which proves the quality of the chaos-based optimization algorithm. It can be seen that almost for all functions, excluding F4 and F5, the values x ¯ 1 and x ¯ 2 are close to the coordinates of the global minimum but only after a relatively large number of iterations.

2.4. Fine-Tuning of the COA Method (R-COA)

Specific problems that have dozens of variables in their objective functions or dynamic systems that need the resolution of differential equations could see their computing time exponentially increased by an high number of iterations. In the present research, to avoid the need for a large number of iterations, it has been proposed the introduction of a third carrier wave, which corresponds to an additional search stage, hereafter denoted as refined search. The whole refined chaos-based optimization algorithm (R-COA) may be outlined as follows:
  • Generation of chaotic variables.
  • Launch the first carrier wave and perform the rough search (identical to that of the COA).
  • Launch the second carrier wave and perform the fine search (identical to that of the COA).
  • Launch the third carrier wave: It is described by the equation x i , k = x ¯ i + g i + h i x i , k , where x ¯ i is the current best solution as found by the second carrier wave, g i and h i are new amplification constants that allow the search in small ergodic areas around x ¯ i , and x i , k is a new chaotic wave. Perform the refined search: Start to iterate over the time k, increase the time-index (iteration counter) k, generate a new attempted solution x i , k , calculate the value of the objective function f = f ( x i , k ) and evaluate the solution: If f f ¯ , then set f ¯ = f and x ¯ i = x i , k . Continue with the iteration until f ¯ does not improve in K 3 search loops.
  • The process ends and the parameter values ( x ¯ 1 , x ¯ 2 , , x ¯ n ) together with the criterion value f ¯ will be the best global solution to the optimization problem, as found by the refined COA.
For the new analysis of the test functions, a lower value of the constant K 2 was used. In addition, different K 3 values were tried. The amplifiers g i , h i are defined as:
g i = δ | α i | , h i = 2 δ | β i α i | ,
where δ represents the shrinkage percentage of analyzed area (2.5%, 1.5% or 0.5%).
The obtained results are displayed in Table 11. In the table, x ¯ 1 and x ¯ 2 again denot the values found by the R-COA, whereas P is the precision index defined before. Comparing the results displayed in Table 11 with the results displayed in Table 10, it can be noticed that the COA endowed with a third carrier wave always achieves a better (or comparable) precision, albeit with a lower number of Total Loops, than its original version. The values x ¯ i found by the algorithm, excluded the ones pertaining to the F6 function (since it has a wide search interval), displays an error in the fourth or fifth decimal digit.
Table 12 displays the values f ¯ of the test function achieved by the refined COA: The differences between the value found by the optimization algorithm and the actual minimum are around 10 4 and 10 5 .

2.5. Comparison with the Genetic and the Simulated-Annealing Algorithms

The performances of the chaos-based optimization algorithm R-COA have been compared with those exhibited by Genetic Algorithm and Simulated Annealing algorithm.
Genetic algorithms (GAs) are computer programs that mimic the processes of biological evolution in order to solve problems and to model evolutionary systems (see, e.g., [10,11], for an overview). The continuing performance improvements of computational systems has made them attractive for some types of optimization. In particular, genetic algorithms work well on mixed continuous/discrete combinatorial problems. GAs are less susceptible to getting stuck at local optima than gradient search methods, although they tend to be computationally expensive. To run a genetic algorithm, one must represent a solution to a given problem as a genome (or chromosome) and the quality of an individual in relation to a given optimization problem is measured by a fitness function. The genetic algorithm then creates a population of solutions and applies genetic operators such as mutation and crossover to evolve the solutions in order to find the best ones. Although it is not guaranteed that a GA will find a global optimal solution, in general, a GA is able to find good solutions within an acceptable time span. Genetic algorithms present several parameters to set and operators to tune in order to optimize their optimization performances. In the present research, the GA solver of MATLAB’s Global Optimization Toolbox (version 2015a) was utilized maintaining without changes its already optimized options. See, e.g., [12], for a recent application to system identification and [13] for an application to magnetic resonance image segmentation.
Simulated annealing (SA) is a method for finding a good (albeit not necessarily optimal) solution to an optimization problem [14]. Simulated annealing’s strength is that it avoids getting caught at local optima, namely, at solutions that are better than any others nearby, and yet are not the global optima. In fact, generally speaking, an optimization algorithm searches for the best solution by generating a random initial solution and “exploring” the area nearby: If a neighboring solution is better than the current one, then it moves to it, otherwise the algorithm stays put. This basic mechanism can lead to situations where it gets stuck at a sub-optimal place. Simulated annealing injects some sort of randomness into the basic mechanism to escape local optima early in the process without getting off when an optimal solution is nearby. The simulated annealing algorithm used in the experiments is the one implemented within MATLAB (version 2015a). See, e.g., [15] for an application to indoor wireless local area network (WLAN) localization.
These algorithms were run 10 times on independent trials. Table 13 shows the results obtained by comparing the R-COA algorithm and the simulated annealing algorithm, while Table 14 shows the results obtained by comparing the COA algorithm and the genetic algorithm. In particular, for the SA and the GA, the tables show the average values x i over the independent trials and the standard deviation of the found global minimum’s coordinates, the average number of iterations required to converge and the number of times that the SA or the GA algorithm achieved convergence.
When tested over the functions F1–F5, the SA always converges towards the global minimum and obtains the exact solution for the benchmark functions F2, F3 and F5 (as it can be seen from the values of the standard deviation). For the functions F1, F4 and F7, the SA is less accurate than the R-COA, moreover, the average number of iterations for five functions out of seven is higher than the number of total loops required by the R-COA. The search for the test function F6 did not converge (the algorithm probably was trapped in a local minima due to the large search’s area).
GA’s results are worse compared with those of SA, even if, generally, the number of iterations is lower. The GA algorithm, not only does not always converge when tested with the benchmark functions F6 and F7, but for seven times out of ten, with F3, it achieves values that lay far apart from the global minimum. GA and R-COA exhibit the same accuracy for the problem F4. The number of Total Loops is higher than the average number of iterations only in two cases.
In conclusion, the chaos optimization algorithm exhibits two important properties that distinguish it from the optimization algorithms commonly invoked in the scientific literature:
  • Convergence: The chaos-based optimization algorithm always converges towards the global minimum since: (a) it does not use variables related to probability; and (b) it is able to examine the whole search area, thanks to its mapping action;
  • Low computation time: Since it does not utilize gradient’s calculation to evaluate the tendency of the objective function, it exhibits a computational complexity equal to Θ ( n ) (the notation Θ ( f ( n ) ) (commonly referred to as “big-Theta of f ( n ) ”) is used to define the computational complexity of a problem: It expresses how fast an algorithm temporal cost increases as a function of The problem size n).

2.6. Generic Refined Chaos-Based Optimization Algorithm

The chaos-based algorithm, similar to other optimization algorithms, has parameters (ergodic areas’ amplifiers and stopping criteria) that must be adapted according to the faced problem. It was possible to model these parameters in order to create a generic algorithm, thanks to the large amount of data that were collected during the above tests.
The Euclidean distance between the minimum approximately found at the end of the rough search and the actual global minimum could be considered as a circle’s radius. This radius represents the sub-area where the second carrier wave will perform a second search. The same measure can be used for the fine search and the third carrier wave. Some general values were found while comparing all the sub-areas (regarded as a percentage of the total search area) and their relative stopping criteria. These values allow solving a generic optimization problem, even if it will perform less efficiently than a R-COA tailored to a specific problem.
The suggested amplifiers for the second carrier wave are:
e i = 0.025 d i 2 π , f i = 2 | e i | ,
and the suggested amplifiers for the third carrier wave are:
g i = 0.00015 d i 2 π , h i = 2 | g i | ,
where the d i ’s are the amplifiers for the first carrier wave. The stopping criteria are:
K 1 = 400 , K 2 = 500 .
Table 15 shows the outcomes of the optimization performed by the generic R-COA versus the results obtained with a specific R-COA. The global minimum’s coordinates are, in general, less accurate compared with those of specific R-COA (except for the first function). Moreover, the number of Total Loops is higher. However, since the error stays confined to the third or fourth decimal digit, the results can be considered acceptable.

3. Application to the Modeling of a Direct-Current Electrical Motor

The refined chaos-based optimization algorithm was applied to a real-world physical problem, namely, the estimation of the electro-mechanical parameters of a model of a direct-current (DC) electrical motor.
Direct-current motors found wide applications in industrial systems because they are easy to control and to model. For analytic control system design and optimization, a precise model of a DC motor may be desirable, since the specifications provided by the motor manufacturer may not be considered adequate, especially for cheaper DC motors which tend to exhibit relatively large tolerances in their electrical and mechanical parameters [16].
The present section recalls the fundamental equations describing the electro-mechanical dynamics of a direct-current electrical motor in Section 3.1. Section 3.2 casts the identification of the parameters of a DC motor model as an optimization problem. Section 3.3 describes and comments the results about the motor’s parameters identification when the motor is driven by a step signal.

3.1. Direct-Current Electrical Motors

Figure 11 shows an electro-mechanical model of a DC electrical motor. It is assumed that the stator has only a pair of polar expansions, characterized by an inductance L e associated to its relative winding and by a resistance R e associated to conductor’s leakage. The analytic equation that describes the stator-equivalent electrical circuit is:
v e ( t ) = L e d i e ( t ) d t + R e i e ( t ) ,
where i e denotes the stator current intensity. Likewise, it is assumed that the rotor has only a pair of polar expansions, characterized by an inductance L a and by a resistance R a . Furthermore, the model takes into account the electromotive force’s effect, e, that correspond to an induced voltage proportional to the rotation speed. The following equation describes the rotor’s electrical circuit:
v a ( t ) = L a d i a ( t ) d t + R a i a ( t ) + e ( t ) ,
where i a denotes the rotor current intensity. According to the physical properties of the motor, while the motor rotates with angular velocity ω m , a proportional electromotive force is generated in the armature circuit, whose intensity is given by
e ( t ) = K ω m ( t ) ,
where K is a proportionality constant whose value depends on the physical design of the DC motor.
Whenever an electrical current of intensity i a flows through the rotor winding, magnetic attraction/repulsion forces appear between the stator’s and rotor’s windings, which cause the motor shaft to rotate. The magnetic field inside the motor transfers a mechanical torque to the drive shaft. If we denote the rotor inertia by J, the viscous friction constant by B m and the torque load applied to the drive shaft by T L (that is supposed to be constant), the equation relating the torque T m available at the drive shaft may be written as
T m ( t ) = J d ω m ( t ) d t + B m ω m ( t ) + T L .

3.2. Formulation of the Motor’s Parameters Identification as an Optimization Problem

To estimate, through an optimization algorithm, the values of the parameters of the above model of a DC motor, it is necessary to know its output values corresponding to a given input drive signal [17]. A DC motor was simulated by a Simulink model, assuming constant stator voltage, controlled by the armature voltage. Figure 12 shows the Simulink model block-diagram.
The armature voltage v a is the only reference signal, whereas the constant load torque T L is regarded as a system disturbance. In the Simulink model, they are multiplied by V r and T r , respectively, to amplify their values. In the two sections of the diagram that represent the electrical dynamics and the mechanical dynamics, there is an integrator block with initial conditions i a 0 and ω m 0 . The outputs of the model are the armature current i a and the angular speed ω m of the drive shaft.
The state-space representation of the model is
d i a ( t ) d t = v a ( t ) R a i a ( t ) K ω m ( t ) L a , d ω m ( t ) d t = T m ( t ) B m w m ( t ) T L J .
The initial value of the armature current and of the angular speed of the drive shaft are i a 0 = 22.8284 A and ω m 0 = 122.3259 rad/s.
A step input response was analyzed and its corresponding outputs are shown in Figure 13.
To estimate the best values of the electro-mechanical parameters of a DC motor model, the problem is formulated in terms of model performance, to be optimized by means of the R-COA.
In [18], two performance criteria were tested, namely, a quadratic criterion and an absolute criterion. A comparison reported in the paper [18] reveals that the results obtained though the optimization of the quadratic criterion appear as better than those obtained by means of the absolute criterion. However, in the study [19], it was observed how these criteria perform nearly the same. In the present research, the following absolute criterion was selected as objective function:
C = j = 1 D ( k a | i a , j i a s , j | + k b | ω m , j ω m s , j | ) ,
where i a , j and ω m , j represent the values of the variables computed by means of the Simulink model, upon time discretization, while i a s , j and ω m s , j denote the values of the variables obtained by the optimization algorithm, upon time discretization, at discrete-time step j = 1 , 2 , , D . The constants k a > 0 and k b > 0 are weights, which attribute a different importance to the error in the estimate of the armature current and to the error in the estimate of the drive shaft angular speed, that will be discussed later.
Table 16 shows the actual values of the electro-mechanical parameters of the considered DC motor.
When launching the R-COA, it is necessary to take into account that each parameter is constrained to lie in a given interval, summarized in Table 17. Therefore, the R-COA must adapt the ergodic areas of six chaotic waves to these constraints.

3.3. Results for an Electrical Motor Driven by a Step Armature Voltage

In the first set of experiments, a reference step signal of 240 V of amplitude was applied as armature voltage. In this test condition, the velocity of the drive shaft ω m rises until it reaches the steady-state value of 124.9103 rad/s, while the armature current intensity decreases until it stabilizes at a value of 14.8600 A. The transient lasts approximately 0.5 second.

3.3.1. Separated Electrical and Mechanical Model Parameters Estimation

As a first test, only the electrical dynamics of the DC motor was considered as unknown, whereas the mechanical parameters were considered as exactly known. Namely, the armature resistance R a and armature inductance L a were taken as the only parameters to optimize for. In this test, the exact values of the torque load T L and of the torque coefficient K were made use of. In Table 18, the Error represents the difference between the value of the physical characteristic found by the algorithm and the value of the actual one, whereas the Total Error is the value of the objective function C in (23).
As a second test, only the mechanical dynamics of the DC motor was considered as unknown, whereas the electrical parameters were considered as exactly known. Namely, in the second case, the optimization variables are the viscous friction B m and the rotor inertia J. In this test, the exact values of the torque load T L and of the torque coefficient K were again made use of. In Table 19, the Error represents the difference between the value of the physical characteristic found by the algorithm and the value of the actual one, whereas the Total Error is the value of the objective function C in (23).
As can be readily seen in the above tables, the COA can obtain values that are very close to the actual ones, even without a refined search. The number of iterations needed to achieve such results appears as low.

3.3.2. Joint Electrical and Mechanical Model Parameters Estimation and Weights Analysis

In the definition of the absolute criterion C (23), the weights k a and k b are assigned to the armature current mismatch and to the drive shaft velocity mismatch, respectively. In the third test, different values for such weights were considered in order to achieve a joint electrical and mechanical model parameters estimation. In this test, the torque load T L was assumed equal to zero.
In Table 20, the parameters’ values are shown as obtained with both a low and an high number of iterations. In the latter case, it was evaluated the way that the quality of results improves while increasing the computation time. In the first two cases, even if slightly different results were obtained with few loops (521), it can be seen that both runs converge to the same values by increasing the iterations number. In the third case, the estimations are worse compared to the first two. Therefore, the values k a = 1 and k b = 1 will be used for the subsequent tests.
As a side note, one might consider that the weights k a and k b may be regarded as constants that represent the measurements’ quality. For instance, if it is known in advance that the samples of the drive shaft’s velocity are affected by a measurement error, the Total Error could be reduced by modifying the constant k b . That is to say, if it is known that a sensor is not particularly sophisticated or accurate, the weight related to the corresponding physical quantity could be adjusted. As a numerical example, Table 21 shows results obtained by adding a Gaussian noise drawn from a normal distribution to the drive shaft velocity and by adjusting the coefficient k b . The Total Error obtained by setting k b = 1 increases drastically due to the noises added to measurements. By decreasing the weight k b by 70 % (namely, to k b = 0.3 ), the Total Error is reduced.

3.3.3. Full DC Motor Modeling

After the preliminary tests on modeling separated electrical/mechanical dynamics and after the evaluation of the criterion’s weights, the R-COA was used to obtain a solution to the full DC motor model’s parameter estimation. Several ergodic areas amplifiers were tested. Table 22 illustrates the best solutions found with a low and an high number of iterations.
In the table, Total Error is defined as the sum of the differences between samples from the actual motors outputs and the outputs corresponding to the inferred parameters values. The Parameters Error is a performance index that allows one to evaluate the quality of the found parameters values and is defined as a weighted sum, namely
Param . Error = | R a R a s | + | B m B m s | + 10 | J J s | + 100 | L a L a s | + 1 10 | T L T L s | .
Note that the weights of each term in the above sum were adjusted to balance the different values-ranges of the parameters.
The R-COA always converges towards the correct set of parameters, and the results are satisfactory. Comparing the obtained values of Total Error to the Parameters Error, two observations arise:
  • These errors do not decrease proportionally: From the second and third carrier waves, the Parameters Error drops by 62.3 % and by 26.4 %, while the Total Error drops by 31.4 % and by 58.9 %.
  • An improvement in the matching of the model responses to the actual responses does not always imply an improvement towards the correct parameters values. This may be due to the fact that there exists an infinite number of combinations of the parameters that correspond to the same model. In fact, from the state-space representation (22), it can be noticed that the model depends on a ratio between variables, namely, L a divides the electrical parameters while J divides the mechanical parameters: For this reason, infinite potential combinations are possible.
In Figure 14, a comparison between the response of the motor model, in terms of the armature current and of the velocity of the drive shaft, is shown. It is readily observed how the the steady-state responses, also reported in Table 23 for the convenience of the reader, look identical. The small differences of the two motors’ transient compose the Total Error.

3.3.4. Expansion of the Search Areas

To challenge the R-COA by increasing the problem’s difficulty, additional tests were conducted on expanded search areas of the electric motor’s physical characteristics compared to the ranges indicated in Table 17. Namely, in order to test the behavior of the R-COA when the search areas are badly estimated, the actual ranges were doubled and quadrupled, making them as indicated in Table 24. The R-COA always converges in spite of the enlarged search areas. The refined chaos-based optimization algorithm could find solutions that lay slightly apart from the actual ones and yet that minimize the absolute criterion. Table 25 presents the results of parameter estimation obtained after doubling the search areas, while Table 26 presents the results of parameter estimation obtained after quadrupling the search areas. Both Total and Parameters Error raise in the first and second carrier waves, while they decrease in the third one. If the number of iterations is quadrupled, the curves of the motor’s response get close to one another although the parameters’ precision does not improve. It can be observed that, even if the search areas increase, the number of iterations remains almost the same. In this case, the performance degrades, especially in the identification of the armature resistance R a , with a minimum error of 0.0762 on the actual value, and in the identification of the load torque T L . The Total Error is 6.7405 , while in the previous tests did not exceed 5.
The two panels of Figure 15 correspond to the values of the parameters reported in Table 25 and Table 26: The shown results seem to suggest that the more the search areas increases, the more the dynamics tend to differ, especially in the transient, while they always converge towards the same steady state.

3.4. Comparison with the Genetic Algorithm and the Simulated-Annealing Based Algorithm

As a last test, the solutions computed by means of the R-COA were compared to the solutions found by means of a Genetic Algorithm and a Simulated Annealing algorithm. Each optimization algorithm was run 10 times for each input type. Table 27 illustrates the average results. In particular, the bold-faced values are the best estimations for every parameter (e.g., for the Total loops, the parameter R a and the parameter L a ). In this way, it is also clear what algorithm performs the best on a particular parameter or figure.
The GA is always able to keep the number of total loops low, whereas the SA algorithm in a few cases gets over 3000 iterations. On the other hand, the R-COA exhibits considerably reduced computing time. Quantitatively, it took about 120 min to run the GA- and SA-based experiments, while it took only 10 min to run the R-COA-based experiments (all the computer codes were written in MATLAB and were run on a 2.40 GHz, 4 GB RAM, Intel-based PC).
Genetic algorithm and Simulated Annealing Algorithm rarely converge towards the correct solution; in fact, R-COA’s Total Error is equal to the 5 % of the other two algorithms’ error. These two algorithms, due to the parameters’ imprecision, show a mean-square error, for every measurement, equal to 3.7808 for the GA and equal to 4.4943 for the SA.
Figure 16 illustrates the motor responses modeled by the three optimization algorithms.
In general, COA and R-COA manage, on the one hand, to map the whole constrained area (maintaining a low computational complexity thanks to algorithm design) and, on the other hand, to explore gradually smaller sections by the chaotic wave mechanism that allow to refine the solution without getting lost in parameter space and in local minima. This can be an intuitive explanation of why it outperforms both GA and SA in the benchmark function analysis and in the motor modeling problem. For these reasons, chaotic methods have successfully been exploited in other areas, such as neural-network training, thanks to their excellent optimization abilities (see, for example, [20]).

4. Conclusions

The aim of the research endeavor summarized in the present paper was to investigate the COA, a global optimization algorithm based on the properties of chaotic waves.
The optimization features of the COA were first tested by means of a series of numerical experiments performed on seven benchmark functions, drawn from the scientific literature, that afforded getting an insight into the fundamental properties of this algorithm, which is based on the notion of two carrier waves. Such preliminary numerical tests also led to an interesting refinement that consisted in introducing a third carrier wave, which significantly improves the algorithm’s efficacy as it allows refining the solution at a little expense in terms of optimization loops.
The obtained R-COA algorithm was then applied to a real-world optimization problem, namely, a direct-current electric motor modeling. Several experiments were carried out, where the electrical and mechanical dynamics were analyzed separately or the search areas of the carrier waves were expanded. The same tests were performed using two different optimization algorithms: the genetic algorithm and the simulated annealing algorithm. The comparison with GA and SA highlights COA’s good features: convergence and low computing time.
It is important to emphasize that the effectiveness of COA and R-COA is highly related to the choice of its more important operator, the chaotic map. Further investigations could be carried out to find maps that would grant even better performances. Moreover, for future research, it could be interesting to carry out a more extensive study about both evolutionary strategies to set the ergodic areas’ amplifiers and about the extension of the modeling problem to control.

Acknowledgments

The authors wish to gratefully thank G. Ippoliti for the discussions on the modeling of a DC motor and for suggesting some interesting numerical tests. The authors also wish to express their gratitude to the Academic Editor, A. Mohammad-Djafari, and to the anonymous reviewers for providing a quick and accurate review turnaround.

Author Contributions

The authors have contributed equally in this research and in the writing of this paper. Simone Fiori and Ruben Di Filippo conceived and designed the experiments; Ruben Di Filippo prepared the computer codes and performed the numerical experiments; Simone Fiori and Ruben Di Filippo analyzed the numerical results; Simone Fiori and Ruben Di Filippo wrote the paper. All authors have read and approved the final version of the manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ditto, W.; Munakata, T. Principles and applications of chaotic systems. Commun. ACM 1995, 38, 96–102. [Google Scholar] [CrossRef]
  2. Hagmair, S.; Bachler, M.; Braunisch, M.C.; Lorenz, G.; Schmaderer, C.; Hasenau, A.-L.; von Stülpnagel, L.; Bauer, A.; Rizas, K.D.; Wassertheurer, S.; et al. Challenging recently published parameter sets for entropy measures in risk prediction for end-stage renal disease patients. Entropy 2017, 19, 582. [Google Scholar] [CrossRef]
  3. Ott, E.; Sauer, T.; Yorke, J.A. Coping with Chaos—Analysis of Chaotic Data and Exploitation of Chaotic Systems. In Wiley Series in Nonlinear Science, 1st ed.; John Wiley: New York, NY, USA, 1994. [Google Scholar]
  4. Pecora, L.M.; Carroll, T.L. Synchronization in chaotic systems. Phys. Rev. Lett. 1990, 64, 821. [Google Scholar] [CrossRef] [PubMed]
  5. Rubinstein, R.Y.; Kroese, D.P. The Cross-Entropy Method—A Unified Approach to Combinatorial Optimization, Monte-Carlo Simulation and Machine Learning; Springer: New York, NY, USA, 2004. [Google Scholar]
  6. Li, B.; Jiang, W. Optimizing complex functions by chaos search. Cybern. Syst. Int. J. 1998, 29, 409–419. [Google Scholar]
  7. Yang, D.; Li, G.; Cheng, G. On the efficiency of chaos optimization algorithms for global optimization. Chaos Solitons Fractals 2007, 34, 1366–1375. [Google Scholar] [CrossRef]
  8. Yang, D.; Liu, Z.; Zhou, J. Chaos optimization algorithms based on chaotic maps with different probability distribution and search speed for global optimization. Commun. Nonlinear Sci. Numer. Simul. 2014, 19, 1229–1246. [Google Scholar]
  9. Brachem, C. Implementation and Test of a Simulated Annealing Algorithm in the Bayesian Analysis Toolkit (BAT). Bachelor’s Thesis, Georg-August-Universität Göttingen, Göttingen, Germany, 2009. [Google Scholar]
  10. Mitchell, M. Genetic algorithms: An overview. Complexity 1995, 1, 31–39. [Google Scholar] [CrossRef]
  11. Goldberg, D.E. Genetic Algorithms in Search, Optimization & Machine Learning; Addison-Wesley: Boston, MA, USA, 1989. [Google Scholar]
  12. Zhou, S.; Cao, J.; Chen, Y. Genetic algorithm-based identification of fractional-order systems. Entropy 2013, 15, 1624–1642. [Google Scholar] [CrossRef]
  13. Zang, W.; Zhang, W.; Zhang, W.; Liu, X. A kernel-based intuitionistic fuzzy C-means clustering using a DNA genetic algorithm for magnetic resonance image segmentation. Entropy 2017, 19, 578. [Google Scholar] [CrossRef]
  14. Kirkpatrick, S.; Gelatt, C.D.; Vecchi, M.P. Optimization by simulated annealing. Science 1983, 220, 671–680. [Google Scholar] [CrossRef] [PubMed]
  15. Zhou, M.; Qiu, F.; Tian, Z.; Wu, H.; Zhang, Q.; He, W. An information-based approach to precision analysis of indoor WLAN localization using location fingerprint. Entropy 2015, 17, 8031–8055. [Google Scholar] [CrossRef]
  16. Wu, W. DC motor parameter identification using speed step responses. Model. Simul. Eng. 2012, 2012, 30. [Google Scholar] [CrossRef]
  17. Alonge, F.; D’Ippolito, F.; Raimondi, F.M. Least squares and genetic algorithms for parameter identification of induction motors. Control Eng. Pract. 2001, 9, 647–657. [Google Scholar] [CrossRef]
  18. Razik, H.; Rezzoug, A. Identification of electrical parameters of an induction motor, a comparison between R.L.S. and genetic algorithm. In Proceedings of the International Conference on Electrical Machines (ICEM 2000), Espoo, Finland, 28–30 August 2000; pp. 1433–1437. [Google Scholar]
  19. Benaïdja, N. Softcomputing identification techniques of asynchronous machine parameters: Evolutionary strategy and chemotaxis algorithm. Turk. J. Electr. Eng. Comput. Sci. 2009, 17, 69–85. [Google Scholar]
  20. Wang, S.; Zhang, Y.; Ji, G.; Yang, J.; Wu, J.; Wei, L. Fruit classification by wavelet-entropy and feedforward neural network trained by fitness-scaled chaotic ABC and biogeography-based optimization. Entropy 2015, 17, 5711–5728. [Google Scholar] [CrossRef]
Figure 1. Logistic map. Left-hand panel: Comparison between system’s output with two different initial conditions. (The wave with initial value x 0 = 0.1 is represented by the red line and the other with initial value x 0 = 0.1001 by the blue line.) Right-hand panel: Occurrence histogram of the values of the states of the dynamical system (1).
Figure 1. Logistic map. Left-hand panel: Comparison between system’s output with two different initial conditions. (The wave with initial value x 0 = 0.1 is represented by the red line and the other with initial value x 0 = 0.1001 by the blue line.) Right-hand panel: Occurrence histogram of the values of the states of the dynamical system (1).
Entropy 19 00665 g001
Figure 2. Quadratic map. Left-hand panel: Comparison between system’s output with two different, albeit very close, initial conditions. Right-hand panel: Occurrence histogram of the values of the states of the dynamical system (2).
Figure 2. Quadratic map. Left-hand panel: Comparison between system’s output with two different, albeit very close, initial conditions. Right-hand panel: Occurrence histogram of the values of the states of the dynamical system (2).
Entropy 19 00665 g002
Figure 3. Tent map. Left-hand panel: Comparison between system’s output with two different, although very close, initial conditions. Right-hand panel: Occurrence histogram of the values of the states of the dynamical system (3).
Figure 3. Tent map. Left-hand panel: Comparison between system’s output with two different, although very close, initial conditions. Right-hand panel: Occurrence histogram of the values of the states of the dynamical system (3).
Entropy 19 00665 g003
Figure 4. Sine map. Left-hand panel: Comparison between system’s output with two different initial conditions, very close to one another. Right-hand panel: Occurrence histogram of the values of the states of the dynamical system (4).
Figure 4. Sine map. Left-hand panel: Comparison between system’s output with two different initial conditions, very close to one another. Right-hand panel: Occurrence histogram of the values of the states of the dynamical system (4).
Entropy 19 00665 g004
Figure 5. Rendering of the generalized Rosenbrock function (benchmark function “F1”) for the case of n = 2 independent variables.
Figure 5. Rendering of the generalized Rosenbrock function (benchmark function “F1”) for the case of n = 2 independent variables.
Entropy 19 00665 g005
Figure 6. Rendering of the Goldstein–Price function (benchmark function “F3”).
Figure 6. Rendering of the Goldstein–Price function (benchmark function “F3”).
Entropy 19 00665 g006
Figure 7. Rendering of the Schaffer function (benchmark function “F4”).
Figure 7. Rendering of the Schaffer function (benchmark function “F4”).
Entropy 19 00665 g007
Figure 8. Rendering of the “step” function (benchmark function “F5”) for the case of n = 2 independent variables.
Figure 8. Rendering of the “step” function (benchmark function “F5”) for the case of n = 2 independent variables.
Entropy 19 00665 g008
Figure 9. Rendering of the Schwefel function (benchmark function “F6”) for the case of n = 2 independent variables.
Figure 9. Rendering of the Schwefel function (benchmark function “F6”) for the case of n = 2 independent variables.
Entropy 19 00665 g009
Figure 10. Rendering of the Rastring function (benchmark function “F7”) for the case of n = 2 independent variables.
Figure 10. Rendering of the Rastring function (benchmark function “F7”) for the case of n = 2 independent variables.
Entropy 19 00665 g010
Figure 11. An electro-mechanical model of the direct-current electrical motor (we assumed no angular speed change after gear).
Figure 11. An electro-mechanical model of the direct-current electrical motor (we assumed no angular speed change after gear).
Entropy 19 00665 g011
Figure 12. A Simulink model of an electrical direct-current motor with constant stator voltage.
Figure 12. A Simulink model of an electrical direct-current motor with constant stator voltage.
Entropy 19 00665 g012
Figure 13. Example of output values of the model of a direct-current electrical motor driven by a step input voltage. Left-hand panel: Armature current i a versus time. Right-hand panel: Drive shaft angular velocity ω m versus time.
Figure 13. Example of output values of the model of a direct-current electrical motor driven by a step input voltage. Left-hand panel: Armature current i a versus time. Right-hand panel: Drive shaft angular velocity ω m versus time.
Entropy 19 00665 g013
Figure 14. Motor driven by a step armature voltage: Comparison of the output curves. Blue-solid line: Actual motor response. Red-crosshatched line: Inferred motor model response.
Figure 14. Motor driven by a step armature voltage: Comparison of the output curves. Blue-solid line: Actual motor response. Red-crosshatched line: Inferred motor model response.
Entropy 19 00665 g014
Figure 15. Motor driven by a step armature voltage: Comparison of responses corresponding to the values of the parameters reported in Table 25 and Table 26. Blue-solid lines: Actual response. Red-crosshatched line: Response corresponding to the values of the parameters inferred upon doubling the search areas. Black dot-dashed lines: Response corresponding to the values of the parameters inferred upon quadrupling the search areas.
Figure 15. Motor driven by a step armature voltage: Comparison of responses corresponding to the values of the parameters reported in Table 25 and Table 26. Blue-solid lines: Actual response. Red-crosshatched line: Response corresponding to the values of the parameters inferred upon doubling the search areas. Black dot-dashed lines: Response corresponding to the values of the parameters inferred upon quadrupling the search areas.
Entropy 19 00665 g015
Figure 16. Estimation of a DC motor parameters: A comparison of motor responses between models instantiated with parameters estimated by the R-COA, GA and SA algorithm. In the two panels, the solid (blue) lines denote the expected responses, the crosshatched (red) lines denote the results corresponding to the R-COA-based optimization, the dashed (black) lines denote the results pertaining to the GA-based optimization, while the dot-dashed (purple) lines correspond to the SA-based optimization.
Figure 16. Estimation of a DC motor parameters: A comparison of motor responses between models instantiated with parameters estimated by the R-COA, GA and SA algorithm. In the two panels, the solid (blue) lines denote the expected responses, the crosshatched (red) lines denote the results corresponding to the R-COA-based optimization, the dashed (black) lines denote the results pertaining to the GA-based optimization, while the dot-dashed (purple) lines correspond to the SA-based optimization.
Entropy 19 00665 g016
Table 1. Constraints and global minimum of the generalized Rosenbrock function (benchmark function “F1”) for the case of n = 2 independent variables.
Table 1. Constraints and global minimum of the generalized Rosenbrock function (benchmark function “F1”) for the case of n = 2 independent variables.
ConstraintsGlobal Minimum
2 x i 2 f 1 = 0 ;     x i = 1
Table 2. Constraints and global minimum of the Goldstein–Price function (benchmark function “F3”).
Table 2. Constraints and global minimum of the Goldstein–Price function (benchmark function “F3”).
ConstraintsGlobal Minimum
2 x i 2 f 3 = 1.0156 × 10 5 ;     x 1 = 1.7693 , x 2 = 2
Table 3. Constraints and global minimum of the Schaffer function (benchmark function “F4”).
Table 3. Constraints and global minimum of the Schaffer function (benchmark function “F4”).
ConstraintsGlobal Minimum
4 x i 4 f 4 = 0.0025 ;     r * = 1.5692
Table 4. Constraints and global minimum of the “step” function (benchmark function “F5”) for the case of n = 2 independent variables.
Table 4. Constraints and global minimum of the “step” function (benchmark function “F5”) for the case of n = 2 independent variables.
ConstraintsGlobal Minimum
5.1 x i 5.1 f 5 = 12 ;     x i 5
Table 5. Constraints and global minimum of the Schwefel function (benchmark function “F6”) for the case of n = 2 independent variables.
Table 5. Constraints and global minimum of the Schwefel function (benchmark function “F6”) for the case of n = 2 independent variables.
ConstraintsGlobal Minimum
512 x i 512 f 6 = 418.982887 ;     x i = 420.968746
Table 6. Constraints and global minimum of the Rastring function (benchmark function “F7”) for the case of n = 2 independent variables.
Table 6. Constraints and global minimum of the Rastring function (benchmark function “F7”) for the case of n = 2 independent variables.
ConstraintsGlobal Minimum
5 x i 5 f 7 = 0 ;     x i = 0
Table 7. Results of the application of a rough search (only first carrier wave) on the benchmark functions F1 and F2.
Table 7. Results of the application of a rough search (only first carrier wave) on the benchmark functions F1 and F2.
F1F2
Map K 1 Total LoopsErrorMap K 1 Total LoopsError
Logistic4005537.90%Logistic1001002.00%
Quadratic100032103.50%Quadratic1001002.00%
Sine1001089.80%Sine1001002.00%
Tent1001088.00%Tent1001002.00%
Table 8. Results of the application of a rough search (only first carrier wave) on the benchmark functions F3 and F4.
Table 8. Results of the application of a rough search (only first carrier wave) on the benchmark functions F3 and F4.
F3F4
Map K 1 Total LoopsErrorMap K 1 Total LoopsError
Logistic1001331.75%Logistic1001000.00%
Quadratic60014310.20 %Quadratic10035260.00 %
Sine4006321.60%Sine8001410.00%
Tent4008261.80%Tent1001140.00%
Table 9. Results of the application of a rough search (only first carrier wave) on the benchmark functions F6 and F7.
Table 9. Results of the application of a rough search (only first carrier wave) on the benchmark functions F6 and F7.
F6F7
Map K 1 Total LoopsErrorMap K 1 Total LoopsError
Logistic60010101.26%Logistic1001002.00%
Quadratic1001772.60 %Quadratic4005889.90 %
Sine60010672.00%Sine1001002.00%
Tent80015164.00%Tent1001161.80%
Table 10. Results of the application of a fine search (both first and second carrier waves) on the benchmark functions F1–F7.
Table 10. Results of the application of a fine search (both first and second carrier waves) on the benchmark functions F1–F7.
Benchmark FunctionsCoordinates of GlobalMinimum K 2 Total Loops x ¯ 1 x ¯ 2 P
F1 x i = 1 2001032 1.0268 1.0534 1.0547 × 10 2
200045000.99960.9976 3.7846 × 10 4
F2 x i = 0 10001568 0.0010 0.0094 5.4751 × 10 4
200078060.00060.0002 3.2104 × 10 5
F3 x 1 = 1.7693 1001228 1.7707 1.9999 1.2678 × 10 4
x 2 = 2.0000 20,00060,013 1.7963 2.0000 1.8475 × 10 24
F4 r * = 1.5692 2005380.73121.3886 1.6854 × 10 27
F5 x i 5 100137 5.0356 5.0939 -
F6 x i = 420.9687 1001239421.1416420.9193 1.1557 × 10 4
15,00034,855420.9656420.9927 1.3554 × 10 5
F7 x i = 0 5001068 0.0010 0.0094 5.9844 × 10 4
21007606 0.0005 0.0001 3.2575 × 10 5
Table 11. Results of the application of a refined search (first, second and third carrier waves) on the benchmark functions F1–F3, F6, F7 (Since the functions F4 and F5 did not need a better optimization, they were not tested with the new algorithm).
Table 11. Results of the application of a refined search (first, second and third carrier waves) on the benchmark functions F1–F3, F6, F7 (Since the functions F4 and F5 did not need a better optimization, they were not tested with the new algorithm).
Benchmark FunctionsCoordinates of Global Minimum K 3 Total Loops x ¯ 1 x ¯ 2 P
F1 x i = 1 100028120.99991.0008 1.0657 × 10 4
F2 x i = 0 3008510.00000.0001 5.7872 × 10 7
F3 x 1 = 1.7693 8002705 1.7691 1.9999 1.3514 × 10 5
x 2 = 2.0000
F6 x i = 420.9687 6001884420.9641420.9719 3.2198 × 10 6
F7 x i = 0 1004380.00020.0005 5.7785 × 10 5
Table 12. Values of the test functions resulting from the application of a refined search (first, second and third carrier waves) on the benchmark functions F1–F3, F6, F7. (The functions F4 and F5 were not tested).
Table 12. Values of the test functions resulting from the application of a refined search (first, second and third carrier waves) on the benchmark functions F1–F3, F6, F7. (The functions F4 and F5 were not tested).
Test FunctionsGlobal Minimum f ¯
F10 8.9131 × 10 5
F20 1.0032 × 10 6
F3 1.0156 × 10 5 1.0154 × 10 5
F6 418.9828 418.9829
F70 5.2471 × 10 5
Table 13. Results of a comparison between the refined COA and the SA (Simulated Annealing Algorithm) on the benchmark functions F1–F7.
Table 13. Results of a comparison between the refined COA and the SA (Simulated Annealing Algorithm) on the benchmark functions F1–F7.
FunctionRefined COASimulated Annealing
EstimatesTotal LoopsAverage EstimatesStandard DeviationAverage IterationsConvergence
F1 x ¯ 1 = 0.9999 2812 x 1 = 1.0329 0.1038140710
x ¯ 2 = 1.0008 x 2 = 1.0771 0.2240
F2 x ¯ 1 = 0.0000 851 x 1 = 0.0000 0.0000101110
x ¯ 2 = 0.0001 x 2 = 0.0000 0.0000
F3 x ¯ 1 = 1.7961 2705 x 1 = 1.7693 0.0000101110
x ¯ 2 = 1.9999 x 2 = 2.0000 0.0000
F4 r ¯ = 1.5692 538 r = 1.5688 0.0010163910
F5 x ¯ i 5 137 x i 5 0.0000101110
F6 x ¯ 1 = 420.9641 1884 x 1 = 109.1471 218.577724370
x ¯ 2 = 420.9719 x 2 = 7.9124 189.6553
F7 x ¯ 1 = 0.0002 438 x 1 = 0.0963 0.315921678
x ¯ 2 = 0.0005 x 2 = 0.0979 0.3152
Table 14. Results of a comparison between the refined COA and the GA (Genetic Algorithm) on the benchmark functions F1–F7.
Table 14. Results of a comparison between the refined COA and the GA (Genetic Algorithm) on the benchmark functions F1–F7.
FunctionRefined COAGenetic Algorithm
EstimatesTotal LoopsAverage EstimatesStandard DeviationAverage IterationsConvergence
F1 x ¯ 1 = 0.9999 2812 x 1 = 0.8822 0.0974139210
x ¯ 2 = 1.0008 x 2 = 0.7867 0.1745
F2 x ¯ 1 = 0.0000 851 x 1 = 0.0000 0.0000101110
x ¯ 2 = 0.0001 x 2 = 0.0000 0.0000
F3 x ¯ 1 = 1.7961 2705 x 1 = 0.6067 1.246411863
x ¯ 2 = 1.9999 x 2 = 0.4933 0.7942
F4 r ¯ = 1.5692 538 r = 1.5692 0.0000104010
F5 x ¯ i * 5 137 x i 5 0.0000101010
F6 x ¯ 1 = 420.9641 1884 x 1 = 204.4482 123.354713240
x ¯ 2 = 420.9719 x 2 = 150.4521 126.5412
F7 x ¯ 1 = 0.0002 438 x 1 = 0.0995 0.314610408
x ¯ 2 = 0.0005 x 2 = 0.1990 0.4195
Table 15. Results of a comparison between the specific refined COA and the generic refined COA on the benchmark functions F1–F7.
Table 15. Results of a comparison between the specific refined COA and the generic refined COA on the benchmark functions F1–F7.
FunctionsSpecific Refined COAGeneric Refined COA
EstimatesTotal Loops Estimates Total Loops
F1 x ¯ 1 = 0.9999 2812 x ¯ 1 = 1.0014 2055
x ¯ 2 = 1.0008 x ¯ 2 = 1.0026
F2 x ¯ 1 = 0.0000 851 x ¯ 1 = 0.0001 2055
x ¯ 2 = 0.0001 x ¯ 2 = 0.0009
F3 x ¯ 1 = 1.7961 2705 x ¯ 1 = 1.7695 5283
x ¯ 2 = 1.9999 x ¯ 2 = 1.9998
F4 r ¯ = 1.5692 538 r ¯ = 1.5694 2055
F6 x ¯ 1 = 420.9641 1884 x ¯ 1 = 421.0366 3662
x ¯ 2 = 420.9719 x ¯ 2 = 420.8288
F7 x ¯ 1 = 0.0002 438 x ¯ 1 = 0.0001 2105
x ¯ 2 = 0.0005 x ¯ 2 = 0.0009
Table 16. Application to DC motor model parameters estimation: Actual values of motor’s physical electro-mechanical parameters.
Table 16. Application to DC motor model parameters estimation: Actual values of motor’s physical electro-mechanical parameters.
ParameterParameter SymbolParameter Value
Armature Resistance R a 0.6 Ω
Armature Inductance L a 0.03 H
Viscous Friction B m 0.1 N·m·s/rad
Rotor Inertia J 0.6 kg·m 2 /rad
Torque Coefficient K 1.85 V·s/rad
Torque Load T L 15 N·m
Table 17. Application to DC motor model parameters estimation: Search range for each parameter based on physical constrains.
Table 17. Application to DC motor model parameters estimation: Search range for each parameter based on physical constrains.
ParameterParameter SymbolParameter Search Range
Armature Resistance R a 0.1–0.8 ( Ω )
Armature Inductance L a 0.01–0.05 (H)
Viscous Friction B m 0.05–0.5 (N·m·s/rad)
Rotor Inertia J 0.1–0.8 (kg·m 2 /rad)
Torque Coefficient K 1–2 (V·s/rad)
Torque Load T L 10–20 (N·m)
Table 18. Motor driven by a step armature voltage: Test on estimating the electrical model parameters only.
Table 18. Motor driven by a step armature voltage: Test on estimating the electrical model parameters only.
Carrier WaveTotal Loops R a Error L a ErrorTotal Error
First2020.61560.01560.03200.00201.1003
Second9530.60030.00030.03010.00010.0601
Table 19. Motor driven by a step armature voltage: Test on estimating the mechanical model parameters only.
Table 19. Motor driven by a step armature voltage: Test on estimating the mechanical model parameters only.
Carrier WaveTotal Loops B m ErrorJErrorTotal Error
First2720.09910.00090.60850.00851.0095
Second4320.09910.00090.60080.00080.0704
Table 20. Motor driven by a step armature voltage: Weights analysis for the joint electrical and mechanical model parameters estimation.
Table 20. Motor driven by a step armature voltage: Weights analysis for the joint electrical and mechanical model parameters estimation.
Total Loops R a L a B m JKTotal Error
Actual Value 0.60.030.10.61.85
k a = 1 , k b = 1 5210.59040.02950.09820.63881.85169.9568
52290.59420.03120.10010.65031.85057.3798
k a = 1 , k b = 0.5 5210.60290.02950.10050.65211.85047.9601
52290.59420.03120.10010.65031.85057.3798
k a = 1 , k b = 2 5820.59040.02950.09820.63991.851612.6972
71430.59260.03210.10010.65031.85058.9209
Table 21. Motor driven by a step armature voltage: Joint electrical and mechanical model parameters estimation in the presence of a disturbance on the drive shaft velocity measurements.
Table 21. Motor driven by a step armature voltage: Joint electrical and mechanical model parameters estimation in the presence of a disturbance on the drive shaft velocity measurements.
Total Loops R a L a B m JKTotal Error
Actual value 0.60.030.10.61.85
k b = 1 4110.80000.03440.07030.24241.8328139.2181
k b = 0.5 4110.80000.03440.07030.24241.8328139.2181
k b = 0.3 8460.79930.03570.08880.53661.8010114.2229
Table 22. Motor driven by a step armature voltage: Full motor modeling results.
Table 22. Motor driven by a step armature voltage: Full motor modeling results.
Loops R a L a B m JK T L ErrorParameter Error
Actual value 0.60.030.10.61.8515
First carrier16080.76930.02350.20530.76871.828513.529628.94365.1975
Second carrier17660.63460.03390.09540.52531.843215.975718.02431.6334
Third carrier19280.55940.02800.10040.57051.855115.52254.75740.9623
46240.62740.33700.09920.56451.846515.63794.36001.0743
Table 23. Motor driven by a step armature voltage: Actual steady-state values versus predicted values.
Table 23. Motor driven by a step armature voltage: Actual steady-state values versus predicted values.
Armature CurrentDrive Shaft Velocity
i a 148,600 ω m 1,249,102
i a s 148,379 ω m s 1,249,340
Table 24. Motor driven by a step armature voltage: Ranges for the parameters search areas doubled and quadrupled with respect to the values indicated in Table 17.
Table 24. Motor driven by a step armature voltage: Ranges for the parameters search areas doubled and quadrupled with respect to the values indicated in Table 17.
Constrains Expansion R a L a B m JK T L
Doubling0.1–1.50.01–0.090.05–0.950.1–1.51–310–20
Quadrupling0.1–2.90.01–0.170.05–1.850.1–2.91–510–50
Table 25. Motor driven by a step armature voltage: Results of parameters estimation by R-COA on doubled search areas.
Table 25. Motor driven by a step armature voltage: Results of parameters estimation by R-COA on doubled search areas.
Total Loops R a L a B m JK T L Total ErrorParameter Error
Actual value 0.60.030.10.61.8515
First carrier2710.70770.02910.22560.38881.917411.6350337.288210.7149
Second carrier4290.60410.03350.10510.64201.866815.9271104.44694.9715
Third carrier15310.58300.03060.10080.66451.855315.093012.66430.8976
90370.62600.02970.09960.65891.846916.19674.76911.0058
Table 26. Motor driven by a step armature voltage: Results of parameters estimation by R-COA on quadrupled search areas.
Table 26. Motor driven by a step armature voltage: Results of parameters estimation by R-COA on quadrupled search areas.
Total Loops R a L a B m JK T L Total ErrorParameter Error
Actual value 0.60.030.10.61.8515
First carrier1231.33660.14480.20940.81151.860916.45778375.597822.2117
Second carrier2331.04050.04730.09590.23301.799616.100872.06537.0065
Third carrier13490.53380.03120.09980.63711.853416.193724.62211.2778
25820.75160.04290.09730.59871.853415.47166.74051.8114
Table 27. Estimation of a DC motor parameters: Comparison of the R-COA with the SA algorithm and the GA.
Table 27. Estimation of a DC motor parameters: Comparison of the R-COA with the SA algorithm and the GA.
Results
AlgorithmR-COAGASA
Total Loops462415789767
R a 0.60.62740.40480.3873
L a 0.030.03370.01830.0139
B m 0.10.09920.05110.0500
J 0.60.56450.51470.4123
K 1.851.84651.85081.8475
T L 1515.637913.220314.8479
Total Error4.3600204.1683242.6949
Parameter Error1.07434.64036.1314

Share and Cite

MDPI and ACS Style

Fiori, S.; Di Filippo, R. An Improved Chaotic Optimization Algorithm Applied to a DC Electrical Motor Modeling. Entropy 2017, 19, 665. https://doi.org/10.3390/e19120665

AMA Style

Fiori S, Di Filippo R. An Improved Chaotic Optimization Algorithm Applied to a DC Electrical Motor Modeling. Entropy. 2017; 19(12):665. https://doi.org/10.3390/e19120665

Chicago/Turabian Style

Fiori, Simone, and Ruben Di Filippo. 2017. "An Improved Chaotic Optimization Algorithm Applied to a DC Electrical Motor Modeling" Entropy 19, no. 12: 665. https://doi.org/10.3390/e19120665

APA Style

Fiori, S., & Di Filippo, R. (2017). An Improved Chaotic Optimization Algorithm Applied to a DC Electrical Motor Modeling. Entropy, 19(12), 665. https://doi.org/10.3390/e19120665

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop