Next Article in Journal
Multivariate Multiscale Dispersion Entropy of Biomedical Times Series
Next Article in Special Issue
A Statistical Method for Estimating Activity Uncertainty Parameters to Improve Project Forecasting
Previous Article in Journal
Design and Analysis of the Domestic Micro-Cogeneration Potential for an ORC System Adapted to a Solar Domestic Hot Water System
Previous Article in Special Issue
A Neutrosophic Forecasting Model for Time Series Based on First-Order State and Information Entropy of High-Order Fluctuation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Evolved-Cooperative Correntropy-Based Extreme Learning Machine for Robust Prediction

1
Department of Instrument Science and Technology, University of Electronic Science and Technology of China, Chengdu 611731, China
2
Department of Applied Linguistics, University of Electronic Science and Technology of China, Chengdu 611731, China
*
Author to whom correspondence should be addressed.
Entropy 2019, 21(9), 912; https://doi.org/10.3390/e21090912
Submission received: 6 August 2019 / Revised: 4 September 2019 / Accepted: 12 September 2019 / Published: 19 September 2019
(This article belongs to the Special Issue Entropy Application for Forecasting)

Abstract

:
In recent years, the correntropy instead of the mean squared error has been widely taken as a powerful tool for enhancing the robustness against noise and outliers by forming the local similarity measurements. However, most correntropy-based models either have too simple descriptions of the correntropy or require too many parameters to adjust in advance, which is likely to cause poor performance since the correntropy fails to reflect the probability distributions of the signals. Therefore, in this paper, a novel correntropy-based extreme learning machine (ELM) called ECC-ELM has been proposed to provide a more robust training strategy based on the newly developed multi-kernel correntropy with the parameters that are generated using cooperative evolution. To achieve an accurate description of the correntropy, the method adopts a cooperative evolution which optimizes the bandwidths by switching delayed particle swarm optimization (SDPSO) and generates the corresponding influence coefficients that minimizes the minimum integrated error (MIE) to adaptively provide the best solution. The simulated experiments and real-world applications show that cooperative evolution can achieve the optimal solution which provides an accurate description on the probability distribution of the current error in the model. Therefore, the multi-kernel correntropy that is built with the optimal solution results in more robustness against the noise and outliers when training the model, which increases the accuracy of the predictions compared with other methods.

1. Introduction

With the rapid development of powerful computing environments and rich data sources, artificial intelligence (AI) technology such as neural networks [1,2,3], adaptive filtering [4,5,6] and evolutionary algorithms [7,8,9] has become increasingly more applicable for forecasting problems in various scenarios, such as medicine [10,11,12], economy [13,14,15] and electronic engineering [16,17,18]. The methods have acquired high reputations due to their great approximation abilities.
Although AI methods perform well when solving real world problems, most corresponding models adapt the mean squared error (MSE) as the criterion for training hidden nodes or building the cost functions, assuming that the data satisfy a Gaussian distribution. Moreover, the MSE is a global similarity measure where all the samples in the joint space have the same contribution [19]. Therefore, the MSE is likely to be badly affected by the noise and outliers that are hiding in the samples and this happens commonly in applications, such as speech signals, images, real-time traffic signals and electronic signals from ill-conditioned devices [20,21,22]. Therefore, MSE-based models are likely to result in poor performance in real world applications.
To conquer the weaknesses of the least mean squares (LMS), over the past decades, a number of studies have proposed methods to improve the robustness of the model against the noise and outliers that are contained in the data [23,24,25,26,27]. Among the existing technologies, M-estimators have been the focus of many academic studies. By detecting the potential outliers during training procedures, the M-estimator can eliminate the negative influences from the output weights that adversely affect the predictions [28]. Using these advantages, Zhou et al. [29] proposed a novel data-driven standard least-squares support vector regression (LSSVR) applying the M-estimator, which reduces the interference of outliers and enhances the robustness. However, there are difficulties accessing clean learning data without noises so that the application on the M-estimator-based forecasting models based is limited.
Recently, information theoretic learning (ITL) has drawn considerable attention due to its good performance avoiding the effect of the noise and outliers [30,31,32,33,34,35] and it has become an effective alternative to the MSE criterion. In [36], the authors presented a novel training criterion based on the minimum error entropy (MEE) to replace the MSE. By taking advantages of the higher order description on entropy, MEE has become superior for non-Gaussian signal processing compared with traditional training criteria. Inspired by the entropy and Parzen kernel estimator, Liu et al. [37] proposed an extended definition of the correlation function for random processes using a generalized correlation function, known as correntropy. Although different from global measurements, such as the mean squared error (MSE), the correntropy is regarded as a local similarity measurement where its value is primarily determined by the kernel function along x = y line [38], leading to high robustness against noise and outlies. Moreover, the correntropy has many great properties such as symmetry, nonnegativity and boundness. Most of all, it is easy to form convex cost functions based on the correntropy, which is very convenient for training the models [39,40,41,42]. Therefore, the correntropy has been widely used in forming robust models [43,44,45].
To enhance the forecasting ability of the model, in [46], the correntropy was introduced into the affine projection (AP) algorithm to overcome the degradation of the identification performance with impulsive noise environments. From the simulation results, it is easy to verify that the proposed algorithm has achieved better performance than other methods. Another approach to improve the robustness via the correntropy is enhancing the feature selection efficiencies [47,48,49]. In [50], the kernel modal regression and gradient-based variable identification were integrated together using the maximum correntropy criterion, which guarantees the robustness of the algorithm. Additionally, in [51], a novel principal component analysis (PCA), based on the correntropy and known as the correntropy-optimized temporal PCA (CTPCA), was adapted to enhance the robustness for rejecting the outlier. The outlier improves the models training in simulation experiments. In addition to providing the extractions of the features in neural networks and filtering methods, the correntropy turns out to be a powerful tool for developing robust training methods that generate and adjust the weights in the model. In [52], Wang et al. introduced a feedback mechanism using the kernel recursive maximum correntropy to provide a novel kernel adaptive filters known as the kernel recursive maximum correntropy with multiple feedback (KRMC-MF). The experiments show that the generated filters have high robustness against outliers. In [53], Ahmad et al. proposed the correntropy based conjugate gradient backpropagation (CCG-BP), which can achieve high robustness in environments with both impulsive noise and heavy-tailed noise distributions. Unfortunately, most of the neural networks have to adjust the weights of each node during each training iterations which wastes time during the training process.
Recently, forecasting models with parameters that are free from adjustments have gained increasingly more attention due to their fast training speeds for the models [54,55,56]. Combined with the correntropy, these algorithms have shown great potential in real-world applications. For example, Guo et al. [57] developed a novel training method for echo state networks (ESNs) based on a correntropy induced loss function (CLF), which provides robust predictions for time-series signals. Similar to ESNs, extreme learning machines (ELMs) have received great attention on fast learning due to the random assignments of the hidden layer and being equipped with simpler structures, such as single layer feedback networks (SLFNs) [58,59,60]. It has been proven that the hidden nodes can be assigned with any continuous probability distribution, while the model satisfies the universal approximation and classification capacity [61]. In particular, the extreme learning machine has been applied and received a high reputation for predicting production processes [62,63], system anomalies [64], etc. [65]. In [66], the authors first developed the correntropy-based ELM that uses the regularized correntropy criterion in place of the MSE with half quadratic (HQ) optimization which is called the regularized correntropy criterion for an extreme learning machine (RCC-ELM). Later, Chen et al. [67] extended the dimensions of the correntropy by combining two kinds of correntropy together to enhance the flexibility of the model to generate more robust ELM called ELM by maximum mixture correntropy criterion (MMCC-ELM). The experimental results show that the learning method performs better than the conventional maximum correntropy method. Although the RCC-ELM and MMCC-ELM possess high robustness compared with other ELM methods, the corresponding correntropy is constrained by no more than two kernels. The kernel bandwidth required for the assignments by users in advance is likely to degrade the model due to the improper description on the probability distribution of the signal with the correntropy.
To conquer the weakness of the existing correntropy-based ELMs, this paper focuses on providing a more robust predicting model with adaptive generation based on multi-kernel correntropy which can bring an accurate description of the current errors of ELM. This study developed a more flexible and robust forecasting ELM based on a newly developed adaptive multi-dimension correntropy using evolving cooperation. In the proposed method, the output weights of the ELM are trained based on the maximum multi-dimension correntropy with no constraints on the dimensions of the kernels. To achieve the most appropriate assignment of the parameters of each kernel in the correntropy, a novel evolving cooperation method is developed to concurrently optimize the bandwidths and the corresponding influence coefficients to achieve the best estimations of the residual errors of the model. Furthermore, the training approach has been developed based on the properties of the multi-dimension correntropy. The main contribution of the paper can be summarized as follows.
  • The proposed method develops a novel correntropy criterion with multiple kernels to improve the flexibility for depicting the probability distribution of the current error of the predicting model. Then, a convex cost function has been developed based on the multiple kernel correntropy, which can provide a more robust training strategy for ELMs, resulting in high performance on the predictions against noise and outliers.
  • To accurately describe the probability distribution of the current error, the proposed method develops a cooperating evolution strategy to adaptively generate proper bandwidths and coefficients to suit the error distribution which enhances the accuracy on the approximation for the correntropy, leading to more robust training.
The experiments compare the performance of the proposed method and several state-of-art methods using both simulated data and real-world data, which show that the proposed method obtains more the robust predictions than other methods. Finally, the proposed method is incorporated into the forecasting model for the current transfer ratio (CTR) signals for the optical couplers, and it achieves high accuracies and robustness.
The rest of the paper is as follows. The next section introduces the framework of the proposed method and multi-dimension correntropy. Section 3 describes the evolved cooperation for the kernels with multi-dimension correntropy and Section 4 provides the training procedures of the forecasting model. Then, Section 5 estimates the performance of the proposed method using both simulation data and real-world applications. Finally, the conclusion is drawn in Section 6.

2. The Framework of the Proposed Method

The structure of the prediction model that is built using the proposed method is similar to those of other ELM-based methods. Figure 1 shows the basic structure of the method. Generally, the network includes one input layer, one hidden layer and one output layer. The hidden output is calculated using the given input vectors and the weights and the biases of the hidden nodes which are randomly assigned [54]:
h = f ( w x + b )
where f(.) is the activation function and (w,b) are the weights and bias of the hidden nodes.
With the hidden layer, the network can simulate any kind of function by generating the output weights with the least mean squares (LMS) The cost function is calculated as follows [58]:
J LS = | | Y T | |
where T is the expected output and Y is the predicted output of the model. Y calculated with the hidden outputs h and the output weights β as follows:
Y = β h
Therefore, the output layer is calculated as follows:
β = ( H T H ) 1 H T T
Further, to constrain the output weights, the output layer is calculated as follows:
β = ( H T H + λ I ) 1 H T T
where λ is the constraining coefficient.
Although the output weights that are calculated by Equation (4) or Equation (5) can provide good predictions using the training data, the model has suffered with the outliers and noises in the data which negatively affect the predictions. To overcome the problem, the correntropy, as a high order similarity measurement, has been used in some recently developed methods.
In [62], the cost function built using the correntropy as follows:
J RCC = max β [ p = 1 N G ( t p h β ) λ | | β | | ]
where G(tp) is the Gussian kernel calculated as follows:
G ( t p h β ) = exp ( ( t p h β ) 2 σ 2 )
where σ is the bandwidth of the kernel.
Therefore, the output layer is calculated as follows:
β = ( H T Λ H λ I ) 1 H T Λ T
where Λ is the diagonal matrix of the local optimal solution. It is calculated as follows:
α p τ + 1 = G ( t p h β )
To further improve the flexibility of the correntropy, the cost function with a mixed correntropy is defined in [67] as follows:
J MMCC = 1 1 N i = 1 N [ α G σ 1 ( e i ) + ( 1 α ) G σ 2 ( e i ) ] + λ | | β | |
Therefore, the output is calculated as follows:
β = ( H T Λ H + λ I ) 1 H T Λ T
where the λ = 2 N λ and Λ is the diagonal matrix with elements calculated as follows:
Λ i i = α / σ 1 G σ 1 ( e i ) + ( 1 α ) / σ 2 G σ 2 ( e i )
With two coefficients, Equation (9) gives a more accurate estimation of the costs of the output layer, leading to a higher robustness of the model. Although Equations (7) and (9) can acquire better local similarity measurements compared with Equation (5), both criterions limit the correntropy into two kernels, leading to an inappropriate description on the probability distribution of the data. Additionally, the bandwidths and the coefficients must be assigned by users, thus limiting the performance of the corresponding model in real world applications which can be badly affected since the bandwidths are not suitable for the estimation of the correntropy. To provide a more flexible criterion for the training strategy with a more appropriate description of the probability distribution of the data, the proposed method develops a multi-kernel correntropy criterion that is calculated as follows:
k ( T β H ) = i = 1 K α i G σ i ( T β H )
where αi is the influence coefficients controlling the weight of each kernel. By using multiple kernels to construct the correntropy, the proposed method brings a more accurate approximation on the probability distribution of the samples, leading to a high prediction performance of the model. Based on the corretropy using Equation (13), the proposed method built a convex cost function for training the output weights, which has been analyzed in Section 4. For the suitable assignments of the parameters in Equation (13), a novel generation strategy using an evolved cooperating process based on SDPSO with the MIE to generate the parameters adaptively has been developed. Therefore, the framework of the proposed method can be summarized in Figure 2. The proposed method developed an evolved-cooperation strategy to generate the optimized solution of the influence coefficients and the bandwidths which suits the distribution of the prediction errors. To achieve an accurate estimation, the bandwidth was generated based on switching delayed particle swarm optimization (SDPSO) [68] and the influence coefficients were calculated based on the cost function for estimating the probability distribution function of errors.
The basic procedures of the method are as follows. Supposing that the input vector of the samples is represented as x = {x1, x2, …, xN}, calculate the output of hidden nodes with randomly assigned weights and biases as Equation (1). Then, adapt the cooperating evolution technology for training the output weights. For each iterations of the evolution, the output of the predicting model can be generated using Equation (3). Compared with the actual outputs, the predicted outputs result in current error e with the model. Based on the current error e, the proposed method makes the best assignments of the bandwidths in the correntropy with SDPSO and accesses the optimal coefficients based on MIE. This is shown in the next section. Using the generated correntropy, a list of diagnostic kernels can be calculated which effects the updating of the output layer to reach higher accuracy. This is presented in Section 4. The processes stop when the cost function of the model is stable.
More details are presented in the next section.

3. The Cooperating Evolution Process for the Bandwidth and Influence Coefficients of the Kernel

For the correntropy that is defined by Equation (12), the bandwidth and the influence coefficients are for the similarity measurements since the bandwidths act as the zoom lens for the measurements and the coefficients determine the effect that each kernel has on the estimation of the correntropy according to the assigned bandwidth. They are defined as follows:
σ = { σ 1 , σ 2 , σ 3 , , σ M }
A = { α 1 , α 2 , α 3 , , α M }
Therefore, the bandwidth and the influence coefficients should be carefully assigned to match the probability distribution of the samples to achieve the best effect of the correntropy on generating the output weights of the prediction model. Since the correntropy depicts the probability distribution of the distance between the actual output and the model response, the bandwidth and the coefficients are able to form the probability distribution (pdf) function as follows:
f ^ ( e ) = i = 1 N α 1 G σ 1 ( e i ) + α 2 G σ 2 ( e i ) + + α N G σ n ( e i )
In applications, the real joint probability distribution for the cases are unknown. Therefore, the joint pdf can only be estimated for a finite number of samples{(ti,yi)}, where i = 1, 2, …, N:
f ( e ) = 1 N g ( { ( t k , y k ) | | t k y k | = e } )
where g(S) is the cardinal number of the set S.
Using the kernel contrasts between the pdf estimated with the assigned parameters and the pdf estimated using the data, the least mean integrated error (MIE) can be calculated as follows:
MIE = E ( ( f ^ ( e ) f ( e ) ) 2 d e )
Based on the MIE, the performance of the bandwidth and coefficients can be estimated using the contrasts with the pdf from the data. Therefore, the optimization of these parameters can be transformed to finding the solution with the minimum MIE.
In the proposed method, the switching delay particle swarm optimization is adapted to search for the best bandwidth. To achieve this, the particles are initialized with a list of potential bandwidth setting σc = {σc,1, σc,2, …, σc,N}. With respect to each bandwidth of the particle, the velocities for the evolution of the particles are defined as follows:
c = {vσc,1, vσc,2,…, vσc,N}
Meanwhile, the influence coefficient is denoted as vector A:
A c = { α c , 1 , α c , 2 , α c , 3 , , α c , M }
where αi is the influence coefficient according to σc,i.
Since the samples provide disperse values of the outputs, the pdf from the data is estimated using the discrete version of Equation (16):
F = { f ( m 1 ) , f ( m 2 ) , , f ( m k ) }
f ( m ) = 1 N g ( { ( t k , y k ) | m ε | t k y k | m + ε } )
where the vector m = {m1, m2, …, mk} is a list of values that satisfy m1 < m2 < … < mk and |mi − mi−1| = ε. ε is the step length of the estimation.
Accordingly, the values from Equation (15) with respect to m are equivalent to the following set:
F ^ = { f ^ ( m 1 ) , f ^ ( m 2 ) , , f ^ ( m k ) }
They can be calculated as:
F ^ = A K
where K is the kernel matrix, which is as follows:
K = [ G σ 1 ( e 1 ) G σ 1 ( e 2 ) G σ 2 ( e 1 ) G σ 2 ( e 2 ) G σ 1 ( e N ) G σ 2 ( e N ) G σ M ( e 1 ) G σ M ( e 2 ) G σ M ( e N ) ]
By inserting Equations (20) and (22) into Equation (17), the following cost function can be obtained:
MIE = ( AK F ) ( AK F ) T
Then, the following differential equations with respect to A are calculated:
2 ( AK F ) = 0
Therefore, the coefficient can be calculated using the assigned bandwidth as follows:
A = F K T ( K K T ) 1
Since each particle contains one solution for the kernels’ parameters, the personal best solution pσ and the global best solution gσ is updated by minimizing the costs. Then, the particles are updated as follows:
v σ c ( k + 1 ) = wv σ c + c 1 ( k ) × r 1 ( p σ ( k ( k ) ) σ c ( k ) ) ) + c 2 ( k ) × r 2 ( g σ ( k τ 2 ( k ) σ c ( k ) ) )
σ c ( k + 1 ) = σ c ( k ) + v σ c ( k + 1 )
where c1(k) and c2(k) are the acceleration coefficients and τ1(k) and τ2(k) are the time delays. All the parameters are adjusted based on the evolution factor, Ef, which determines the evolutionary states, and it is calculated as follows:
Ef = ( d g d min ) / ( d max d min )
where dg is the global best particle among the mean distance. It is calculated as:
d g = 1 N i = 1 N | | σ c , i g σ | |
With the estimate on Ef, the parameters can be selected as shown in Table 1.
The final solution of the bandwidth and the influence coefficients are determined as the solution that minimizes the costs during the evolution procedures.
In summary, the cooperative evolution process is shown in Algorithm 1. First, the bandwidth and the corresponding velocity of each particle are randomly assigned. Then, for each iteration of the process, the influence coefficients are evolved using the bandwidth based on the MIE and the particles are updated using the cost function. Finally, the algorithm finds the best solutions for the bandwidth and the influence coefficients, from which the kernel depicts the pdf from the data. Based on the generated kernel, the correntropy can lead to a model with good robustness.
Algorithm 1 Evolved cooperation for the kernel parameters
Input: the samples {xi,ti},i = 1, 2, …, N
Output: the vector of bandwidth σ and the vector of influence coefficients A
  Parameters: the step length and the number of iterations L
  Initialization: Set the cost function of the best solution MIEbest to and randomly assign the bandwidth of the kernels σc = {σc,1, σc,2, …, σc,N} and the corresponding velocity c = {vσc,1, vσc,2, …, vσc,N}.
1: for k = 1, 2, … L do
2:    Generate the best influence coefficients Ac using Equation (26) for each particles.
3:    Calculate value of cost function for each particle MIEc based on Equation (24)
4:    Update the personal best solution p σ and the global best solution g σ based on minimizing the cost function.
5:    Calculate the Ef of the iteration with Equation (29)
6:    Access the parameters for evolution based on Table 1
7:    Update the swarm with Equations (27) and (28)
8: end for
9: Return the global best bandwidth g σ and the corresponding influence coefficients

4. Training the Extreme Learning Machine Using the Multi-Dimension Correntropy

To improve the robustness of the extreme learning machine, in the proposed method, the training procedure of the output layer as Equation (5), is replaced by the developed calculation using the mixture correntropy that is generated using the evolved kernel from Section 3. The loss function for the output layer is developed according to the following properties.
Property 1.
K(Y,T) is symmetric, which means the following: K(Y,T) = K(T,Y).
Property 2.
K(T,Y) is positive and bounded, which means the following: 0 < K(Y,T) < = 1 and K(T,Y) = 1 if and only if T = Y.
Property 3.
K(T,Y) involves all the even moments of e, which means the following:
K ( T , Y ) = E [ e 2 n ] n = 0 ( 1 ) n i = 1 M α i σ i 2 n 2 n i = 1 M ( σ i ) 2 n n !
Property 4.
When the first bandwidth is large enough, it satisfies the following:
K ( T , Y ) i = 1 M α i i = 1 M α i σ i 2 2 i = 1 M σ i 2 E [ e 2 ]
Proof. 
For lim x 0 exp ( x ) 1 + x , suppose that σ1 is large enough, K(T,Y) can be approximated as follows:
K ( T , Y ) = α 1 G σ 1 ( e ) + α 2 G σ 2 ( e ) + + α m G σ m ( e ) = α 1 ( 1 e 2 2 σ 1 2 ) + α 2 ( 1 e 2 2 σ 2 2 ) + + α m ( 1 e 2 2 σ m 2 ) = i = 1 m α i i = 1 M α i σ i 2 2 i = 1 M σ i 2 E [ e 2 ]
that completes the proof. □
Remark 1.
Based on Property 4, the mixed C-loss is defined as L(T,Y) = 1 − K(T,Y), which is approximately equivalent to the mean square error (MSE) with a large enough bandwidth.
Property 5.
The empirical mixed C-loss L(e) that is a function of e is convex at any point satisfying | | e | | = max | e i | σ 1 .
Proof. 
Build the Hessian matrix of the C-loss function L(e) with respect to e as follows:
H L ( e ) = [ L ( e ) e i e j ] = diag ( ξ 1 , ξ 2 , , ξ N )
The elements of matrix ξ is calculated as follows:
ξ i = i = 1 m α i σ i 4 e i 4 N σ i 4 G σ i ( e i )
It is obvious that ξ i is positive. Therefore, L(e) is convex. □
Remark 2.
Using Property 4 and Property 5, the loss function of the output weights is based on the empirical mixed C-loss L(e) from the data observations, which can be defined as follows:
J = L ( T , Y ) + λ | | β | | 2 = 1 1 N i = 1 N j = 1 M α j G σ j ( e i ) + λ | | β | | 2
Based on Equation (38), the training criterion is generated for improvement on the robustness of the model.
Taking the differential of the loss function, it is easy to get the following:
J ( β ) β = 0 i = 1 N { [ j = 1 M α j σ j 2 G σ j ( e i ) ] e i h i T } + 2 N λ β = 0 i = 1 N ( φ ( e i ) h i T h i β φ ( e i ) t i h i T ) + λ β = 0 i = 1 N ( φ ( e i ) h i T h i β + λ β = i = 1 N ( φ ( e i ) t i h i T ) β = [ H T Λ H + λ I ] 1 H T Λ T
where λ = 2 N λ , φ ( e i ) = j = 1 M α j σ j 2 G σ j ( e i ) and Λ is a diagonal matrix with diagonal elements Λ i i = φ ( e i ) , which provides the local similarity measurements between the predicted output and the actual outputs. When the training data contain large noise or many outliers, the corresponding diagonal elements are relatively low which induce the effects of such samples. Therefore, the algorithm can achieve high robustness against noises and outliers in the signals.
Since Equation (37) is a fixed-point equation because the diagonal matrix depends on the weight vector, the optimal solution should be solved by applying the evolved cooperation using Equation (37).
Therefore, combined with the kernel optimization in Section 3, the whole training process can be summarized in Algorithm 2, which is referred to as the ECC-ELM algorithm in this paper.
Algorithm 2 ECC-ELM
Input: the samples {xi,ti}, i = 1, 2, …, N
Output: output weights
Parameters: the number of hidden nodes N, the number of iterations L, the iterations T and termination tolerance ε
Initialization: Randomly set the weights and bias of the hidden nodes and initialize the output weights β using Equation (5)
1: for t = 1, 2, …, T do
2:    Calculate the residual error: ei = tihi β , i = 1, 2, …, N
3:    Calculate the kernel parameters { σ , A } using Algorithm 1
4:    Calculate the diagonal matrix Λ: Λ i i = φ ( e i ) = j = 1 M α j G σ j ( e i )
5:    Update the output weight using Equation (37)
6:    Until |Jk( β ) − Jk−1( β )| < ε
7: end for

5. Analysis on Time Complexity and Space Complexity of ECC-ELM

In this section, the time complexity of the proposed method is analyzed and compared with the other algorithms. The main time complexity of the ECCELM comes from the cooperating evolution process and the training process of the model. The cooperative evolution contains the calculations of the influence coefficients and the particles updating with the time complexity of O(ItNK2), where It is the number of iterations, N is the number of particles and K is the number of disperse values of the outputs. To train the ELM, the procedures share the same time complexity as the RCC-ELM and MMCC-ELM, which is O(IhNl(5M+M2)), where Ih is the amount of iterations for training and Nl is the number of training data. Additionally, M is the number of hidden nodes. Therefore, the time complexity of ECC-ELM is O(IhNl(5M+M2+ItNK2)), which is slightly higher than those of the RCC-ELM and MMCC-ELM but it satisfies the requirements in most applications.
With respect to the spatial complexity, the ECC-ELM has the same complexity as the prediction models using the RCC-ELM, which is O(N+(N+2)M+Nl2). Additionally, the space complexity consumed by evolving process is O(2N+K). Therefore, the space complexity of ECC-ELM is O(N+(N+2)M+Nl2+2N+K), which has the same order as RCC-ELM and MMCC-ELM.
In summary, the time complexity and spatial complexity are practical for most applications.

6. Experiments

6.1. The Simulation of the Sinc Function with Sas noises

In this section, the simulation experiments using the Sinc function with random noises are presented. They compare between serval state-of-art algorithms with the proposed method, which are the R-ELM, the RCC-ELM, the MMCC-ELM and our method. The training and test samples were randomly assigned according to the Sinc function and random noises were added with respect to alpha-stable distribution. This is represented as follows:
y = α Sinc ( x ) + ρ
where α is the scale of the function which is set to 8.0 and Sinc(x) is the Sinc function. The Sinc function is represented as follows:
Sinc ( x ) = { sin ( x ) / x x 0 1 x = 0
Moreover, ρ is the noise that satisfies the following characteristic function [69]:
ρ = { exp ( δ α | θ | α ( 1 j β s i g n ( θ ) tan ( π α 2 ) ) ) + j μ θ α 1 exp ( δ 1 | θ | 1 ( 1 j β ( π / 2 ) s i g n ( θ ) log ( π α 2 ) ) ) α = 1
The parameters α, β, γ and μ are real and characterize the distribution of the random variable X. Here, the alpha-stable probability distribution function is denoted as S(α,β,γ,μ). In these experiments, the four parameters were assigned to three different conditions to provide three types of noises. The assignment of the parameters in each sample is presented in Table 2.
Each sample contained 200 data, with half of the data being used for training and another half for testing. To get a proper estimation of the performances of each method, the experiments were operated with the best optimization of parameters. This is presented in Table 3.
Each experiment was conducted 30 times and the averages were taken. The comparison of the accuracies of these algorithms is presented in Table 4. Compared with other algorithms, the R-ELM and ECC-ELM achieve lower mean square errors due to the advantages of the correntropy. The performance of R-ELM is relatively poor due to the effect of noises. The performance of MMCC-ELM also improved by the correntropy. However, since the fixed dimension of the correntropy, the accuracy can be badly influenced by unnecessary assignments on the second order of the bandwidth. Furthermore, it is clear that the proposed algorithm achieves the lowest training MSE, which means that it is the most accurate method for simulation of the Sinc function.
To further analyze the predictive abilities of these four algorithms, Figure 3 depicts the differences between the actual function and the predicted function for each algorithm. It is clear that all the algorithms achieve relatively good prediction on the Sinc function. However, the prediction results of the ELM have been badly influenced by the noises in all three samples. Additionally, the MMCC-ELM performance is poor on sample 2 and sample 3, which is probably due to the assignments with high dimension parameters. The RCC-ELM and ECC-ELM provide good predictions, which are almost identical to the actual functions in all three samples. The ECCELM has the closet predicted function with the Sinc function, which also proves that the method has high reliability against noise.
Furthermore, an experiment on sample 1 was conducted to compare the cost function for the output weights with the MMCC-ELM and ECC-ELM since they share similar cost functions. The results are shown in Figure 4, which show that the cost function of ECC-ELM is quite lower than the cost of MMCC-ELM. Additionally, the costs of the ECC-ELM become stable for less than 25 iterations for all three examples than MMCC-ELM. This shows the improvements on training the model with ECC-ELM taking the cooperating evolution technique. Since both algorithms finish the generation of the model when the cost function becomes stable, it can be concluded that the proposed model has faster convergence on training the prediction model.
Figure 5 illustrates the effects of the evolutionary process on the optimization of the kernel bandwidth and influence coefficients. From Figure 5, it can be seen that the cost function for the kernel bandwidth quickly drops during the evolution process. Moreover, Ef continuously decreases during the process, which means that the particle swarm become stable and the best solution occurs. Figure 6 compares the actual pdf function and the estimated pdf function. It can be seen that the algorithm achieves a comparatively accurate estimation of the distribution of the errors.

6.2. The Performance Comparison on Benchmark datasets

To further assess the proposed algorithm, the performance of the ECC-ELM and other methods were compared using the data set from the UCI machine learning repository [70], awesome public dataset [71] and the United Nations development program [72], which are listed in Table 5. The assignments of the parameters are shown in Table 6, all of which refer to the best performance of each algorithm. Each experiment was conducted 30 times and the average performance was reported.
The performance is compared in Table 7, which shows that the proposed algorithm is able to achieve better prediction accuracies than other methods. Additionally, the performance of the proposed method is relatively stable compared with other correntropy-based extreme learning machines.
Figure 7 compares the actual output value and the predicted value for the Servo data set. It is clear that the predicted values are basically identical to the actual output values, and it has not been influenced by the outliers in the data.
To illustrate the evolutionary processes for optimizing the bandwidth, Figure 8 depicts the distributions of the particles and the evolution of the optimal solutions. It can be seen that the distribution of the particles dynamically changes based on the state of the PSO process. The optimal solution is adjusted and stabilizes during the process, which allows the optimal solution of the bandwidth assignments to generate a more accurate model.

6.3. The Performance Estimations for Forecasting the CTR of Optical Couplers

Finally, to estimate the performance of a real application, the proposed method has been used to predict the current transfer ratio for optical couplers. This is one type of transmission device for electric signals and optical signals with wide applications to the isolation transfer of signals, A/D transmission, D/A transmission, digital communications and high-pressure control. For optical couplers, the CTR is an essential factor for estimating the operating status of optical couplers. In this section, the proposed method was used to give the predictions of CTR for the optical couplers to predict the health condition of the devices.
For the experiments, the degenerating signals of four optical couplers were recorded and transformed into the samples historical CTR value as input vectors and the CTR value of the next time as the expected output. The training data was the samples that were generated from the optical couplers’ records over the first ten years and the testing data were the samples that were generated from the last ten years.
Figure 9 depicts the evolutionary process of the PSO procedure. It shows that the Ef value quickly decreases during the evolutionary process and stabilizes within 17 iterations, resulting in the optimal solution that is provided by the swarm.
Finally, the predicted results of the four optical couplers are shown in Figure 10. It is clear that the generated ELM network accurately predicts the CTR value of each optical coupler and is robust with the noises of the signals. Therefore, the proposed method is able to achieve good performance for the optical couplers.
Table 8 presents the numerical results of the CTR prediction, which compares the actual CTR and the predicted CTR. It is clear that the proposed method can very accurately provide the prediction on the state of Optical Couplers (OCs). Additionally, the time consumption is presented in Table 8 which shows that the proposed method is able to obtain high accuracy on the prediction of the future CTR of the OC and the predicting time is quite low within 5 ms. Therefore, the proposed method can achieve high performance on real applications.

7. Conclusions

To improve the robustness of the forecasting model, the paper provides a novel correntropy-based ELM called the ECC-ELM. It uses a multi-dimension correntropy criterion and the evolved cooperation method to adaptively generate the parameters for kernels. In the proposed algorithm, SDPSO is integrated by minimizing the MIE to determine the proper bandwidths and their corresponding influence coefficients to estimate the probability distributions of the residual error of the model. A novel training process was developed based on the properties of the multi-dimension correntropy and it was able to build the convex cost function to calculate the output weights for the ELM. The experiments on the simulated data and real-world application were conducted to estimate the accuracy of the probability distribution of the signal and robustness on predicting the samples. The simulation results with the Sinc function proved that the proposed method can generate the multi-kernel correntropy with high accuracy on describing the probability distribution of the signals and fast converge on the evolution process. This leads to high robustness of the proposed method compared with the other methods. The performance comparisons on the benchmark datasets show that the proposed method can achieve higher accuracy and more stability than the other methods. Finally, the CTR prediction experiments show the proposed method can achieve high accuracy within acceptable time consumption on real world applications. Although the proposed algorithm has predictive advantages, there are still several limitations on the study. One limitation is the proposed method is only applicable for an ELM with one hidden layer, which requires extensions on multi-layer networks. The other limitation is that the proposed method only provides an offline training model. Therefore, how to update the online prediction model becomes another interesting topic for future research. The codes and data of the research are available at https://github.com/mwj1997/ECC-ELM.

Author Contributions

Conceptualization, W.M.; Data curation, L.D.; Funding acquisition, Z.L.; Investigation, J.H.; Methodology, W.M. and Y.S.; Project administration, Z.L.; Resources, J.H.; Software, W.M.; Supervision, Z.L.; Validation, Y.S.; Writing—original draft, W.M.; Writing—review & editing, L.D.

Funding

This work was supported by the National Natural Science Foundation of China under Grants No. U1830133 (NSAF) and No. 61271035.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Heddam, S.; Keshtegar, B.; Kisi, O. Predicting total dissolved gas concentration on a daily scale using kriging interpolation, response surface method and artificial neural network: Case study of Columbia river Basin Dams, USA. Nat. Resour. Res. 2019, 2, 1–18. [Google Scholar] [CrossRef]
  2. Ahmadi, N.; Nilashi, M.; Samad, S.; Rashid, T.A.; Admadi, H. An intelligent method for iris recognition using supervised machine learning techniques. Opt. Laser Thechnol. 2019, 120, 105701. [Google Scholar] [CrossRef]
  3. Aeukumar, R.; Karthigaikumar, P. Multi-retinal disease classification by reduced deep learning features. Neural Comput. Appl. 2017, 28, 329–334. [Google Scholar]
  4. Pentapati, H.K.; Teneti, M. Robust speaker recognition systems with adaptive filter algorithms in real time under noisy conditions. Adv. Decis. Sci. Image Process. Secur. Comput. Vis. 2020, 4, 1–18. [Google Scholar]
  5. Eweda, E. Stability bound of the initial mean-square division of high-order stochastic gradient adaptive filtering algorithms. IEEE Trans. Signal Process. 2019, 6, 4168–4176. [Google Scholar] [CrossRef]
  6. Huang, X.; Wen, G.; Liangm, L.; Zhang, Z.; Tan, Y. Frequency phase space empirical wavelet transform for rolling bearing fault diagnosis. IEEE Access. 2019, 7, 86306–86318. [Google Scholar] [CrossRef]
  7. Yang, J.; Zhu, H.; Liu, T. Secure and economical multi-cloud storage policy with NSGA-II-C. Appl. Soft Comput. 2019, 83, 105649. [Google Scholar] [CrossRef]
  8. Albasri, A.; Abdali-Mohammadi, F.; Fathi, A. EEG electrode selection for person identification thru a genetic-algorithm method. J. Med. Syst. 2019, 43, 297. [Google Scholar] [CrossRef]
  9. Dermanaki Farahani, Z.; Ahmadi, M.; Sharifi, M. History matching and uncertainty quantification for velocity dependent relative permeability parameters in a gas condensate reservoir. Arab. J. Geosci. 2019, 12, 454. [Google Scholar] [CrossRef]
  10. Shah, P.; Kendall, F.; Khozin, S.; Goosen, R.; Hu, J.; Laramine, J.; Ringel, M.; Schork, N. Artificial intelligence and machine learning in clinical development: A translational perspective. Nature 2019, 2, 1–5. [Google Scholar] [CrossRef]
  11. Shirwaikar, R.D.; Dinesh, A.U.; Makkithaya, K.; Suruliverlrajan, M.; Srivastava, S.; Leslie, E.S.; Lewis, U. Optimizing neural network for medical data sets: A case study on neonatal apnea prediction. Artif. Intell. Med. 2019, 98, 59–76. [Google Scholar] [CrossRef] [PubMed]
  12. Lucena, O.; Souza, R.; Rittner, L.; Frayne, R.; Lotufo, R. Convolutional neural network for skull-stripping in brain MR imaging using silver standard masks. Artif. Intell. Med. 2019, 98, 48–58. [Google Scholar] [CrossRef] [PubMed]
  13. Guan, H.; Dai, Z.; Guan, S.; Zhao, A. A neutrosophic forecasting model for time series based on first-order state and information entropy of high-order fluctuation. Entropy 2019, 21, 455. [Google Scholar] [CrossRef]
  14. Tymoshchuk, O.; Kirik, O.; Dorundiak, K. Comparative analysis of the methods for assessing the probability of bankruptcy for Ukrainian enterprises. In Lecture Notes in Computational Intelligence and Decision Making; Springer: Basel, Switzerland, 2019; pp. 281–293. [Google Scholar]
  15. Yang, T.; Jia, S. Research on artificial intelligence technology in computer network technology, International conference on artificial intelligence and security. In Proceedings of the 5th International Conference on Artificial Intelligence and Security (ICAIS 2019), New York, NY, USA, 26–28 July 2019; pp. 488–496. [Google Scholar]
  16. Senguta, E.; Jain, N.; Garg, D.; Choudhury, T. A review of payment card fraud detection methods using artificial intelligence. In Proceedings of the International Conference on Computational Techniques, Electronics and Mechanical Systems (CTEMS), Belagavi, India, 21–23 December 2018; pp. 494–499. [Google Scholar]
  17. Ampatzidis, Y.; Partel, V.; Meyering, B.; Albercht, U. Citrus rootstock evaluation utilizing UAV-based remote sensing and artificial intelligence. Comput. Electron. Agric. 2019, 164, 104900. [Google Scholar] [CrossRef]
  18. Yue, D.; Han, Q. Guest editorial special issue on new trends in energy internet: Artificial intelligence-based control, network security and management. IEEE Trans. Syst. Man Cybern. Syst. 2019, 49, 1551–1553. [Google Scholar] [CrossRef]
  19. Liu, W.; Pokharel, P.P.; Principe, J.C. The kernel least mean square algorithm. IEEE Trans. Signal Process. 2008, 56, 543–554. [Google Scholar] [CrossRef]
  20. Vega, L.R.; Rey, H.; Benesty, J.; Tressens, S. A new robust variable step-size NLMS algorithm. IEEE Trans. Signal Process. 2008, 56, 1878–1893. [Google Scholar] [CrossRef]
  21. Vega, L.R.; Rey, H.; Benesty, J.; Tressens, S. A fast robust recursive least-squares algorithm. IEEE Trans. Signal Process. 2008, 57, 1209–1216. [Google Scholar] [CrossRef]
  22. Ekpenyong, U.E.; Zhang, J.; Xia, X. An improved robust model for generator maintenance scheduling. Electr. Power Syst. Res. 2012, 92, 29–36. [Google Scholar] [CrossRef]
  23. Huang, Y.; Lee, M.-C.; Tseng, V.S.; Hsiao, C.; Huang, C. Robust sensor-based human activity recognition with snippet consensus neural networks. In Proceedings of the IEEE 16th International Conference on Wearable and Implantable Body Sensor Networks (BSN), Chicago, IL, USA, 19–22 May 2019. [Google Scholar]
  24. Ning, C.; You, F. Deciphering latent uncertainty sources with principal component analysis for adaptive robust optimization. Comput. Aided Chem. Eng. 2019, 46, 1189–1194. [Google Scholar]
  25. He, C.; Zhang, Q.; Tang, Y.; Liu, S.; Liu, H. Network embedding using semi-supervised kernel nonnegative matrix factorization. IEEE Access. 2019, 7, 92732–92744. [Google Scholar] [CrossRef]
  26. Bravo-Moncayo, L.; Lucio-Naranjo, J.; Chavez, M.; Pavon-Garcia, I.; Garzon, C. A machine learning approach for traffic-noise annoyance assessment. Appl. Acoust. 2019, 156, 262–270. [Google Scholar] [CrossRef]
  27. Santos, J.D.A.; Barreto, G.A. An outlier-robust kernel RLS algorithm for nonlinear system identification. Nonlinear Dyn. 2017, 90, 1707–1726. [Google Scholar] [CrossRef]
  28. Guo, W.; Xu, T.; Tang, K. M-estimator-based online sequential extreme learning machine for predicting chaotic time series with outliers. Neural Comput. Appl. 2017, 28, 4093–4110. [Google Scholar] [CrossRef]
  29. Zhou, P.; Guo, D.; Wang, H.; Chai, T. Data-driven robust M-LS-SVR-based NARX modeling for estimation and control of molten iron quality indices in blast furnace ironmaking. IEEE Trans. Neural Netw. Learn. 2018, 29, 4007–4021. [Google Scholar] [CrossRef]
  30. Ma, W.; Qiu, J.; Liu, X.; Xiao, G.; Duan, J.; Chen, B. Unscented Kalman filter with generalized correntropy loss for robust power system forecasting-aided state estimation. IEEE Trans. Ind. Inf. 2019. [Google Scholar] [CrossRef]
  31. Safarian, C.; Ogunfunmi, T. The quaternion minimum error entropy algorithm with fiducial point for nonlinear adaptive systems. Signal Process. 2019, 163, 188–200. [Google Scholar] [CrossRef]
  32. Dighe, P.; Asaei, A.; Bourlard, H. Low-rank and sparse subspace modeling of speech for DNN based acoustic modeling. Speech Commun. 2019, 109, 34–45. [Google Scholar] [CrossRef]
  33. Li, L.-Q.; Wang, X.-L.; Xie, W.-X.; Liu, Z.-X. A novel recursive T-S fuzzy semantic modeling approach for discrete state-space systems. Neurocomputing 2019, 340, 222–232. [Google Scholar] [CrossRef]
  34. Hajiabadi, M.; Hodtani, G.A.; Khoshbin, H. Robust learning over multi task adaptive networks with wireless communication links. IEEE Trans. Comput. Aided Des. 2019, 66, 1083–1087. [Google Scholar]
  35. Kutz, N.J. Neurosensory network functionality and data-driven control. Curr. Opin. Syst. Biol. 2019, 3, 31–36. [Google Scholar] [CrossRef]
  36. Chen, B.; Xing, L.; Zheng, N.; Principe, J.C. Quantized minimum error Entropy criterion. IEEE Trans. Neural Netw. Learn. Syst. 2019, 30, 1370–1380. [Google Scholar] [CrossRef]
  37. Liu, W.; Pokharel, P.P.; Principe, J.C. Correntropy: Properties and applications in non-guassian signal processing. IEEE Trans. Signal Process. 2007, 55, 5286–5298. [Google Scholar] [CrossRef]
  38. Kuliova, M.V. Factor-form Kalman-like implementations under maximum correntropy criterion. Signal Process. 2019, 160, 328–338. [Google Scholar] [CrossRef]
  39. Ou, W.; Gou, J.; Zhou, Q.; Ge, S.; Long, F. Discriminative Multiview nonnegative matrix factorization for classification. IEEE Access. 2019, 7, 60947–60956. [Google Scholar] [CrossRef]
  40. Wang, Y.; Yang, L.; Ren, Q. A robust classification framework with mixture correntropy. Inform. Sci. 2019, 491, 306–318. [Google Scholar] [CrossRef]
  41. Moustafa, N.; Turnbull, B.; Raymond, K. An ensemble intrusion detection technique based on proposed statical flow features for protecting network traffic of internet of things. IEEE Internet Things J. 2019, 6, 4815–4830. [Google Scholar] [CrossRef]
  42. Wang, G.; Zhang, Y.; Wang, X. Iterated maximum correntropy unscented Kalman filters for non-Gaussian systems. Signal Process. 2019, 163, 87–94. [Google Scholar] [CrossRef]
  43. Peng, J.; Li, L.; Tang, Y.Y. Maximum likelihood estimation-based joint sparse representation for the classification of hyperspectral remote sensing images. IEEE Trans. Neural Netw. Learn. Syst. 2019, 30, 1790–1802. [Google Scholar] [CrossRef]
  44. Masuyama, N.; Loo, C.K.; Wermter, S. A kernel Bayesian adaptive resonance theory with a topological structure. Int. J. Neural Syst. 2019, 29, 1850052. [Google Scholar] [CrossRef]
  45. Shi, W.; Li, Y.; Wang, Y. Noise-free maximum correntropy criterion algorithm in non-Gaussian environment. IEEE Trans. Circuits Syst. II Express Briefs 2019. [Google Scholar] [CrossRef]
  46. Jiang, Z.; Li, Y.; Hunag, X. A correntropy-based proportionate affine projection algorithm for estimating sparse channels with impulsive noise. Entropy 2019, 21, 555. [Google Scholar] [CrossRef]
  47. He, R.; Zheng, W.-S.; Hu, B.-G. Maximum correntropy criterion for robust face recognition. IEEE Trans. Patt. Anal. Mach. Intell. 2019, 33, 1561–1576. [Google Scholar]
  48. Macheshwari, S.; Pachori, R.B.; Rajendra, U. Automated diagnosis of glaucoma using empirical wavelet transform and correntropy features extracted from fundus images. IEEE J. Biol. Health Inf. 2017, 21, 803–813. [Google Scholar] [CrossRef]
  49. Moharmmadi, M.; Noghabi, H.S.; Hodtani, G.A.; Mashhadi, H.R. Robust and stable gene selection via maximum minimum correntropy criterion. Geomics 2016, 107, 83–87. [Google Scholar]
  50. Guo, C.; Song, B.; Wang, Y.; Chen, H.; Xiong, H. Robust variable selection and estimation based on modal regression. Entropy 2019, 21, 403. [Google Scholar] [CrossRef]
  51. Luo, X.; Xu, Y.; Wang, W.; Yuan, M.; Ban, X.; Zhu, Y.; Zhao, W. Towards enhancing stacked extreme learning machine with sparse autoencoder by correntropy. J. Frankl. Inst. 2018, 355, 1945–1966. [Google Scholar] [CrossRef]
  52. Wang, S.; Dang, L.; Wang, W.; Qian, G.; Chi, K.T.S.E. Kernel adaptive filters with feedback based on maximum correntropy. IEEE Access. 2018, 6, 10540–10552. [Google Scholar] [CrossRef]
  53. Heravi, A.R.; Hodtani, G.A. A new correntropy-based conjugate gradient backpropagation algorithm for improving training in neural networks. IEEE Trans. Neural Netw. Learn. Syst. 2018, 29, 6252–6263. [Google Scholar] [CrossRef]
  54. Jaeger, H.; Lukosevicious, M.; Popovivi, D.; Siewert, U. Optimization and applications of echo state networks with leaky integrator neurons. Neural Netw. 2007, 20, 335–352. [Google Scholar] [CrossRef]
  55. Tanaka, G.; Yamane, T.; Heroux, J.B.; Nakane, R.; Kanazawa, N.; Takeda, S.; Numata, H.; Nakano, D.; Hirose, A. Recent advances in physical reservoir computing: A review. Neural Netw. 2019, 115, 100–123. [Google Scholar] [CrossRef]
  56. Obst, O.; Trinchi, A.; Hardin, S.G.; Chawick, M.; Cole, I.; Muster, T.H.; Hoschke, N.; Ostry, D.; Price, D.; Pham, K.N. Nano-scale reservoir computing. Nano Commun. Netw. 2013, 4, 189–196. [Google Scholar] [CrossRef] [Green Version]
  57. Guo, Y.; Wang, F.; Chen, B.; Xin, J. Robust echo state network based on correntropy induced loss function. Neurocomputing 2017, 267, 295–303. [Google Scholar] [CrossRef]
  58. Huang, G.; Chen, L. Convex incremental extreme learning machine. Neurocomputing 2007, 70, 3056–3062. [Google Scholar] [CrossRef]
  59. Huang, G.; Zhou, H.; Ding, X.; Zhang, R. Extreme learning machine for regression and multiclass classification. IEEE Trans. Syst. Man Cybern. Part B 2012, 42, 513–529. [Google Scholar] [CrossRef]
  60. Tang, J.; Deng, C.; Huang, G. Extreme learning machine for multilayer perceptron. IEEE Trans. Neural Netw. Learn. Syst. 2016, 27, 809–821. [Google Scholar] [CrossRef]
  61. Huang, G.; Bai, Z.; Lekamalage, L.; Vong, C.M. Local receptive fields based extreme learning machine. IEEE Comput. Intell. Mag. 2015, 10, 18–29. [Google Scholar] [CrossRef]
  62. Arabilli, S.F.; Najafi, B.; Alizamir, M.; Mosavi, A.; Shamshirband, S.; Rabczuk, T. Using SVM-RSM and ELM-RSM Approaches for optimizing the production process of Methyl and Ethyl Esters. Energies 2018, 11, 2889. [Google Scholar]
  63. Ghazvinei, P.T.; Darvishi, H.H.; Mosavi, A.; Yusof, K.b.W.; Alizamir, M.; Shamshirband, S.; Chau, K.-W. Sugarcane growrh prediction based on meteorological parameter using extreme learning machine and artificial neural network. Eng. Appl. Comp. Fluid. 2018, 12, 738–749. [Google Scholar]
  64. Shamshirband, S.; Chronopoulos, A.T. A new malware delectation system using a high performance ELM method. In Proceedings of the 23rd international database applications & engineering symposium, Athens, Greece, 10–12 June 2019; p. 33. [Google Scholar]
  65. Bin, G.; Yan, X.; Yang, X.; Gary, W.; Shuyong, L. An intelligent time-adaptive data-driven method for sensor fault diagnosis in induction motor drive system. IEEE Trans. Ind. Electr. 2019, 66, 9817–9827. [Google Scholar]
  66. Xing, H.; Wang, X. Training extreme learning machine via regularized correntropy criterion. Neural Comput. Appl. 2013, 23, 1977–1986. [Google Scholar] [CrossRef]
  67. Chen, B.; Wang, X.; Lu, N.; Wang, S.; Cao, J.; Qin, J. Mixture correntropy for robust learning. Pattern Recognit. 2018, 79, 318–327. [Google Scholar] [CrossRef]
  68. Zeng, N.; Zhang, H.; Liu, W.; Liang, J.; Alsaadi, F.E. A switching delayed PSO optimized extreme learning machine for short-term load forecasting. Neurocomputing 2017, 240, 175–182. [Google Scholar] [CrossRef]
  69. Weron, A.; Weron, R. Computer simularion of Levy alpha-stable variables and processes. In Lecture Notes in Pihysics; Springer: Berlin/Heidelberg, Germany, 1995; pp. 379–392. [Google Scholar]
  70. Frank, A.; Asuncion, A. UCI Machine Learning Repository; University of California, School of Information and Computer Science: Irvine, CA, USA, 2010. [Google Scholar]
  71. Awesome Data. Available online: http://www.awesomedata.com/ (accessed on 16 September 2015).
  72. Human Development Reports. Available online: http://hdr.undp.org/en/data# (accessed on 15 September 2019).
Figure 1. The structure of the prediction model.
Figure 1. The structure of the prediction model.
Entropy 21 00912 g001
Figure 2. The framework of the proposed method.
Figure 2. The framework of the proposed method.
Entropy 21 00912 g002
Figure 3. The performance comparison of each algorithm (a) comparison with sample 1; (b) comparison with sample 2 and (c) comparison with sample 3.
Figure 3. The performance comparison of each algorithm (a) comparison with sample 1; (b) comparison with sample 2 and (c) comparison with sample 3.
Entropy 21 00912 g003
Figure 4. The comparison on the cost function values of the extreme learning machine by maximum mixture correntropy criterion (MMCC-ELM) and ECC-ELM (a) comparison with sample 1; (b) comparison with sample 2; (c) comparison with sample 3.
Figure 4. The comparison on the cost function values of the extreme learning machine by maximum mixture correntropy criterion (MMCC-ELM) and ECC-ELM (a) comparison with sample 1; (b) comparison with sample 2; (c) comparison with sample 3.
Entropy 21 00912 g004
Figure 5. The dynamic changes of the evolution factor (Ef) and costs during the cooperative evolution.
Figure 5. The dynamic changes of the evolution factor (Ef) and costs during the cooperative evolution.
Entropy 21 00912 g005
Figure 6. Comparison between the estimated pdf and actual pdf.
Figure 6. Comparison between the estimated pdf and actual pdf.
Entropy 21 00912 g006
Figure 7. The comparison between the actual values and the predicted values under the data set, Servo.
Figure 7. The comparison between the actual values and the predicted values under the data set, Servo.
Entropy 21 00912 g007
Figure 8. The evolutionary process of the particles.
Figure 8. The evolutionary process of the particles.
Entropy 21 00912 g008
Figure 9. Dynamic changes on Ef and costs during the cooperative evolution.
Figure 9. Dynamic changes on Ef and costs during the cooperative evolution.
Entropy 21 00912 g009
Figure 10. The comparison with the predicted current transfer ratio (CTR) and actual CTR.
Figure 10. The comparison with the predicted current transfer ratio (CTR) and actual CTR.
Entropy 21 00912 g010
Table 1. The strategies for selecting the parameters.
Table 1. The strategies for selecting the parameters.
StateRange of Efc1c2 p σ g σ τ 1 τ 2
Convergence 0 Ef < 0.25 22 p σ ( k ) g σ ( k ) 00
Exploitation 0.25 Ef < 0.5 2.11.9 p σ ( k τ 1 ( k ) ) g σ ( k ) [ k · r a n d 1 ] 0
Exploration 0.5 Ef < 0.75 2.21.8 p σ ( k ) g σ ( k τ 2 ( k ) ) 0 [ k · r a n d 2 ]
Jumping outEf > 0.751.82.2 p σ ( k τ 1 ( k ) ) g σ ( k τ 2 ( k ) ) [ k · r a n d 1 ] [ k · r a n d 2 ]
Table 2. The assignments of the parameters in each sample.
Table 2. The assignments of the parameters in each sample.
Sample # α β γ μ
Sample 1100.0010
Sample 20.700.00010
Sample 31.200.0010
Table 3. The assignment of the parameters for each algorithms.
Table 3. The assignment of the parameters for each algorithms.
AlgorithmParameterSample 1Sample 2Sample 3
R-ELMN100100100
λ 0.000010.00010.0001
RCC-ELMN100100100
λ 0.000010.000010.00001
Ihq303030
ε0.00010.00010.0001
σ11.21.2
MMCC-ELMN100100100
λ0.000010.000010.00001
Ihq303030
ε0.00010.00010.0001
Σ122.24.3
Σ20.80.88.5
α 0.80.80.9
ECC-ELMN100100100
λ0.000010.000010.00001
Ihq303030
ε0.00010.00010.0001
Table 4. The comparison of the accuracies of the four algorithms.
Table 4. The comparison of the accuracies of the four algorithms.
SamplesELMRCC-ELMMMCC-ELMECCC-ELM
Training MSETesting MSETraining MSETesting MSETraining MSETesting MSETraining MSETesting MSE
Sample 10.3360.66010.13390.35050.72251.10850.14150.3595
Sample 20.08280.110.05070.08921.3632.1890.02570.0576
Sample 30.22190.25720.20760.23390.8680.75830.20460.2237
Table 5. The information on the data sets.
Table 5. The information on the data sets.
Data SetFeaturesObservations
Training NumbersTesting Numbers
Servo58383
Slump105251
Concrete9515515
Housing14253253
Yacht6154154
Airfoil5751751
Soil moisture124340340
HDI129393
HIV106565
Table 6. Parameter settings of each algorithm.
Table 6. Parameter settings of each algorithm.
AlgorithmParameterServoSlumpConcreteHousingYachtAirfoilSoil MoistureHDIHIV
R-ELMN90190185180185200200100100
λ 0.00010000 0.00050000 0.00020000 0.00020000 0.00002000 0.00002000 0.00002000 0.00001000 0.00001000
RCC-ELMN120100200200200180180150120
λ 0.00001000 0.00010000 0.00000100 0.00010000 0.00000001 0.00000001 0.00000001 0.00000001 0.00000001
Ihq303030303030303030
ε0.00010000 0.00010000 0.00010000 0.00010000 0.00010000 0.00010000 0.00010000 0.00010000 0.00010000
σ0.00100000 0.00001000 0.00005000 0.01000000 0.00000100 0.00000100 0.00000100 0.00000100 0.00000130
MMCC-ELMN90165200200195150150150150
λ 0.00100000 0.00001000 0.00005000 0.01000000 0.00000100 0.00000100 0.00000100 0.00000100 0.00000100
Ihq303030303030303030
ε0.00010000 0.00010000 0.00010000 0.00010000 0.00010000 0.00010000 0.00010000 0.00010000 0.00010000
Σ10.20.50.50.50.50.21.01.20.7
Σ22.81.62.6222.70.70.80.3
α 0.80.30.50.80.80.50.60.70.6
ECC-ELMN90180180180180200200200200
λ 0.00100000 0.00001000 0.00005000 0.01000000 0.00000100 0.00000100 0.00000100 0.00000100 0.00000100
Ihq303030303030303030
ε0.00010000 0.00010000 0.00010000 0.00010000 0.00010000 0.00010000 0.00010000 0.00010000 0.00010000
Table 7. The performance comparison.
Table 7. The performance comparison.
Data SetR-ELMRCC-ELMMMCC-ELMECCC-ELM
Training RMSETesting RMSETraining RMSETesting RMSETraining RMSETesting RMSETraining RMSETesting RMSE
Servo0.0590 ± 0.0090.1039 ± 0.01640.0740 ± 0.01060.1031 ± 0.01480.0839 ± 0.01740.0989 ± 0.01870.1047 ± 0.01810.8742 ± 0.0131
Slump0.0081 ± 0.00110.0461 ± 0.00950.0000 ± 0.00000.0422 ± 0.00940.0001 ± 0.00000.0408 ± 0.01010.0001 ± 0.00010.354 ± 0.1890
Concrete0.0738 ± 0.00210.0917 ± 0.00450.0561 ± 0.00180.0872 ± 0.00660.0560 ± 0.00210.0867 ± 0.00640.0561 ± 0.00180.0852 ± 0.0053
Housing0.0439 ± 0.00430.0896 ± 0.01240.0495 ± 0.00450.0830 ± 0.01100.0554 ± 0.00450.0821 ± 0.01010.0352 ± 0.00130.0791 ± 0.0110
Yacht0.0366 ± 0.00930.0529 ± 0.00900.0125 ± 0.00080.0349 ± 0.01130.0125 ± 0.00080.0328 ± 0.00740.0172 ± 0.00270.0268 ± 0.0031
Airfoil0.0974 ± 0.00740.1031 ± 0.00770.0736 ± 0.00220.0906 ± 0.00540.0736 ± 0.00250.0898 ± 0.00510.0736 ± 0.00230.0889 ± 0.0046
Soil moisture0.0032 ± 0.00110.0095 ± 0.00130.0007 ± 0.00010.0015 ± 0.00030.0006 ± 0.00000.0012 ± 0.00020.0006 ± 0.00000.0009 ± 0.0001
HDI0.0004 ± 0.00010.0006 ± 0.00020.0001 ± 0.00000.0003 ± 0.00010.0001 ± 0.00000.0003 ± 0.00010.0001 ± 0.00000.0003 ± 0.0001
HIV0.0376 ± 0.02200.0599 ± 0.01300.0050 ± 0.00170.0079 ± 0.00090.0047 ± 0.00060.0065 ± 0.00040.0059 ± 0.00070.0059 ± 0.0006
Table 8. The performance of the predicted model that is generated using the ECCELM.
Table 8. The performance of the predicted model that is generated using the ECCELM.
Time (year)Actual CTRPredicted CTRNormalized ErrorPredicting Time (ms)
187.90 88.03 0.0037 2.98
287.70 88.01 0.0068 4.02
387.40 87.94 0.0274 3.92
485.50 87.15 0.0095 4.98
586.30 87.02 0.0122 2.26
685.93 86.61 0.0084 3.22
785.86 85.40 0.0188 5.74
884.73 85.30 0.0145 4.48
984.01 84.33 0.0115 5.85
1083.31 83.38 0.0023 4.87

Share and Cite

MDPI and ACS Style

Mei, W.; Liu, Z.; Su, Y.; Du, L.; Huang, J. Evolved-Cooperative Correntropy-Based Extreme Learning Machine for Robust Prediction. Entropy 2019, 21, 912. https://doi.org/10.3390/e21090912

AMA Style

Mei W, Liu Z, Su Y, Du L, Huang J. Evolved-Cooperative Correntropy-Based Extreme Learning Machine for Robust Prediction. Entropy. 2019; 21(9):912. https://doi.org/10.3390/e21090912

Chicago/Turabian Style

Mei, Wenjuan, Zhen Liu, Yuanzhang Su, Li Du, and Jianguo Huang. 2019. "Evolved-Cooperative Correntropy-Based Extreme Learning Machine for Robust Prediction" Entropy 21, no. 9: 912. https://doi.org/10.3390/e21090912

APA Style

Mei, W., Liu, Z., Su, Y., Du, L., & Huang, J. (2019). Evolved-Cooperative Correntropy-Based Extreme Learning Machine for Robust Prediction. Entropy, 21(9), 912. https://doi.org/10.3390/e21090912

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop