1. Introduction
Energy demand has always been one of the main problems of human development since the increasing consumption of energy with the improvement of living standards. In recent years, renewable energy has gradually become a research hotspot. Wind energy is valued for its clean, pollution-free, renewable, and abundant availability. However, wind is highly random and volatile, which may affect the stability of the power system and hinder the efficient use of wind energy [
1]. Accurate ultra-short-term wind speed prediction models are therefore crucial in power dispatch planning and power market operations [
2]. Thus, reliable wind speed prediction has drawn a lot of interest.
The three common wind speed prediction models are physical models, statistical models, and hybrid models. Physical models take into account the physical conditions and locations of wind farms, which require abundant meteorological data. Numerical weather prediction is a typical physical model, as it takes into account temperature pressure and obstacles for wind speed prediction, so it has a long calculation period [
3]. Physical models are efficient and accurate for long-term forecasting, but they are computationally intensive and expensive for small-scale forecasting.
Statistical models make better use of historical wind speed data to predict future wind speeds than physical models. Statistical models include both traditional statistical models and neural network-based models. Traditional statistical models include the autoregressive moving average model [
4], the autoregressive integrated moving average model [
5], the Bayesian model [
6], etc. However, the non-linear nature of wind makes it difficult for traditional statistical models to extract the deeper features of wind speed data. Neural networks are introduced into the field of wind speed prediction for their ability to fit the non-linear part of the data well. Neural network-based models can extract deeper features from wind speed data than traditional statistical models—for example, BP [
7], RBF [
8], artificial neural network [
9], SVR [
10], etc. To improve the learning ability and predictive ability of predictive models, deep neural networks are introduced into wind speed prediction, such as the deep belief network [
11], RNN [
12], GNN [
13], and LSTM [
14].
In recent years, hybrid models have gradually become the mainstream wind speed prediction models. Hybrid models typically use one or more auxiliary strategies to assist the main forecasting network in wind speed prediction. Therefore, hybrid models can achieve better prediction performance than physical models and statistical models. The auxiliary strategies involved in hybrid models include data preprocessing techniques, optimization algorithms, error correction, and weighting strategies.
(1) Data preprocessing techniques. Zhang et al. [
15] used EMD for data pre-processing of wind speed, which effectively reduced the volatility of the wind speed series. However, EMD suffers from the problem of modal confusion, which leads to unsatisfactory decomposition results. Santhosh et al. [
16] used EEMD to process the raw wind speed series, which effectively mitigated the EMD problem. However, EEMD has a noise residual problem affected by noise residuals. Wang et al. [
17] used CEEMD for wind speed prediction. CEEMD cancels out the residual noise with a pair of white noises, effectively improving the efficiency of the calculation. Ren et al. [
18] experimentally demonstrated that the CEEMDAN-based model always performs best compared to the EMD-based model.
(2) Optimization algorithms. Optimization algorithms can be used to optimize the hyperparameters, weights, network structure, and thresholds of predictive models. Li et al. [
19] used PSO to optimize two hyper-parameters of LSTM, which solved the problem of wide intervals caused by interval superposition and thus improved the wind speed prediction accuracy. Tian [
20] used PSO to optimize the weight coefficients of each prediction model, and the experimental results demonstrate the necessity of introducing the weight coefficient optimization strategy. Liu et al. [
21] used GA to optimize the internal parameters of LSTM, which improved the efficiency and accuracy of the prediction model. Cui et al. [
22] used the Bat algorithm to optimize the thresholds of BP networks, effectively improving the generalization ability and nonlinear mapping ability of BP networks.
(3) Error correction. Error correction is a post-processing technique for wind forecasting. It predicts the residuals and superimposes the predictions on the original predictions to obtain the final predictions. Duan et al. [
23] used improved CEEMDAN to decompose the errors, and the experimental results showed that the error decomposition correction method can significantly enhance the prediction accuracy. Liu et al. [
24] proposed an adaptive multiple error correction method, which makes full use of the deeper predictable components and effectively improves the reliability and accuracy of the model. Zhang et al. [
25] demonstrated experimentally that the final predicted values after Markov chain correction are closer to the original wind field data, which proves that the use of the Markov chain is effective.
(4) Weighting strategies. To scientifically determine the weights of different prediction networks in a hybrid model, many scholars have proposed different weighting strategies. To alleviate the adverse effects of multi-collinearity in combinatorial prediction models, Jiang et al. [
26] used a GMDH neural network to automatically identify the weights of three nonlinear models. The application of GMDH can significantly improve the predictive capability compared to the widely used equal-weighting scheme. Wang et al. [
27] used MTO to minimize the error sum of squares of the IOWA operator, which obtains the optimal weight vector for the combined prediction model and ensures the stability of the prediction results. Altan et al. [
28] optimized the weighting coefficients for each IMF using the gray wolf optimizer algorithm.
Although the above models achieve good predictive performance, they still have some problems. Methods involving deep neural networks [
27] cause huge computational costs. Hybrid methods based on weighting strategy [
28] may have the problem of multicollinearity, which reduces the prediction accuracy. The performance of hybrid methods based on parameter optimization [
26] is largely influenced by the understanding of the researcher of the optimization algorithm.
Considering the above issues, we propose a hybrid model combining data preprocessing techniques and optimization algorithms for ultra-short-term wind speed prediction. We design the hybrid generative adversarial network (HGANN) as the prediction master network for the proposed hybrid model. The contributions and innovations of this research are concluded as follows:
- (1)
A hybrid generative adversarial network model (HGANN) is proposed for ultra-short-term wind speed prediction, which learns the distribution of wind data and predicts it through a continuous game between generators and discriminators.
- (2)
To improve the error convergence of the model, the OBLS was developed as a generator for HGANN. The IPSO was used to optimize the hyperparameters of the OBLS. To maintain the stability of the generated samples, we used the discriminator of WGAN as the discriminator of HGANN.
- (3)
A wind data decomposition and denoising process was carried out using CEEMDAN to reduce the randomness and instability in the original wind series.
The rest of this article is organized as follows.
Section 2 introduces the model framework and methods involved in this article in detail. In
Section 3, the experimental cases and prediction results are elaborated in detail, which verifies the validity of the framework we propose.
Section 4 contains a discussion of the results of the experiment. The conclusions are presented in
Section 5.
2. Proposed Predictive Framework
2.1. Overall Framework of HGANN
Generative adversarial networks (GANs) [
29] are deep learning networks, which are composed of a generator and discriminator that confront each other. The role of the generator is to generate false samples that are close to the real ones. The role of the discriminator is to distinguish between true and false samples as correctly as possible. However, GANs often suffer from the problem of target confusion. Our proposed HGANN alleviates this problem to a great extent.
We developed a hybrid generative adversarial network model (HGANN) for ultra-short-term wind speed prediction, which uses the two networks to compete with each other to achieve highly accurate wind speed predictions. The proposed model is shown in
Figure 1. First, CEEMDAN decomposes the raw wind speed data into multiple modalities. These modalities are separately fed into the generator of HGANN, the OBLS. The generator is used to obtain virtual samples that are similar to real samples. The virtual samples and real samples are then fed into the discriminator, which consists of convolutional layers and fully connected layers. The discriminator extracts the high-dimensional features of the input samples through the convolutional layer, and then further extracts the effective features by the fully connected layer. The outputs scalars “1” or “0” of the discriminator are passed to the generator and the discriminator to perform the iterative update of HGANN. Via the continuous iterative update, OBLS obtains the best parameters and performs wind speed prediction. Finally, the final wind speed forecast can be obtained by stacking all forecast values.
2.2. CEEMDAN Model
Due to the high volatility of the wind speed series, CEEMDAN [
30] is introduced to smooth the wind speed data. CEEMDAN decomposes a signal into some modalities.
The original wind speed series is defined as
. CEEMDAN decomposes
into
and residue
.
Figure 2 shows the flow chart of the CEEMDAN algorithm. The specific steps of the algorithm are as follows.
Randomly generate white noise with (0, 1), which is defined as . Define an operator which generates the IMFs by EMD. We set the noise standard deviation to = 0.2 and the ensemble size to = 500.
Add
to the
and generate a new series with noisy signals
. For
j, the first-order
that is decomposed by EMD is expressed as:
The first-order residue is computed as follows:
For
, calculate the
and the
jth residue as follows:
Decompose until the residue cannot be decomposed and has only one extreme value. Then we can get and the final residue .
The original wind speed time series can be decomposed as , where or can represent different features of the wind speed.
2.3. Generator OBLS for HGANN
BLS [
31] can provide incremental structural learning. It achieves better forecasting results in time-series forecasting. Furthermore, because of its shallow network structure, BLS has higher error convergence performance than CNN. Compared with BLS, OBLS can provide both higher convergence performance and predictive accuracy. This is because OBLS uses IPSO to improve the network hyper-parameter of optimization. Therefore, OBLS has faster convergence and higher error convergence than CNN. Therefore, instead of using CNN as the generator of GAN, we use OBLS as the generator of HGANN to solve the problem of target confusion during HGANN training, which can improve the generalization ability and error convergence of HGANN, and thus make HGANN more suitable for wind speed prediction. The following is the detailed process of the OBLS algorithm.
Randomly generate n particles so that the dimensions of the particles are a three-dimensional vector corresponding to the three parameters of BLS, respectively. Initialize the particle position and speed . Determine the learning factors and , inertia weights and , and the maximum number of iterations .
Assume the input wind speed series data
and project the data using
to represent
th mapped feature
, where
represents random weight with the proper dimensions. The
th group of enhancement nodes
is denoted as
.
and
can be different functions. The
th mappings can be denoted as:
The feature nodes are denoted as
, where
and
are random weights. The enhanced nodes are denoted as:
Let
where the symbol
means “noted as”; then the output of the BLS can be denoted as:
where the
is the final target weight needed by OBLS and is obtained through the ridge regression algorithm, that is,
.
Let
; then
can be expressed as follows:
where λ is
regularization.
The IPSO [
32] is introduced to iterate to optimize the parameters of BLS:
. When the iteration of IPSO is consistently performed, the position and speed of the particles are continually updated through the following equation:
Here,
is the velocity coefficient; the value of the inertia weight
is
. When reaching the maximum iterative number
, the iteration is stopped and the optimal value of
can be obtained.
The generator takes the wind speed subsequence as input, which is generated by CEEMDAN. Then the generator generates a new wind speed sequence , which is statistically similar to the wind speed sequence .
From Equations (8)–(10), OBLS does not require layer-to-layer coupling. Since there are no multi-layer connections, OBLS does not need to use gradient descent to update the weights, so the computational cost of OBLS is significantly lower than that of deep learning. When the accuracy of OBLS does not meet the requirements, its accuracy can be improved by increasing the “width” of the network nodes. Compared with the increase in the amount of calculation by increasing the number of layers in the deep network, the increase in that by increasing the “width” of the network nodes in OBLS is negligible.
2.4. Discriminators for HGANN
To maintain the stability of the generated samples, we used the discriminator of WGAN [
33] as the discriminator of HGANN. In HGANN, the discriminator takes
or
as input. The training goal of discriminator is to discriminate
as false and
as true. The discriminator is trained by minimizing the distance function (
) (loss function), which is defined as follows:
here,
represents the output of the D;
is the binary cross entropy, defined as:
Based on this loss function, the discriminator can achieve an output of 1 when the input is and an output of 0 when the input is , and then discriminates the wind speed sequence .
The discriminator outputs a scalar of “0” or “1.” The scalar of “0” or “1” has two purposes: (1) It can influence and then adjust the weights of the neural network in the discriminator and maximize Equation (14) through a backpropagation algorithm. (2) It can be passed to the generator to assist the PSO algorithm to find the optimal hyper-parameters of the OBLS and then calculate the value of the fitness function
, which is defined as follows:
where
represents the output of the generator.
2.5. Prediction Steps of the Proposed HGANN Model
We propose the HGANN model for ultra-short-term wind speed prediction. The flow chart of the prediction process of the proposed model is shown in
Figure 3. CEEMDAN is used to decompose the raw wind speed data
into multiple modes
. These
are separately sent into the generator (OBLS) of HGANN to obtain virtual samples
. The discriminator (WGAN) takes
or
as input and then outputs scalars “1” or “0.” The scalars “1” or “0” are passed to the generator (OBLS) and the discriminator (WGAN) to participate in iterative model updates. Through the continuous iterative update, OBLS obtains the optimal value of
. The final wind speed forecasting values
can be obtained by stacking all forecast values.
4. Discussion
Our model was compared with five advanced models to evaluate its performance and advantages in various wind sequence experiments. Experimental results show that the proposed model had better predictive performance. The reasons behind this fact are given as follows.
First, the wind speed data were one-year data from wind farms in Germany and China, which cover complex fluctuation characteristics. Therefore, our HGANN model uses CEEMDAN to smoothen the volatility of the data and improve the predictive performance.
Second, HGANN uses OBLS as the generator to provide a special shallow broad incremental learning network structure, which can not only be beneficial for improving prediction accuracy for one-dimensional wind speed prediction compared to CNN but also greatly decrease computational cost using pseudo-inverse operations to determine the network weights instead of using convolution operations.
Third, in our HGANN model, the proposed OBLS uses an improved PSO to optimize the hyper-parameters of its network, which can search in a wider range and obtain the optimal parameters over BLS. Therefore, OBLS has better generalization ability than BLS.
Finally, HGANN can better extract the deeper features of wind speed data by playing a minimum–maximum game between the generator and discriminator for wind speed prediction.
5. Conclusions
Although existing various hybrid predictive models have provided competitive performance in ultra-short-term wind speed prediction, they still need to be further improved—for instance, how to effectively reduce the computational cost of hybrid predictive models, and how to effectively deal with the multicollinearity problem of the hybrid forecasting model based on weighted strategy, which leads to the problem of reduced forecasting accuracy. To enhance the predictive power and decrease the computational cost, this paper proposes the HGANN model for ultra-short-term wind speed forecasting. HGANN is a generative adversarial network in which the generator and discriminator play against each other to obtain wind speed predictions with high accuracy. In HGANN, we developed OBLS and the convolutional structures as the generator and the discriminator, respectively, which enables them to obtain effective synergies to improve predictive performance. Particularly, OBLS involves a special shallow broad incremental learning network structure, which can effectively deal with one-dimensional wind speed data. Furthermore, the shallow network structure of OBLS can also significantly decrease computational cost via using pseudo-inverse operations rather than convolution operations. In addition, the proposed OBLS applies an improved PSO to obtain the optimal network hyper-parameters. CEEMDAN performs noise reduction and decomposition of the wind data. Via the above rational combination, the proposed HGANN provides high predictive accuracy and generalization ability with low computational cost in ultra-short-term wind speed prediction. The experimental results indicate the above fact. For instance, the RMSE predictive errors of the proposed model were 29.35%, 49.22%, 38.09%, and 30.10% compared to the four state-of-art predictive models PSO-ANFIS, VMD-GA-BP, EEMD-GPR-LSTM, and MWS-CE-ENN on the spring wind data of the HER dataset, respectively. In the future, we plan to use parallel computing to speed up the process of PSO optimization of BLS during training. Furthermore, the proposed HGANN will be extended to a wider range of applications, such as financial time-series forecasting, electricity-load forecasting, traffic forecasting, etc.