Next Article in Journal
Application of Dissolved Gas Analysis in Assessing Degree of Healthiness or Faultiness with Fault Identification in Oil-Immersed Equipment
Next Article in Special Issue
Oriented Fermentation of Food Waste towards High-Value Products: A Review
Previous Article in Journal
A Comparison of Dispatchable RES Technoeconomics: Is There a Niche for Concentrated Solar Power?
Previous Article in Special Issue
Soil Bioremediation: Overview of Technologies and Trends
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Neural-Based Ensembles and Unorganized Machines to Predict Streamflow Series from Hydroelectric Plants

by
Jônatas Belotti
1,2,
Hugo Siqueira
1,
Lilian Araujo
1,
Sérgio L. Stevan, Jr.
1,
Paulo S.G. de Mattos Neto
3,
Manoel H. N. Marinho
4,
João Fausto L. de Oliveira
4,
Fábio Usberti
2,
Marcos de Almeida Leone Filho
5,
Attilio Converti
6 and
Leonie Asfora Sarubbo
7,8,*
1
Graduate Program in Computer Sciences, Federal University of Technology–Parana (UTFPR), Ponta Grossa 84017-220, Brazil
2
Institute of Computing, State University of Campinas (UNICAMP), Campinas 13083-852, Brazil
3
Departamento de Sistemas de Computação, Centro de Informática, Universidade Federal de Pernambuco, (UFPE), Recife 50740-560, Brazil
4
Polytechnic School of Pernambuco, University of Pernambuco, Recife 50100-010, Brazil
5
Venidera Pesquisa e Desenvolvimento, Campinas 13070-173, Brazil
6
Department of Civil, Chemical and Environmental Engineering, University of Genoa (UNIGE), 16126 Genoa, Italy
7
Department of Biotechnology, Catholic University of Pernambuco (UNICAP), Recife 50050-900, Brazil
8
Advanced Institute of Technology and Innovation (IATI), Recife 50751-310, Brazil
*
Author to whom correspondence should be addressed.
Energies 2020, 13(18), 4769; https://doi.org/10.3390/en13184769
Submission received: 2 July 2020 / Revised: 1 September 2020 / Accepted: 4 September 2020 / Published: 12 September 2020
(This article belongs to the Special Issue Environmental and Energetic Valorization of Renewable Resources)

Abstract

:
Estimating future streamflows is a key step in producing electricity for countries with hydroelectric plants. Accurate predictions are particularly important due to environmental and economic impact they lead. In order to analyze the forecasting capability of models regarding monthly seasonal streamflow series, we realized an extensive investigation considering: six versions of unorganized machines—extreme learning machines (ELM) with and without regularization coefficient (RC), and echo state network (ESN) using the reservoirs from Jaeger’s and Ozturk et al., with and without RC. Additionally, we addressed the ELM as the combiner of a neural-based ensemble, an investigation not yet accomplished in such context. A comparative analysis was performed utilizing two linear approaches (autoregressive model (AR) and autoregressive and moving average model (ARMA)), four artificial neural networks (multilayer perceptron, radial basis function, Elman network, and Jordan network), and four ensembles. The tests were conducted at five hydroelectric plants, using horizons of 1, 3, 6, and 12 steps ahead. The results indicated that the unorganized machines and the ELM ensembles performed better than the linear models in all simulations. Moreover, the errors showed that the unorganized machines and the ELM-based ensembles reached the best general performances.

Graphical Abstract

1. Introduction

Planning the operation of a power generation system is defined by establishing the use of energy sources in the most efficient way [1,2,3]. Renewable sources are those with the lowest operation cost since the fuel is provided free of charge by nature. Good predictions of river streamflows allow resource management according to their future availability [4]. Therefore, this is mandatory for countries where there are hydroelectric plants [5,6,7].
The International Hydropower Association published the Hydropower Status Report 2020 [8], showing that 4306 TWh of electricity was generated in the world using hydroelectric plants in 2019. This amount represents the single most significant contribution from a renewable energy source in history. The document summarizes data from 13,000 stations in 150 countries. The top countries in hydropower installed capacity are China (356.40 GW), Brazil (109.06 GW), United States (102.75 GW), and Canada (81.39 GW).
In this context, it is important to predict accurate information about rivers’ monthly seasonal streamflow, since it makes the turbines spin, transforming kinetic into electric energy [5,9]. These series present a specific seasonal behavior due to the volume of water throughout the year being mostly dependent on rainfall [10,11]. Ensuring efficient operation of such plants is needed, since it significantly impacts cost of production and suitable use of water [12,13]. Additionally, their operation leads to a smaller environmental impact than burning carboniferous fuel. Due to this, many pieces of research have presented investigations on such fields for countries such as China [14], Canada [15], Serbia [16], Norway [17], Malaysia [7], and Brazil [9].
Linear and nonlinear methodologies have been proposed to solve this problem. As discussed in [5,12], and [18], the linear methods of the Box-Jenkins family are widely used [19]. The autoregressive model (AR) is highlighted because its easy implementation process allows the calculation of its free coefficients in a simple and deterministic manner. An extended proposal for this task is the autoregressive and moving average model (ARMA), a more general methodology that uses the errors of past predictions to form the output response [19,20].
However, artificial neural networks (ANN) are prominent for this kind of problem [9,21,22,23,24]. They were inspired by the operation of the nervous system of superior organisms, recognizing data regularities and patterns through training and determining generalizations based on the acquired knowledge [18,25,26,27].
In recent times, some studies have indicated that the best results for time series forecasting can be achieved by combining different predictors using ensembles [28,29,30]. Many authors have applied these techniques to similar tasks [31,32,33]. However, the approaches commonly explore only the average of the single models output or the classic neural networks (multilayer perceptron (MLP) and radial basis function networks (RBF)) as a combiner. The specialized literature regarding streamflow forecasting shows that ANN approaches stand out, but some authors use linear models [16], support vector regression [14], and ensembles [15].
This work proposes using a special class of neural networks, the unorganized machines (UM), to solve the aforementioned forecasting task. The term UM defines the extreme learning machines (ELM) and echo state network (ESN) collectively. In this investigation, we addressed six versions of UMs: ELM with and without regularization coefficient (RC) as well as ESN using the reservoir designs from Jaeger’s and Ozturk et al. with and without RC. Additionally, we addressed the ELM and the ELM (RC) as the combiner of a neural-based ensemble.
To realize an extensive comparative study, we addressed two linear models (AR and ARMA models); four well-known artificial neural networks (MLP, RBF, Elman network, Jordan Network); and four other ensembles, using as combiners the average, the median, the MLP, and the RBF. To the best of our knowledge, the use of ELM ensembles in this problem and similar repertoires of models is an investigation not yet accomplished. Therefore, we would like to fill this gap.
In this study, the database is from Brazil. In the country, electric energy is mostly generated by hydroelectric plants, these being responsible for 60% of all electric power produced in 2018 [8,34,35]. In addition, Brazil is the one of the largest producers of hydropower in the world. Therefore, the results achieved can be extensible for other countries.
The remainder of this work is organized as follows: Section 2 discusses the linear models from the Box-Jenkins methodology; Section 3 presents the artificial neural networks and the ensembles; Section 4 shows the case study, the details on the seasonal streamflow series, the computational results, and the result analysis; and Section 5 shows the conclusions.

2. Linear Forecasting Models

The definition of linear prediction models by Box-Jenkins makes use of linear filtering concepts [19]. The x t element of a time series results from a linear filter Ψ on a Gaussian white noise a t . Another form to represent a linear model is by weighing the previous signals ( x t 1 , x t 2 , , x 1 ) for the next forecast element ( x t ). To do this, we add a noise a t and the mean of the series μ :
x t = μ + a t + π 1 x t 1 + π 2 x t 2 + + π t x 1
where a t is the noise of t -th term, and π n is the weight assigned to the ( t n ) -th term of series.

2.1. Autoregressive Model

Given any value x t of a time series, the delay p is defined with x t p . An autoregressive process of order p (AR ( p ) ) is defined as the linear combination of p delays of observation x t , with the addition of a white Gaussian noise a t , as shown in Equation (2) [19]:
x ˜ t = ϕ 1 x ˜ t 1 + ϕ 2 x ˜ t 2 + + ϕ p x ˜ t p + a t
where x ˜ t = x t μ , ϕ p is the weighting coefficient for the delay p .
The term a t is considered as the inherent error of the regression process. This is the error of the forecast when the model is used to predict future values. Thus, the optimum ϕ p coefficients must be calculated to minimize the error a t [36].
To determine the optimum values of ϕ p it is necessary to solve a recurrence relation that emerges from its autocorrelation function, as presented in Equation (3):
ρ j = ϕ 1 ρ j 1 + ϕ 2 ρ j 2 + + ϕ p ρ j p ,     j > 0
If we expand this relation to j = 1 , 2 , , p we obtain the set of linear equations denominated Yule–Walker equations, which define ϕ 1 , ϕ 2 , , ϕ p as a function of ρ 1 , ρ 2 , , ρ p for a model AR(p), as in Equation (4) [19]:
ρ 1 = ϕ 1 ρ 0 + ϕ 2 ρ 1 + + ϕ p ρ p 1 ρ 2 = ϕ 1 ρ 1 + ϕ 2 ρ 0 + + ϕ p ρ p 2 ρ 3 = ϕ 1 ρ 2 + ϕ 2 ρ 1 + + ϕ p ρ p 3 ρ p = ϕ 1 ρ p 1 + ϕ 2 ρ p 2 + + ϕ p ρ 0

2.2. Autoregressive and Moving Average Model

Unlike the autoregressive, in the moving average model (MA), white noise signals are combined [19]. A model MA is said to be of order q if the prediction of x t utilizes q samples of white noise signals, as in Equation (5):
x t = θ 1 a t 1 θ 2 a t 2 θ q a t q + a t
where θ t , t 1 , 2 , , q are the parameters of model.
An ARMA model is the union of AR and MA. To predict using the ARMA of order p , q , it is necessary to address p prior signals (AR) and q white noise signals (MA). Mathematically, an ARMA ( p , q ) model is described as in Equation (6):
x t = ϕ 1 x ˜ t 1 + ϕ 2 x ˜ t 2 + + ϕ p x ˜ t p θ 1 a t 1 θ 2 a t 2 θ q a t q + a t
where ϕ t and θ t are the model parameters.
Unlike the AR, the calculation of the ARMA coefficient is done by solving nonlinear equations. However, it is possible to achieve an optimal linear predictor if the choice of these coefficients is adequate [18,19].

3. Artificial Neural Network

Artificial neural networks (ANN) are distributed and parallel systems composed of simple data processing units. These units are denominated artificial neurons and are capable of computing mathematical functions, which, in most cases, are nonlinear [37,38]. Artificial neurons are connected by normally unidirectional connections and can be arranged in one or more layers [25].
The ANNs present a learning ability through the application of a training method. They can generalize the knowledge acquired through the solution of problem instances for which no answer is known [39]. Neural networks are widely used in many areas of science, engineering computing, medicine, and others [9,25,40,41,42,43,44,45].

3.1. Multilayer Perceptron (MLP)

The multilayer perceptron (MLP) consists of a set of artificial neurons arranged in multiple layers so that the input signal propagates through the network layer by layer [25]. It is considered one of the most versatile architectures for applicability and is used in the universal approximation of functions, pattern recognition, process identification and control, time series forecasting, and system optimization [25,29].
The training process consists of adjusting the synaptic weights of the artificial neuron to find the set that achieves the best mapping of the desired event [46,47]. The most known training method for MLP is the steepest descent in which the gradient vector is calculated using the backpropagation algorithm [48,49].
The error signal of a neuron j in iteration t is given by Equation (7):
e j ( t ) = d j ( t ) y j ( t )
where e j ( t ) is the error, d j ( t ) is the expected result (desired output), and y j ( t ) is the output of the network.
Finally, the rule for updating the synaptic weights of each neuron is done using Equation (8):
w i j m ( t + 1 ) = w i j m ( t ) α E ( t ) w i j m ( t )
where w i j m ( t ) is the synaptic input weight i of the neuron j of the layer m in iteration t , and E ( t ) is the partial derivative of the error.
The training algorithm consists of two phases. Initially, the input data are propagated by the network to obtain its outputs. These values are then compared with the desired ones to obtain the error. In the second step, the opposite path is performed from the output layer to the input layer. In this case, all the synaptic weights are adjusted according to the rule of error correction assumed, so that the output given by the network in the following iteration is closer to the expected one [25].

3.2. Radial Basis Function Network (RBF)

Radial basis function networks (RBF), unlike MLPs, have only two layers, one hidden and one output layer. In the first, all the kernel (activation) functions are radial-based [25]. One of the functions most used is Gaussian function expressed in Equation (9):
φ ( u ) = e ( u c ) 2 2 σ 2
in which c is the center of Gaussian and σ 2 its variance as a function of the center.
The training of RBFs is performed in two stages. First, the weights of the intermediate layer are calculated, and the center is adjusted to the value of the base variance of each function. Subsequently, the weights of the output layer are tuned in a supervised process similar of the MLP [25,50].

3.3. Elman and Jordan Networks

The Elman Network is a recursive neural architecture created by Elman [51] based on an MLP. The author divided the input layer into two parts: the first comprises the network inputs and the second, denominated context unit, consists of the outputs of the hidden layer. An Elman network is shown in Figure 1.
As the context units of an Elman network are treated as inputs, they also have network-associated synaptic weights and can be adjusted by the backpropagation through time algorithm. In this work, we used the truncated backpropagation through time version for one delay [25].
Jordan [52] created the first recurrent neu6ral network based on similar premises. This neural network was initially used for time series recognition but is currently applied to all kinds of problems. Here, the context units are fed by the outputs of the output layer neurons instead of the hidden layer. Figure 2 illustrates this model.
As in the Elman network, the context units are treated as network inputs, also having associated synaptic weights, which allows the use of truncated backpropagation through time [25].

3.4. Extreme Learning Machines (ELM)

Extreme learning machines (ELM), introduced by Huang et al. [53], are an architecture of feedforward neural network with only a hidden layer. The main difference between them and the traditional MLP is that the synaptic weights of the hidden layer are chosen randomly and remain untuned during the training process. For this reason, the ELM and the ESN are classified as unorganized machines.
Adjusting an ELM consists of determining the matrix with the synaptic weights of the output layer W o u t that generates the smallest error for the desired output vector d , which can be done through an analytic solution. This process is summarized in utilizing the Moore-Penrose pseudo-inverse operator, which ensures the minimum mean square error and confers to the ELM a fast training process. This solution is demonstrated in Equation (10):
W o u t = ( X h i d T X h i d ) 1 X h i d T d
where X h i d | x | × m is the matrix with all the outputs of the hidden layer for the training set, and m is the number of neurons in the output layer.
This operator ensures that the ELM training is much more computationally efficient than the application of the backpropagation. However, network performance can be improved by inserting a regularization coefficient C , as in Equation (11) [54]:
W o u t = ( 1 C + X h i d T X h i d ) 1 X h i d T d
where C = 2 λ , where λ { 25 , 24 , , 25 , 26 } .
To determine the best value of C , it is necessary to test all 52 possible values [55].

3.5. Echo State Networks (ESN)

Jaeger proposed echo state networks (ESN) as a new type of recurrent neural network. Recursive networks allow different outputs for the same input since it depends on the internal state of the network. The idea of using the term echo is based on the perception that the most recent samples and the previous states influence more strongly the output [56]. The theoretical proof of the existence of an echo state is denominated echo state propriety [9]
In the original proposal from Jaeger, the ESN presents three layers. The hidden layer is denominated dynamic reservoir and consists of fully interconnected neurons, which generate a nonlinear characteristic. The output layer is responsible for combining the outputs of the dynamic reservoir. The subsequent layer, in turn, corresponds to the linear portion of the network. Unlike other RNNs, which can present feedback in any layer, the original proposal only presents feedback loops in the dynamic reservoir. Figure 3 shows a generic ESN.
For each new input on time t + 1 , the states of the network are updated according to Equation (12):
x t + 1 = f ( W i n u t + 1 + W x t )
in which x t + 1 are the states in the input t + 1 , f ( ) represents the activations of reservoir neurons f ( ) = ( f 1 ( ) , f 2 ( ) , f 3 ( ) , , f N ( ) ) , W i n are the coefficients of the input layer, and u t is the input vector at time t .
In turn, the network output vector y t + 1 is given by Equation (13):
y t + 1 = W o u t x t + 1
where W o u t L × N is the matrix with the synaptic weights of the output layer, and L is the number of network outputs.
As occurs in the ELM, the sy6naptic weights of the ESN dynamic reservoir are not adjusted during training. The Moore-Penrose pseudo-inverse operator is also used to determine the weights of W o u t . Additionally, the performance can be improved by using the regularization coefficient.
In this work, we consider two forms of creating the dynamic reservoir. Jaeger et al. [56] proposed the weight matrix by setting three possible values, which are randomly chosen according to the probabilities described in Equation (14):
W k i i n = { 0.4   with   a   probability   of   0.025 0.4   with   a   probability   of   0.025 0   with   a   probability   of   0.95
Ozturk et al. [57] elaborated a reservoir rich in mean entropy of the echo states. The eigenvalues respect a uniform distribution in the unit circle, creating a canonical matrix as presented in Equation (15):
W k i i n = [ 0 0 0 0 r N 1 0 0 0 0 0 1 0 0 0 0 0 1 0 0 0 0 0 1 0 ]
in which r is the spectral radius, set in the range [ 0 ,   1 ] , and N is the number of neurons present in the dynamics reservoir.

3.6. Ensemble Methodology

An ensemble combines the results of several individually adjusted models with the aim of improving the final response of the system [58]. The idea behind this methodology is that different methods, such as neural networks, produce different behaviors when the same inputs are applied. Therefore, a methodology can present better responses for a given range of data, while another works better in another band. A combination method (average, voting, or another neural network) is applied to produce the final ensemble output [29,30]. Figure 4 presents an example of this model.
The combination of several specialists through an ensemble does not exclude predictors’ need to show good individual performance. The purpose of an ensemble is to improve upon existing good results. Therefore, the essential condition for its accuracy is that its models be accurate and diverse [59]. Over time, ensembles have been used to solve many problems [60,61,62,63].
In this work, we used some distinct combiners. First, we addressed some non-trainable methods: the mean and the median of the outputs [29]. Additionally, we applied feedforward neural models: MLP, RBF, and ELM with and without the regularization coefficient [29,30,64]. We highlight that these methodologies are not used often, especially for seasonal streamflow series forecasting.

4. Case Study

Streamflow series are a kind of time series in which each observation refers to monthly, weekly, daily, or hourly average flow. This work addresses the monthly average flow. These series present seasonality since they follow the rain cycles that occur during the year [9]. Seasonality changes the standard behavior of the series, which must be adjusted to improve the response of predictors [5].
Deseasonalization is an operation that removes the seasonal component of the monthly series. They become stationary, presenting zero mean and unitary standard deviation. The new deseasonalized series is given by Equation (16) [50]:
z i , m = x i , m μ ^ m σ ^ m
in which z i , m is the new standardized value of the i element of the series, x i , m is the value of the element i in the original series, μ ^ m is the average of all elements of the series in the month m , and σ ^ m is the standard deviation of all elements of the series in the month m .
This investigation addressed the series of the following Brazilian hydroelectric plants: Agua Vermelha, Belo Monte, Ilha Solteira, Paulo Afonso, and Tucuru. All data were obtained from the National Electric System Operator (ONS) and are available on its website [65]. These plants were selected because of their location in different regions of the country, and, as shown in Table 1, they have different hydrological behavior. Therefore, it is possible to accomplish a robust performance analysis.
All series comprise samples from January of 1931 to December of 2015, a total of 85 years, or 1020 months. The data were divided into three sets: from 1931 to 1995 is the training set, used to adjust the free parameters of the models; from 1996 to 2005 is the validation set, used in the cross-validation process and to adjust the value of the regularization coefficient of ELMs and ESNs; and from 2006 to 2015 is the test set, utilized to measure the performance of the models. The mean squared error (MSE) was adopted as the performance metric, as done in other works of the area [12].
Predictions were made using up to the last six delays of the samples as inputs. These delays were selected using the wrapper method [66,67]. The predictions were performed for 1 (next month), 3 (next season), 6 (next semester), and 12 (next year) steps ahead, using the recursive prediction technique [68,69].
During the preprocessing stage, the series’ seasonal component was removed through deseasonalization to make the behavior almost stationary. Therefore, the predicted values required an additional postprocessing step, where the data had the seasonal component reinserted. Figure 5 shows step by step the entire prediction process.
In total, this work applied 18 predictive methods, two of which were linear methods of the Box-Jenkins family, 10 artificial neural networks, and six ensembles:
  • AR( q )—autoregressive model of order q , optimized by the Yule–Walker equations;
  • ARMA( p , q )—autoregressive–moving-average model of order p , q , optimized by maximum likelihood estimators;
  • MLP—MLP network with a hidden layer;
  • RBF—RBF network;
  • ELM—ELM network;
  • ELM (RC)—ELM with the regularization coefficient;
  • Elman—Elman network;
  • Jordan—Jordan network;
  • Jaeger ESN—ESN with reservoir from Jaeger;
  • Jaeger ESN (RC)—ESN from Jaeger and regularization coefficient;
  • Ozturk ESN—ESN with reservoir from Ozturk et al.;
  • Ozturk ESN (RC)—ESN with reservoir from Ozturk et al. and regularization coefficient;
  • Average Ensemble—ensemble with mean as arithmetic combiner;
  • Median Ensemble—ensemble with median as arithmetic combiner;
  • MLP Ensemble—ensemble with MLP as combiner;
  • RBF Ensemble—ensemble with an RBF as combiner;
  • ELM Ensemble—ensemble with ELM as combiner;
  • ELM Ensemble (RC)—ensemble with an ELM with regularization coefficient as combiner.
All neural networks used the hyperbolic tangent as activation function with β = 1 . The learning rate adopted for the models, which use backpropagation, was 0.1 . As a stopping criterion, a minimum improvement in MSE of 10 6 or a maximum of 2000 epochs was considered. The networks were tested for the number of neurons from 5 to 200, with an increase of 5 neurons. All these parameter’s values were determined after empirical tests.
The regularization coefficients were evaluated with all the 52 possibilities mentioned in Section 3.4. The one with the lowest MSE in the validation set was chosen for ELM and ESN approaches. The wrapper method was used to select the best lags to the single models as well as which experts were the best predictors when using the ensembles. The holdout cross-validation was also applied to all fully neural networks and ensembles to avoid overtraining and to determine the C of ELM and ESN.
All proposed models were executed 30 times to obtain sample output data for each input configuration and number of neurons, these having been chosen as the best executions. Additionally, following the methodology addressed in [9] and [70], we adjusted 12 independent models, one for each month, for all proposed methods. It is allowed since the mean and the variance of each month are distinct, and this approach can lead to better results [9].

Experimental Results

In this section, we present and discuss the results obtained by all models and four forecasting horizons. Table 2, Table 3, Table 4, Table 5 and Table 6 show the results achieved for Agua Vermelha, Belo Monte, Ilha Solteira, Paulo Afonso, and Tucurui hydroelectric plants, for both real and deseasonalized domains. The best performances are highlighted in bold.
The Friedman [71,72] test was applied to the results for the 30 runs of each predictive model proposed, regarding the MSE in the test set. The p-values achieved were smaller than 0.05. Therefore, we can assume that changing the predictor leads to a significant change in the results.
To analyze the dispersion of the results obtained after 30 executions [73], Figure 6 presents the boxplot graphic [72,74].
Note that AR, ARMA, average ensemble, and median ensemble models did not show dispersion since they presented close-form solution [75]. One can verify the highest dispersion was obtained by RBF, while the lowest was achieved by Elman network.
Many aspects of the general results presented in Table 2, Table 3, Table 4, Table 5 and Table 6 can be discussed. The first is that the best predictor in the real space sometimes was not the same in the deseasonalized domain. This occurred because the deseasonalization process considered all parts of the series as having the same importance. In the literature, the error in the real space is adopted as the most important measure to evaluate the results [12,18].
As the forecasting horizon grows, the prediction process becomes more difficult, and the errors tend to increase for all models. It is directly related to the decreased correlation between the input samples and the desired future response. Therefore, the output estimates tend to achieve the long term mean, or the historical mean [9].
We elaborated Table 7 using the results from Table 2, Table 3, Table 4, Table 5 and Table 6 to show a winner ranking to illustrate which models achieved the best general performance.
For P = 1 step ahead, for all series, the best predictor was always an ensemble, highlighting the ELM-based combiner, which was the best for three of five scenarios (60%). This result indicates that the use of ensembles can lead to an increase in the performance. Moreover, the application of an unorganized machine requires less computational effort than the MLP or the RBF since its training process is based on a deterministic linear approach. It also corroborates that their approximation capability is elevated, overcoming their fully trained counterpart [9,12].
The results varied for P = 3 , for which several architectures—ELM, Jaeger ESN, average ensemble, and MLP ensemble—were the best at least once. We emphasize that the ELM network was better in two cases, and the ESN in one. The UMs were better in 60% of the cases. Regarding P = 6 , the ELM was also the best predictor, achieving the smallest error in four cases (80%), followed by the MLP, which was the best only for Tucurui.
Analyzing the last forecasting horizon, P = 12 , four different neural architectures reached the best performance at least once: MLP, Elman, Jaeger ESN (twice), and Ozturk ESN (RC). An important observation is the presence of recurrent models among them. This horizon is very difficult to predict since the correlation between the input samples and the desired response is small. Therefore, there is an indication that the existence of model’s internal memory is an advantage.
In summary, the unorganized networks (ESN and ELM), in stand-alone versions or as a combiner of an ensemble, provided the best results in 14 of 20 scenarios (70%). This is relevant since such methods are newer than the others and are simpler to implement.
Considering the reservoir design of ESNs, we achieved almost a draw; in nine cases, the proposal from Ozturk et al. was the best, and in 11, the Jaeger model achieved the smallest error. Therefore, we cannot state which one is the most adequate for the problem.
Regarding the feedforward neural models, one can observe for 16 of 20 cases (80%), the ELMs overcame the traditional MLP and RBF architectures. In the same way, the ESNs were superior to the traditional and the fully trained Elman and Jordan proposals in 17 of 20 scenarios (85%). This is strong evidence that the unorganized models are prominent candidates to carry out such problems.
Linear models did not outperform neural networks in any of the 20 scenarios. For the problem of forecasting monthly seasonal streamflow series, the results showed that ANNs were most appropriate. However, it is worth mentioning that linear models are still widely used in current days.
Finally, to provide a visual appreciation of the final simulation, Figure 7 presents the forecast made by the ELM ensemble for Água Vermelha plant with P = 1 .

5. Conclusions

This work investigated the performance of unorganized machines—extreme learning machines (ELM), echo state networks (ESN), and ELM-based ensembles—on monthly seasonal streamflow series forecasting from hydroelectric plants. This is a very important task for countries where power generation is highly dependent on water as a source, such as Canada, China, Brazil, and the USA, among others. Due to the broad use of this kind of energy generation in the world, even a small improvement in the accuracy of the predictions can lead to significant financial resource savings as well as a reduction in the impact of using fossil fuels.
We also used many artificial neural network (ANN) architectures—multilayer perceptron (MLP), radial basis function networks (RBF), Jordan network, Elman network, and the ensemble methodology using the mean, the median, the MLP, and the RBF as combiner. Moreover, we compared the results with the traditional AR and ARMA linear models. We addressed four forecast horizons, P = 1 , 3 , 6 , and 12 steps ahead, and the wrapper method to select the best delays (inputs).
The case study involves a database related to five hydroelectric plants. The tests showed that the neural ensembles were the most indicated for P = 1 since they presented the best performances in all the simulations of this scenario, especially those that employed the ELM. For P = 3 and 6, the ELM was highlighted. For P = 12, it was clear that the recurrent models were outstanding, mainly those with the ESN.
Regarding the linear models, this work showed its inferiority in comparison to the neural ones in all cases. Furthermore, the unorganized neural models (ELM and ESN), in their stand-alone versions or as combiners of the ensemble, prevailed over the others, presenting 14 of the lowest errors (70%).
These results are important since the unorganized machines are easy to implement and require less computational effort than the fully trained approaches. This is related to the use of the Moore-Penrose inverse operator to train their output layer, since it ensures the optimum value of the weights in the mean square error sense. The use of the backpropagation could lead the process to a local minimum, indicating how difficult the problem is. In at least 80% of the cases, the unorganized proposals (ELM and ESN) overcame the fully trained proposals (MLP, RBF, Elman, and Jordan).
Other deseasonalization processes should be investigated in future works. Additionally, the streamflow from distinct plants must be predicted and the results evaluated. Moreover, the use of bio-inspired optimization methods [76,77,78] is encouraged to optimize the ARMA model and the application of the support vector regression method.

Author Contributions

Conceptualization, H.S., S.L.S.J., J.B., L.A.; methodology, H.S., S.L.S.J., J.B., L.A., software, H.S., J.B.; validation, H.S., P.S.G.d.M.N., J.F.L.d.O.; formal analysis, P.S.G.d.M.N., M.H.N.M., J.F.L.d.O.; investigation, J.B., L.A.; resources: M.H.N.M.; writing original draft preparation, H.S., J.B., L.A.S. and I.C.; writing, review and editing, M.d.A.L.F., F.U., L.A.S., A.C.; visualization, P.S.G.d.M.N., J.F.L.d.O.; supervision, A.C. and M.H.N.M. project administration, M.d.A.L.F., M.H.N.M.; funding acquisition, L.A.S., M.d.A.L.F. and M.H.N.M., All authors have read and agreed to the published version of the manuscript.

Funding

This work received funding and technical support from AES and Associated Companies (of the CPFL, Brookfield and Global group) as part of the ANEEL PD-0610-1004/2015 project, “IRIS—Integration of Intermittent Renewables: A simulation model of the operation of the Brazilian electrical system to support planning, operation, commercialization, and regulation”, which is part of an R&D program regulated by ANEEL, Brazil. The authors also thank the Advanced Institute of Technology and Innovation (IATI) for its support, the National Institute of Meteorology (INMET) for providing the data, and the Coordination for the Improvement of Higher Education Personnel—Brazil (CAPES)—Financing Code 001, for support.

Acknowledgments

This work was partially supported by the Brazilian agencies CAPES and FACEPE. The authors thank the Brazilian National Council for Scientific and Technological Development (CNPq), process number 40558/2018-5, and Araucaria Foundation, process number 51497, for their financial support. The authors also thank the Federal University of Technology.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Marcjasz, G.; Uniejewski, B.; Weron, R. Beating the Naïve—Combining LASSO with Naïve Intraday Electricity Price Forecasts. Energies 2020, 13, 1667. [Google Scholar] [CrossRef] [Green Version]
  2. Sigauke, C.; Nemukula, M.M.; Maposa, D. Probabilistic hourly load forecasting using additive quantile regression models. Energies 2018, 11, 2208. [Google Scholar] [CrossRef] [Green Version]
  3. Malfatti, M.G.L.; Cardoso, A.O.; Hamburger, D.S. Linear Empirical Model for Streamflow Forecast in Itaipu Hydroelectric Dam–Parana River Basin. Rev. Bras. Meteorol. 2018, 33, 257–268. [Google Scholar] [CrossRef] [Green Version]
  4. Moscoso-López, J.A.; Turias, I.; Jiménez-Come, M.J.; Ruiz-Aguilar, J.J.; del Cerbán, M.M. A two-stage forecasting approach for short-term intermodal freight prediction. Int. Trans. Oper. Res. 2019, 26, 642–666. [Google Scholar]
  5. Francelin, R.; Ballini, R.; Andrade, M.G. Back-propagation and Box & Jenkins approaches to streamflow forecasting. Simpósio Bras. Pesqui. Oper.-SBPO Lat.-Iber.-Am. Congr. Oper. Res. Syst. Eng.-CLAIO 1996, 3, 1307–1312. [Google Scholar]
  6. Sacchi, R.; Ozturk, M.C.; Principe, J.C.; Carneiro, A.A.F.M.; Silva, I.N. Water inflow forecasting using the echo state network: A Brazilian case study. In Proceedings of the 2007 International Joint Conference on Neural Networks, Orlando, FL, USA, 12–17 August 2007; pp. 2403–2408. [Google Scholar]
  7. Yaseen, Z.; Sulaiman, S.; Deo, R.; Chau, K.W. An enhanced extreme learning machine model for river flow forecasting: State-of-the-art, practical applications in water resource engineering area and future research direction. J. Hydrol. 2018, 569. [Google Scholar] [CrossRef]
  8. International Hydropower Association. Hydropower Status Report; International Hydropower Association: London, UK, 2019. [Google Scholar]
  9. Siqueira, H.; Boccato, L.; Luna, I.; Attux, R.; Lyra, C. Performance analysis of unorganized machines in streamflow forecasting of Brazilian plants. Appl. Soft Comput. 2018, 68, 494–506. [Google Scholar] [CrossRef]
  10. Wu, C.; Chau, K.W. Data-driven models for monthly streamflow time series prediction. Eng. Appl. Artif. Intell. 2010, 23, 1350–1367. [Google Scholar] [CrossRef] [Green Version]
  11. Kisi, O.; Cimen, M. Precipitation forecasting by using wavelet-support vector machine conjunction model. Eng. Appl. Artif. Intell. 2012, 25, 783–792. [Google Scholar] [CrossRef]
  12. Siqueira, H.; Boccato, L.; Attux, R.; Lyra, C. Unorganized machines for seasonal streamflow series forecasting. Int. J. Neural Syst. 2014, 24, 1430009. [Google Scholar] [CrossRef]
  13. Taormina, R.; Chau, K.W. ANN-based interval forecasting of streamflow discharges using the LUBE method and MOFIPS. Eng. Appl. Artif. Intell. 2015, 45, 429–440. [Google Scholar] [CrossRef]
  14. Zhu, S.; Zhou, J.; Ye, L.; Meng, C. Streamflow estimation by support vector machine coupled with different methods of time series decomposition in the upper reaches of Yangtze River, China. Environ. Earth Sci. 2016, 75, 531. [Google Scholar] [CrossRef]
  15. Arsenault, R.; Côté, P. Analysis of the effects of biases in ensemble streamflow prediction (ESP) forecasts on electricity production in hydropower reservoir management. Hydrol. Earth Syst. Sci. 2019, 23, 2735–2750. [Google Scholar] [CrossRef] [Green Version]
  16. Stojković, M.; Kostić, S.; Prohaska, S.; Plavšić, J.; Tripković, V. A New Approach for Trend Assessment of Annual Streamflows: A Case Study of Hydropower Plants in Serbia. Water Resour. Manag. 2017, 31, 1089–1103. [Google Scholar] [CrossRef]
  17. Hailegeorgis, T.T.; Alfredsen, K. Regional statistical and precipitation–runoff modelling for ecological applications: Prediction of hourly streamflow in regulated rivers and ungauged basins. River Res. Appl. 2017, 33, 233–248. [Google Scholar] [CrossRef] [Green Version]
  18. Siqueira, H.V.; Boccato, L.; Attux, R.R.F.; Lyra Filho, C. Echo State Networks in Seasonal Streamflow Series Prediction. Learn. Nonlinear Models 2012, 10, 181–191. [Google Scholar] [CrossRef] [Green Version]
  19. Box, G.E.; Jenkins, G.M.; Reinsel, G.C.; Ljung, G.M. Time Series Analysis: Forecasting and Control; John Wiley & Sons: New York, NY, USA, 2015. [Google Scholar]
  20. Hwang, W.Y.; Lee, J.S. A new forecasting scheme for evaluating long-term prediction performances in supply chain management. Int. Trans. Oper. Res. 2014, 21, 1045–1060. [Google Scholar] [CrossRef]
  21. Singh, P.; Deo, M. Suitability of different neural networks in daily flow forecasting. Appl. Soft Comput. 2007, 7, 968–978. [Google Scholar] [CrossRef]
  22. Jain, A.; Kumar, A.M. Hybrid neural network models for hydrologic time series forecasting. Appl. Soft Comput. 2007, 7, 585–592. [Google Scholar] [CrossRef]
  23. Toro, C.H.F.; Meire, S.G.; Gálvez, J.F.; Fdez-Riverola, F. A hybrid artificial intelligence model for river flow forecasting. Appl. Soft Comput. 2013, 13, 3449–3458. [Google Scholar] [CrossRef]
  24. Kazemi, S.; Seied Hoseini, M.M.; Abbasian-Naghneh, S.; Rahmati, S.H.A. An evolutionary-based adaptive neuro-fuzzy inference system for intelligent short-term load forecasting. Int. Trans. Oper. Res. 2014, 21, 311–326. [Google Scholar] [CrossRef]
  25. Haykin, S.S. Neural Networks and Learning Machines; Prentice Hall: New York, NY, USA, 2009. [Google Scholar]
  26. Singh, P.; Borah, B. An efficient time series forecasting model based on fuzzy time series. Eng. Appl. Artif. Intell. 2013, 26, 2443–2457. [Google Scholar] [CrossRef]
  27. Parsa, N.; Keshavarz, T.; Karimi, B.; Moattar Husseini, S.M. A hybrid neural network approach to minimize total completion time on a single batch processing machine. Int. Trans. Oper. Res. 2020. [Google Scholar] [CrossRef]
  28. Sharkey, A.J.C. Combining Artificial Neural Nets: Ensemble and Modular Multi-Net Systems; Springer: London, UK, 1999. [Google Scholar]
  29. De Mattos Neto, P.S.; Madeiro, F.; Ferreira, T.A.; Cavalcanti, G.D. Hybrid intelligent system for air quality forecasting using phase adjustment. Eng. Appl. Artif. Intell. 2014, 32, 185–191. [Google Scholar] [CrossRef]
  30. Firmino, P.R.A.; de Mattos Neto, P.S.; Ferreira, T.A. Correcting and combining time series forecasters. Neural Netw. 2014, 50, 1–11. [Google Scholar] [CrossRef]
  31. Kasiviswanathan, K.S.; Sudheer, K.P. Quantification of the predictive uncertainty of artificial neural network based river flow forecast models. Stoch. Environ. Res. Risk Assess. 2013, 27, 137–146. [Google Scholar] [CrossRef]
  32. Fan, F.M.; Schwanenberg, D.; Alvarado, R.; Reis, A.A.; Collischonn, W.; Naumman, S. Performance of Deterministic and Probabilistic Hydrological Forecasts for the Short-Term Optimization of a Tropical Hydropower Reservoir. Water Resour. Manag. 2016, 30, 3609–3625. [Google Scholar] [CrossRef] [Green Version]
  33. Thober, S.; Kumar, R.; Wanders, N.; Marx, A.; Pan, M.; Rakovec, O.; Samaniego, L.; Sheffield, J.; Wood, E.F.; Zink, M. Multi-model ensemble projections of European river floods and high flows at 1.5, 2, and 3 degrees global warming. Environ. Res. Lett. 2018, 13, 014003. [Google Scholar] [CrossRef]
  34. EPE—Energy Research Company. National Energy Balance 2019; Ministry of Mines and Energy: Rio de Janeiro, Brazil, 2019.
  35. Siqueira, H.V.; Boccato, L.; Attux, R.R.F.; Lyra Filho, C. Echo state networks and extreme learning machines: A comparative study on seasonal streamflow series prediction. Lect. Notes Comput. Sci. 2012, 7664, 491–500. [Google Scholar]
  36. Haykin, S.O. Adaptive Filter Theory; Pearson Higher: London, UK, 2013. [Google Scholar]
  37. Cao, Q.; Ewing, B.T.; Thompson, M.A. Forecasting wind speed with recurrent neural networks. Eur. J. Oper. Res. 2012, 221, 148–154. [Google Scholar] [CrossRef]
  38. Barrow, D.; Kourentzes, N. The impact of special days in call arrivals forecasting: A neural network approach to modelling special days. Eur. J. Oper. Res. 2018, 264, 967–977. [Google Scholar] [CrossRef] [Green Version]
  39. Cybenko, G. Approximation by superpositions of a sigmoidal function. Math. Control Signals Syst. (MCSS) 1989, 2, 303–314. [Google Scholar] [CrossRef]
  40. Polezer, G.; Tadano, Y.S.; Siqueira, H.V.; Godoi, A.F.; Yamamoto, C.I.; de André, P.A.; Pauliquevis, T.; de Fatima Andrade, M.; Oliveira, A.; Saldiva, P.H.; et al. Assessing the impact of PM2.5 on respiratory disease using artificial neural networks. Environ. Pollut. 2018, 235, 394–403. [Google Scholar] [CrossRef] [PubMed]
  41. Reifman, J.; Feldman, E.E. Multilayer perceptron for nonlinear programming. Comput. Oper. Res. 2002, 29, 1237–1250. [Google Scholar] [CrossRef]
  42. Feng, S.; Li, L.; Cen, L.; Huang, J. Using MLP networks to design a production scheduling system. Comput. Oper. Res. 2003, 30, 821–832. [Google Scholar] [CrossRef]
  43. Dahl, G.E.; Stokes, J.W.; Deng, L.; Yu, D. Large-scale malware classification using random projections and neural networks. In Proceedings of the 2013 IEEE International Conference on Acoustics, Speech and Signal Processing, Vancouver, BC, Canada, 26–31 May 2013; IEEE: Piscataway, NJ, USA, 2013; pp. 3422–3426. [Google Scholar]
  44. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. In Advances in Neural Information Processing Systems; Neural Information Processing Systems Foundation, Inc.: South Lake Tahoe, NV, USA, 2012; pp. 1097–1105. [Google Scholar]
  45. Kristjanpoller, W.; Minutolo, M.C. Gold price volatility: A forecasting approach using the Artificial Neural Network–GARCH model. Expert Syst. Appl. 2015, 42, 7245–7251. [Google Scholar] [CrossRef]
  46. Zhang, G.P.; Qi, M. Neural network forecasting for seasonal and trend time series. Eur. J. Oper. Res. 2005, 160, 501–514. [Google Scholar] [CrossRef]
  47. Rendon-Sanchez, J.F.; de Menezes, L.M. Structural combination of seasonal exponential smoothing forecasts applied to load forecasting. Eur. J. Oper. Res. 2019, 275, 916–924. [Google Scholar] [CrossRef]
  48. Werbos, P.J. Beyond Regression: New Tools for Prediction and Analysis in the Behavioral Sciences. Ph.D. Thesis, Harvard University, Cambridge, MA, USA, 1974. [Google Scholar]
  49. Rumelhart, D.E.; Hinton, G.E.; Williams, R.J. Learning representations by back-propagating errors. Cogn. Model. 1986, 5, 1. [Google Scholar] [CrossRef]
  50. Siqueira, H.; Luna, I. Performance comparison of feedforward neural networks applied to stream flow series forecasting. J. MESA 2019, 10, 41–53. [Google Scholar]
  51. Elman, J.L. Finding structure in time. Cogn. Sci. 1990, 14, 179–211. [Google Scholar] [CrossRef]
  52. Jordan, M. Attractor Dynamics and Parallelism in a Connectionist Sequential Machine. In Proceedings of the Eighth Annual Conference of the Cognitive Science Society, Amherst, MA, USA, 15–17 August 1986; Erlbaum: Hillsdale, NJ, USA, 1986; pp. 531–546. [Google Scholar]
  53. Huang, G.H.; Zhu, Q.Y.; Siew, C.K. Extreme learning machine: A new learning scheme of feedforward neural networks. In Proceedings of the 2004 IEEE International Joint Conference on Neural Networks, Budapest, Hungary, 25–29 July 2004; Volume 2, pp. 985–990. [Google Scholar]
  54. Huang, G.H.; Zhou, H.; Ding, X.; Zhang, R. Extreme learning machine for regression and multiclass classification. Trans. Syst. Man Cybern.-Part B Cybern. 2012, 42, 513–529. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  55. Huang, G.B.; Zhu, Q.Y.; Siew, C.K. Extreme learning machine: Theory and applications. Neurocomputing 2006, 70, 489–501. [Google Scholar] [CrossRef]
  56. Jaeger, H. The “echo state” approach to analysing and training recurrent neural networks-with an erratum note. Ger. Natl. Res. Cent. Inf. Technol. GMD Tech. Rep. 2001, 148, 13. [Google Scholar]
  57. Ozturk, M.C.; Xu, D.; Príncipe, J.C. Analysis and Design of Echo State Networks for Function Approximation. Neural Comput. 2007, 19, 111–138. [Google Scholar] [CrossRef]
  58. Wichard, J.D.; Ogorzalek, M. Time series prediction with ensemble models. In Proceedings of the 2004 IEEE International Joint Conference on Neural Networks, Budapest, Hungary, 25–29 July 2004; IEEE: Piscataway, NJ, USA, 2004; Volume 2, pp. 1625–1630. [Google Scholar]
  59. Hansen, L.K.; Salamon, P. Neural network ensembles. IEEE Trans. Pattern Anal. Mach. Intell. 1990, 12, 993–1001. [Google Scholar] [CrossRef] [Green Version]
  60. Yu, L.; Wang, S.; Lai, K.K. A novel nonlinear ensemble forecasting model incorporating GLAR and ANN for foreign exchange rates. Comput. Oper. Res. 2005, 32, 2523–2541. [Google Scholar] [CrossRef]
  61. West, D.; Dellana, S.; Qian, J. Neural network ensemble strategies for financial decision applications. Comput. Oper. Res. 2005, 32, 2543–2559. [Google Scholar] [CrossRef]
  62. Yang, D. Spatial prediction using kriging ensemble. Sol. Energy 2018, 171, 977–982. [Google Scholar] [CrossRef]
  63. Kim, D.; Hur, J. Short-term probabilistic forecasting of wind energy resources using the enhanced ensemble method. Energy 2018, 157, 211–226. [Google Scholar] [CrossRef]
  64. Kumar, N.K.; Savitha, R.; Al Mamun, A. Ocean wave height prediction using ensemble of Extreme Learning Machine. Neurocomputing 2018, 277, 12–20. [Google Scholar] [CrossRef]
  65. ONS—Electric System Operator. Database of “Hydrological Data/Streamflows”. 2018. Available online: http://ons.org.br/Paginas/resultados-da-operacao/historico-da-operacao/dados-hidrologicos-vazoes.aspx (accessed on 1 February 2018).
  66. Guyon, I.; Elisseeff, A. An introduction to variable and feature selection. J. Mach. Learn. Res. 2003, 3, 1157–1182. [Google Scholar]
  67. Wang, H.; Sun, J.; Sun, J.; Wang, J. Using random forests to select optimal input variables for short-term wind speed forecasting models. Energies 2017, 10, 1522. [Google Scholar] [CrossRef] [Green Version]
  68. Dong, W.; Yang, Q.; Fang, X. Multi-step ahead wind power generation prediction based on hybrid machine learning techniques. Energies 2018, 11, 1975. [Google Scholar] [CrossRef] [Green Version]
  69. Sorjamaa, A.; Hao, J.; Reyhani, N.; Ji, Y.; Lendasse, A. Methodology for long-term prediction of time series. Neurocomputing 2007, 70, 2861–2869. [Google Scholar] [CrossRef] [Green Version]
  70. Stedinger, J.R. Report on the Evaluation of Cepel’s Par Models: Techical Report; School of Civil and Environmental Engineering-Cornell University: Ithaca, NY, USA, 2001. [Google Scholar]
  71. Friedman, M. The use of ranks to avoid the assumption of normality implicit in the analysis of variance. J. Am. Stat. Assoc. 1937, 32, 675–701. [Google Scholar] [CrossRef]
  72. Araujo, N.A.; Belotti, J.T.; Antonini Alves, T.; Tadano, Y.S.; Siqueira, H. Ensemble method based on Artificial Neural Networks to estimate air pollution health risks. Environ. Model. Softw. 2020, 123, 104567. [Google Scholar] [CrossRef]
  73. Siqueira, H.V.; Boccato, L.; Attux, R.R.F.; Lyra Filho, C. Echo state networks for seasonal streamflow series forecasting. Lect. Notes Comput. Sci. 2012, 7435, 226–236. [Google Scholar]
  74. Kachba, Y.; Chiroli, D.M.; Belotti, J.T.; Alves, T.A.; de Souza Tadano, Y.; Siqueira, H. Artificial neural networks to estimate the influence of vehicular emission variables on morbidity and mortality in the largest metropolis in South America. Sustainability 2020, 12, 2621. [Google Scholar] [CrossRef] [Green Version]
  75. Siqueira, H.; Luna, I.; Antonini Alves, T.; de Souza Tadano, Y. The direct connection between Box & Jenkins methodology and adaptive filtering theory. Math. Eng. Sci. Aerosp. (MESA) 2019, 10, 27–40. [Google Scholar]
  76. Santos, P.; Macedo, M.; Figueiredo, E.; Santana, C.S., Jr.; Soares, F.; Siqueira, H.; Maciel, A.; Gokhale, A.; Bastos-Filho, C.J.A. Application of PSO-based clustering algorithms on educational databases. In Proceedings of the 2017 IEEE Latin American Conference on Computational Intelligence (LA-CCI), Arequipa, Peru, 8–10 November 2017; pp. 1–6. [Google Scholar]
  77. Puchta, E.; Lucas, R.; Ferreira, F.R.V.; Siqueira, H.V.; Kaster, M.S. Gaussian adaptive PID control optimized via genetic algorithm applied to a step-down DC-DC converter. In Proceedings of the 2016 12th IEEE International Conference on Industry Applications (INDUSCON), Curitiba, Brazil, 20–23 November 2016; pp. 1–6. [Google Scholar]
  78. Santana, C.J., Jr.; Macedo, M.; Siqueira, H.; Gokhale, A.; Bastos-Filho, C.J.A. A novel binary artificial bee colony algorithm. Future Gener. Comp. Syst. 2019, 98, 180–196. [Google Scholar] [CrossRef]
Figure 1. Elman network.
Figure 1. Elman network.
Energies 13 04769 g001
Figure 2. Jordan network.
Figure 2. Jordan network.
Energies 13 04769 g002
Figure 3. Echo state network (ESN).
Figure 3. Echo state network (ESN).
Energies 13 04769 g003
Figure 4. Schematic of an ensemble.
Figure 4. Schematic of an ensemble.
Energies 13 04769 g004
Figure 5. Preprocessing and postprocessing stages.
Figure 5. Preprocessing and postprocessing stages.
Energies 13 04769 g005
Figure 6. Boxsplot Água Vermelha, 2006 at 2015–30 executions.
Figure 6. Boxsplot Água Vermelha, 2006 at 2015–30 executions.
Energies 13 04769 g006
Figure 7. Best predicted to Água Vermelha of 2006 at 2015 and P = 1.
Figure 7. Best predicted to Água Vermelha of 2006 at 2015 and P = 1.
Energies 13 04769 g007
Table 1. Mean and standard deviation of the sets.
Table 1. Mean and standard deviation of the sets.
SeriesMean Standard Deviation
Agua Vermelha2077.33 m3/s1295.71 m3/s
Belo Monte8045.84 m3/s7769.75 m3/s
Ilha Solteira5281.94 m3/s3100.62 m3/s
Paulo Afonso2698.80 m3/s2026.48 m3/s
Tucuruí10,935.23 m3/s9182.29 m3/s
Table 2. Computational results for Água Vermelha.
Table 2. Computational results for Água Vermelha.
ModelP = 1P = 3
MSEMSE (d)MSEMSE (d)
AR(3)468,513.810.4290635,623.140.6901
ARMA(2,2)459,868.320.4161652,368.140.6562
MLP435,047.550.4321592,399.820.6701
RBF433,015.370.4903616,761.670.7871
ELM413,488.420.3893548,179.420.5522
ELM (RC)393,462.780.3901548,521.400.5862
Elman391,192.040.4013590,235.980.6513
Jordan417,081.010.4181642,432.880.7443
Jaeger ESN383,174.660.3760576,896.700.5940
Jaeger ESN (RC)379,043.860.3781565,927.300.5891
Ozturk ESN401,843.390.3772745,343.330.9742
Ozturk ESN (RC)397,982.110.3772833,991.161.1010
Average Ensemble374,741.570.3683543,746.080.5590
Median Ensemble379,043.860.3782574,877.030.5901
MLP Ensemble381,381.740.3951595,528.640.6240
RBF Ensemble378,694.990.3910582,641.850.6120
ELM Ensemble360,574.550.3771598,571.400.6311
ELM Ensemble (RC)362,264.360.3870596,396.640.6390
ModelP = 6P = 12
MSEMSE (d)MSEMSE (d)
AR(3)920,536.060.95601,032,390.381.3101
ARMA(2,2)857,987.670.8951844,191.841.0820
MLP705,060.110.8360717,730.320.8534
RBF1,003,756.281.30102,139,100.523.1500
ELM602,358.840.6664695,473.230.8123
ELM (RC)643,385.780.7177735,096.120.8840
Elman719,246.600.8161733,779.790.8746
Jordan726,929.580.8792711,201.570.8455
Jaeger ESN650,640.110.7129679,957.560.8021
Jaeger ESN (RC)632,523.340.6990690,611.820.8196
Ozturk ESN671,232.690.8152726,758.250.8643
Ozturk ESN (RC)695,143.950.8861749,393.170.8869
Average Ensemble718,287.320.7420736,530.620.9079
Median Ensemble806,882.940.8172761,767.760.9891
MLP Ensemble796,168.680.8097779,261.850.9861
RBF Ensemble774,132.080.7849738,982.180.9638
ELM Ensemble834,303.170.8658800,998.371.0501
ELM Ensemble (RC)837,861.291.00211,252,464.381.7002
Table 3. Computational results for Belo Monte.
Table 3. Computational results for Belo Monte.
ModelP = 1P = 3
MSEMSE (d)MSEMSE (d)
AR(3)5,136,070.020.36819,481,744.900.7818
ARMA(4,3)5,560,980.020.397810,338,028.240.7803
MLP4,611,515.040.39566,610,976.000.7674
RBF4,189,519.180.39767,374,107.900.7756
ELM3,932,745.860.44555,211,473.530.7751
ELM (RC)4,263,133.640.36205,977,477.680.6716
Elman4,348,164.470.431012,461,369.171.6201
Jordan4,841,287.430.37958,029,604.930.7721
Jaeger ESN3,990,917.220.36195,744,660.750.6405
Jaeger ESN (RC)4,063,963.410.37036,268,470.740.6559
Ozturk ESN4,135,861.200.39866,121,782.430.6100
Ozturk ESN (RC)4,054,930.750.37695,576,790.170.6437
Average Ensemble3,664,436.730.37386,583,650.150.6255
Median Ensemble3,654,331.870.33797,892,943.440.6883
MLP Ensemble3,683,236.200.34959,631,438.350.7448
RBF Ensemble3,716,658.520.40826,756,906.300.7051
ELM Ensemble3,496,325.520.35837,360,295.020.6748
ELM Ensemble (RC)3,537,238.590.35227,038,931.440.6529
ModelP = 6P = 12
MSEMSE (d)MSEMSE (d)
AR(3)13,220,246.311.070113,856,487.041.3102
ARMA(4,3)14,598,946.901.070110,811,736.131.3103
MLP8,503,685.650.91588,834,763.120.8732
RBF9,560,433.480.953410,802,574.681.1402
ELM5,390,642.110.81296,258,443.280.7798
ELM (RC)6,266,406.910.72526,209,801.530.7287
Elman6,837,675.111.04017,177,995.010.9005
Jordan10,176,393.170.974810,698,937.701.0302
Jaeger ESN5,778,132.120.72075,938,616.010.7261
Jaeger ESN (RC)5,681,748.340.74465,942,372.260.7407
Ozturk ESN5,906,402.730.68606,391,806.490.7184
Ozturk ESN (RC)5,764,898.140.73255,911,871.830.7356
Average Ensemble10,217,687.600.83059,363,944.041.1201
Median Ensemble12,782,750.960.996311,326,906.601.3901
MLP Ensemble14,838,424.951.160012,489,587.181.6103
RBF Ensemble11,651,888.441.030112,834,385.611.2302
ELM Ensemble11,834,559.770.961310,523,981.041.3001
ELM Ensemble (RC)11,409,184.740.932310,465,521.501.2700
Table 4. Computational results for Ilha Solteira.
Table 4. Computational results for Ilha Solteira.
ModelP = 1P = 3
MSEMSE (d)MSEMSE (d)
AR(6)2,966,223.640.54513,607,452.300.8363
ARMA(2,2)2,820,456.340.53113,575,263.580.8337
MLP2,673,739.640.53663,385,476.990.7607
RBF2,646,716.440.57453,910,685.330.9068
ELM2,545,002.110.49733,316,801.940.7265
ELM (RC)2,498,945.940.52403,445,130.560.7750
Elman5,689,735.150.57133,719,407.930.8549
Jordan2,722,458.650.55243,557,423.140.8219
Jaeger ESN2,511,512.570.49143,259,442.700.7219
Jaeger ESN (RC)2,568,883.380.49993,294,093.730.7489
Ozturk ESN2,337,398.440.46633,428,320.790.7850
Ozturk ESN (RC)2,313,534.090.47683,646,528.580.8801
Average Ensemble2,277,545.050.45933,316,183.090.7599
Median Ensemble2,313,534.090.47683,230,939.250.7472
MLP Ensemble2,312,233.010.49453,196,000.870.7135
RBF Ensemble2,006,994.790.49033,468,793.600.8388
ELM Ensemble2,084,515.900.45463,377,696.820.7866
ELM Ensemble (RC)1,992,611.390.44913,265,153.090.7606
ModelP = 6P = 12
MSEMSE (d)MSEMSE (d)
AR(6)4,511,751.901.11004,773,379.181.1900
ARMA(2,2)4,473,867.421.12014,494,073.051.1301
MLP3,920,672.520.90254,049,775.430.9433
RBF4,307,512.831.01004,283,245.081.1101
ELM3,761,082.940.84803,873,269.310.8967
ELM (RC)3,823,661.640.89343,815,701.760.8860
Elman4,154,039.410.96984,116,367.320.9615
Jordan4,169,869.490.96624,070,205.280.9373
Jaeger ESN3,838,690.480.88143,477,087.730.8497
Jaeger ESN (RC)3,884,498.960.87273,969,786.680.9293
Ozturk ESN3,938,797.730.96293,712,781.930.8854
Ozturk ESN (RC)3,864,185.960.94794,670,725.870.9975
Average Ensemble4,188,451.091.05034,920,660.691.1702
Median Ensemble4,131,637.371.06014,970,254.981.1602
MLP Ensemble3,982,036.000.98744,739,091.941.0910
RBF Ensemble3,934,481.311.04036,385,242.271.4903
ELM Ensemble4,419,019.181.17026,998,311.412.4601
ELM Ensemble (RC)4,586,375.571.31016,612,087.531.4502
Table 5. Computational results for Paulo Afonso.
Table 5. Computational results for Paulo Afonso.
ModelP = 1P = 3
MSEMSE (d)MSEMSE (d)
AR(6)726,145.700.29311,129,419.580.5697
ARMA(2,1)694,185.050.28741,131,738.270.5453
MLP635,673.560.34371,058,546.460.6355
RBF719,377.510.41032,000,897.181.1801
ELM602,176.480.3312938,475.790.5768
ELM (RC)607,078.510.32371,089,089.420.6389
Elman617,664.770.2869936,344.630.4566
Jordan673,001.750.33981,166,257.860.6914
Jaeger ESN566,929.670.2786929,239.630.5230
Jaeger ESN (RC)627,573.850.34151,490,572.960.9615
Ozturk ESN576,338.260.29571,256,147.760.6778
Ozturk ESN (RC)581,609.270.30951,047,548.670.5800
Average Ensemble551,531.610.29081,230,760.560.5601
Median Ensemble552,252.810.25211,169,891.730.5154
MLP Ensemble548,535.130.29311,185,108.360.5330
RBF Ensemble554,508.820.2986987,644.320.4806
ELM Ensemble535,210.830.27141,088,286.190.4940
ELM Ensemble (RC)535,865.260.30161,193,241.920.5780
ModelP = 6P = 12
MSEMSE (d)MSEMSE (d)
AR(6)1,088,288.610.74601,916,895.361.1101
ARMA(2,1)1,024,438.220.68061,634,060.460.9137
MLP1,113,602.560.68551,118,376.670.6888
RBF3,779,964.352.600113,040,123.348.1102
ELM980,510.070.63571,290,254.850.7766
ELM (RC)1,129,832.900.72201,723,308.911.1100
Elman981,800.010.58801,214,455.570.7626
Jordan1,209,656.520.77081,258,279.090.7960
Jaeger ESN1,106,672.850.68471,364,591.040.8514
Jaeger ESN (RC)1,278,004.660.87671,852,101.701.1703
Ozturk ESN1,040,841.920.61801,250,520.940.7806
Ozturk ESN (RC)1,083,715.180.76171,348,961.530.8507
Average Ensemble1,025,453.520.61821,420,331.330.7761
Median Ensemble1,078,563.810.58531,334,761.670.7067
MLP Ensemble1,063,431.440.59141,311,740.310.7058
RBF Ensemble1,058,247.060.63671,393,776.650.7927
ELM Ensemble989,046.790.53911,267,630.780.6619
ELM Ensemble (RC)997,954.040.59941,346,719.340.7240
Table 6. Computational results for Tucuruí.
Table 6. Computational results for Tucuruí.
ModelP = 1P = 3
MSEMSE (d)MSEMSE (d)
AR(3)7,446,258.520.331013,406,528.210.9131
ARMA(4,1)7,991,096.150.338313,974,802.610.9714
MLP7,100,370.960.345111,346,681.970.7834
RBF6,721,796.960.359211,966,687.620.7750
ELM6,824,285.760.303010,906,829.600.7399
ELM (RC)6,829,099.840.306611,134,575.770.7359
Elman7,351,576.910.410713,167,672.490.8019
Jordan6,980,914.400.325111,752,842.370.8432
Jaeger ESN6,936,607.010.302411,639,311.670.6543
Jaeger ESN (RC)6,975,358.520.303513,382,939.200.7647
Ozturk ESN6,757,516.840.308417,883,748.200.8628
Ozturk ESN (RC)7,119,853.310.331011,180,015.340.6629
Average Ensemble6,254,104.500.308912,708,717.470.7938
Median Ensemble6,757,516.840.308414,733,468.980.8444
MLP Ensemble6,317,164.230.312112,726,659.050.8268
RBF Ensemble5,776,403.720.312014,878,993.970.8811
ELM Ensemble5,923,084.050.279215,279,206.301.0902
ELM Ensemble (RC)5,859,606.760.281014,830,513.861.0101
ModelP = 6P = 12
MSEMSE (d)MSEMSE (d)
AR(3)15,727,475.821.200122,161,343.501.4602
ARMA(4,1)16,315,642.781.370029,987,913.131.9100
MLP11,658,727.900.813111,743,205.590.8254
RBF12,696,569.790.983015,087,729.250.9688
ELM11,989,453.790.823812,304,667.660.8764
ELM (RC)12,143,842.920.823612,284,951.170.8614
Elman14,163,188.221.100210,533,366.450.6894
Jordan12,264,013.860.886012,341,774.820.8899
Jaeger ESN12,040,332.150.754811,766,645.650.8119
Jaeger ESN (RC)12,071,588.190.734512,279,058.450.8578
Ozturk ESN16,299,760.490.825611,238,408.860.7952
Ozturk ESN (RC)11,770,003.630.706210,807,257.910.6965
Average Ensemble16,469,919.610.998319,601,009.991.0703
Median Ensemble20,303,247.251.120020,367,666.361.1801
MLP Ensemble15,554,587.870.971021,292,183.311.0702
RBF Ensemble18,606,965.661.020122,318,965.501.2100
ELM Ensemble21,731,212.231.280231,221,147.231.5203
ELM Ensemble (RC)20,723,489.571.260134,497,724.891.5801
Table 7. Best model by horizon.
Table 7. Best model by horizon.
ModelP = 1P = 3P = 6P = 12Total
AR(3)----00
ARMA(4,1)----0
MLP--1122
RBF----0
ELM-24-66
ELM (RC)----0
Elman---- 1 1
Jordan---- 0
Jaeger ESN-1-234
Jaeger ESN (RC)----0
Ozturk ESN----0
Ozturk ESN (RC)---11
Average Ensemble-1--17
Median Ensemble----0
MLP Ensemble-1--1
RBF Ensemble1---1
ELM Ensemble3---3
ELM Ensemble (RC)1---1

Share and Cite

MDPI and ACS Style

Belotti, J.; Siqueira, H.; Araujo, L.; Stevan, S.L., Jr.; de Mattos Neto, P.S.G.; Marinho, M.H.N.; de Oliveira, J.F.L.; Usberti, F.; Leone Filho, M.d.A.; Converti, A.; et al. Neural-Based Ensembles and Unorganized Machines to Predict Streamflow Series from Hydroelectric Plants. Energies 2020, 13, 4769. https://doi.org/10.3390/en13184769

AMA Style

Belotti J, Siqueira H, Araujo L, Stevan SL Jr., de Mattos Neto PSG, Marinho MHN, de Oliveira JFL, Usberti F, Leone Filho MdA, Converti A, et al. Neural-Based Ensembles and Unorganized Machines to Predict Streamflow Series from Hydroelectric Plants. Energies. 2020; 13(18):4769. https://doi.org/10.3390/en13184769

Chicago/Turabian Style

Belotti, Jônatas, Hugo Siqueira, Lilian Araujo, Sérgio L. Stevan, Jr., Paulo S.G. de Mattos Neto, Manoel H. N. Marinho, João Fausto L. de Oliveira, Fábio Usberti, Marcos de Almeida Leone Filho, Attilio Converti, and et al. 2020. "Neural-Based Ensembles and Unorganized Machines to Predict Streamflow Series from Hydroelectric Plants" Energies 13, no. 18: 4769. https://doi.org/10.3390/en13184769

APA Style

Belotti, J., Siqueira, H., Araujo, L., Stevan, S. L., Jr., de Mattos Neto, P. S. G., Marinho, M. H. N., de Oliveira, J. F. L., Usberti, F., Leone Filho, M. d. A., Converti, A., & Sarubbo, L. A. (2020). Neural-Based Ensembles and Unorganized Machines to Predict Streamflow Series from Hydroelectric Plants. Energies, 13(18), 4769. https://doi.org/10.3390/en13184769

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop