Next Article in Journal
Numerical Modelling of Mechanical Behavior of Coal Mining Hard Roofs in Different Backfill Ratios: A Case Study
Previous Article in Journal / Special Issue
Low-Cost Solar Irradiance Sensing for PV Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Prediction in Photovoltaic Power by Neural Networks †

1
Department of Information Engineering, Electronics and Telecommunications, University of Rome “La Sapienza”, via Eudossiana, 18, Rome 00184, Italy
2
Electrical Engineering Division of Department of Astronautical, Electrical and Energy Engineering, University of Rome “La Sapienza”, via Eudossiana, 18, Rome 00184, Italy
*
Author to whom correspondence should be addressed.
Series for the Prediction in Photovoltaic Power Plants. In Proceedings of the 2016 IEEE 16th International Conference on Environment and Electrical Engineering (EEEIC), Florence, Italy, 7–10 June 2016.
Energies 2017, 10(7), 1003; https://doi.org/10.3390/en10071003
Submission received: 11 April 2017 / Revised: 11 July 2017 / Accepted: 12 July 2017 / Published: 15 July 2017

Abstract

:
The ability to forecast the power produced by renewable energy plants in the short and middle term is a key issue to allow a high-level penetration of the distributed generation into the grid infrastructure. Forecasting energy production is mandatory for dispatching and distribution issues, at the transmission system operator level, as well as the electrical distributor and power system operator levels. In this paper, we present three techniques based on neural and fuzzy neural networks, namely the radial basis function, the adaptive neuro-fuzzy inference system and the higher-order neuro-fuzzy inference system, which are well suited to predict data sequences stemming from real-world applications. The preliminary results concerning the prediction of the power generated by a large-scale photovoltaic plant in Italy confirm the reliability and accuracy of the proposed approaches.

1. Introduction

In recent years, electrical power systems have progressively evolved from centralized bulk systems to decentralized systems in a distributed generation (DG) scenario characterized by smaller generating units connected directly to distribution networks near the consumer [1]. Several economic and environmental reasons, driven by government financial incentives [2], have considerably pushed the widespread adoption of DG in the energy marketplace, especially that from renewable resources (e.g., solar power and wind energy).
However, technical barriers often impede the entrance of DG into the current distribution infrastructure with a significant penetration level due to the intermittent nature of the renewable energy resources that often do not match the energy load demands [3]. Indeed, the transition to an energy economy primarily founded on renewable resources depends on overcoming the difficulties associated with the variability and reliability of these non-dispatchable resources [4]. These issues give rise to substantial critical issues with respect to the usual working habits of the transmission (TSO) and distribution (DSO) system operators, utility companies, as well as power producers: voltage and frequency regulation, islanding detection, harmonic distortion, distribution issues, as well as demand side management of prosumers. Certainly, energy storage systems (EES) could become a valuable response providing specific support services for renewable energy production in order to reduce short-term output fluctuations and, consequently, improve power quality [5].
In this framework, it is of paramount importance to forecast accurately the power output of plants over the next hours or days in order to integrate in an optimal way the DG from non-programmable renewable resources into power systems on a large scale [6]. The ability to forecast the amount of power fed by renewable energy plants into substations is a necessary requirement for the TSO in order to optimize short- and middle-term decisions, such as corrections to out-of-region trade and unit commitments [7]. Besides, accurate short-term and intra-hour forecasting are required for regulatory, dispatching and load-following issues [8], while power system operators are more sensitive to intra-day forecasts [9], especially when handling multiple load zones, since they avoid possible penalties that are incurred due to deviations between forecasted and produced energy. Moreover, power system operators are mostly interested in developing methods to be used in the framework of the daily session of the electricity market where the producers must present energy bid offers for the 24 h of the following day [10].
Forecasting can be broadly divided into four major categories [11], namely long-, medium-, short- and very short-term forecast, on the basis of the time horizon’s scale, which ranges from several years in the long-term forecasting to hours in the medium-term forecasting, till arriving at minutes in the very short-term forecasting. When adopted for photovoltaic (PV) power plants, forecasting techniques can be applied directly to the produced power (direct methods) or indirectly to the solar irradiation, which is the main factor related to the output power that is calculated in a second successive step (indirect methods) [12]. Both cases share similar techniques that can be broadly divided [11] into persistence methods, physical techniques, statistical techniques and hybrid models. Usually, physical methods try to obtain an accurate forecast using a white box approach, that is to say physical parametric equations. On the other hand, statistical or probabilistic methods try to predict the relevant quantities extracting dominant relations from past data that can be used to predict the future behavior. In addition, both methods can be properly mixed, originating the so-called “gray box” approaches [13].
In the prediction of time series, past observations are used to develop a method able to describe underlying relationships among data. Therefore, a mathematical model is adopted to extrapolate future values, especially for those contexts where missing and incomplete data can limit the ability to gain knowledge about the underlying, unknown process. Complexity and dynamics of real-world problems require advanced methods and tools able to use past samples of the sequence in order to make an accurate prediction [14]. Additionally, the problem of forecasting future values of a time series is often mandatory for the cost-effective management of available resources. These are well-known hard problems [6], given the non-stationarity and, often, the non-linearity of time series, which result in a complex dynamics that is hard to model adequately by using standard predictive models.
For many years, regressive statistical approaches [15] have been considered for prediction; among them, moving average (MA), auto regressive (AR), auto regressive moving average (ARMA), auto regressive integrated moving average (ARIMA) and autoregressive integrated moving average with exogenous variables (ARIMAX) are typical examples of these approaches. Unfortunately, standard structural models provide a poor representation of actual data and therefore result in poor accuracy when used for forecasting. Consequently, many worldwide research activities intend to improve the accuracy of prediction models. In this regard, computational intelligence is considered as one of the most fruitful approaches for prediction [16]. Several forecasting methods, with different mathematical backgrounds, such as fuzzy predictors, artificial neural networks (ANN), evolutionary and genetic algorithms and support vector machines, have been considered [17]. Nevertheless, dealing with noisy and missing training examples and the lack of robustness against outliers are still open problems.
In this paper, our attention is mainly concentrated on short-term forecasting with a time horizon of one day. Nevertheless, the proposed techniques can be applied to shorter or longer time horizons, yet always remaining in a short-term forecast. A huge amount of literature is available on short-term forecasting models applied to PV plants, even if most of them deal with irradiation forecasting models: while not claiming to be exhaustive, the proposed techniques include mainly statistical models [18], stochastic predictors [19], fuzzy systems [20], ANN [21], ANN in conjunction with statistical feature parameters [22] and artificial intelligence (AI)-based techniques [23]. Some papers present also the comparison between the predictions obtained through different models based on two or more forecasting techniques [24]. Despite the large amount of literature on forecasting techniques, only a few papers describe forecasting models to be used for directly predicting the daily energy production of the PV plant. These few papers use general techniques appropriately applied to the day-ahead horizon. Starting from the input of past power measurements and meteorological forecasts of solar irradiance, the power output is estimated by means of ANN [25] or soft computing [26], while in [27], the ANN is applied only on past energy production values. Additionally, the ANN can be used in conjunction with numerical weather prediction (NWP) models [28], which can be based on satellite and land-based sky imaging [29]. NWP models have been also used to build an outperforming multi-model ensemble (MME) for day-ahead prediction of PV power generation [30]. More complex schemes are proposed in [31], where an ANN is used to improve the performance of baseline prediction models, i.e., a physical deterministic model based on cloud tracking techniques, an ARMA model and a k-nearest neighbor (kNN) model, and in [32], where a Kalman filter is used in conjunction with a state-space model (SSM).
The use of three inference systems for data regression based on neural and fuzzy neural networks is proposed here. The main distinguishing idea of our work is that we properly tailor new learning approaches to analyze data associated with renewable power sources starting from available techniques based on computational intelligence that have proven to produce very accurate predictions in several fields of application, like for example, risk management, biomedical engineering and financial forecasting [33], but also to analyze accurately the data associated with renewable power sources [34]. Actually, the stochastic and often chaotic behavior of time series in the prediction of output power of PV plants calls for ad hoc procedures to forecast power time series, which are here developed working only on the past measurements of electric quantities, as initially tested in our previous work [35].
We introduce a prediction problem as a two-fold process. First, the observed time series is embedded in order to select, among the available samples, the ones to be used to predict the next one. Then, such samples are used to feed three different function approximation models and techniques based on neural and fuzzy neural networks, which are suited to predict data sequences that often show a chaotic behavior: radial basis function (RBF) [36]; the adaptive neuro-fuzzy inference system (ANFIS) [37]; the higher-order neuro-fuzzy inference system (HONFIS) [38]. As explained in the body of the paper, a predictor can be based on such function approximation models, whose parameters are estimated by data-driven learning procedures.
The rest of the paper is organized as follows. In Section 2, we introduce the function approximation models that are used for time series prediction. The application is ascertained by several benchmark results and extensive computer simulations reported in Section 3 and discussed in Section 4 proving the validity of the proposed learning procedure and showing a favorable comparison with other well-known prediction techniques. Finally, our conclusions are drawn in Section 5.

2. Materials and Methods

There is a huge amount of literature pertaining to time series prediction techniques that rely on neural and fuzzy neural networks [39], which are actually of high importance in this field since they can be used to solve data regression problems [40]. A sequence S ( t ) can be predicted by considering a function approximation y = f ( x ̲ ) , f : I R N I R . For instance, by using linear models each input vector x ̲ t is made of N subsequent samples of S ( t ) , and the output y t is the sample to be predicted:
x ̲ t = S ( t ) S ( t 1 ) S ( t N + 1 ) , y t = S ( t + m ) , f lin ( x ̲ t ) = j = 1 N λ j x t j
Then, we obtain:
S ˜ ( t + m ) = j = 1 N λ j S ( t j + 1 )
where S ˜ ( t + m ) is the prediction of the true value S ( t + m ) at the time distance m. Considering the statistical properties of S ( t ) , as the autocorrelation function, it is possible to determine the parameters λ j , j = 1 N , of the function f lin ( · ) .
Generally speaking, input vectors x ̲ t are obtained through the so-called “embedding technique”, which makes use of previous samples of S ( t ) to build the vectors themselves. There are two parameters to be set in this regard, the embedding dimension D and the time lag T, resulting in the following embedding structure:
x ̲ t = S ( t ) S ( t T ) S ( t 2 T ) S ( t ( D 1 ) T )
These two parameters are estimated in two different ways, by using the false nearest neighbors (FNN) algorithm for the embedding dimension and the average mutual information (AMI) criterion for the time lag [41]. In the following part of the paper, D and T will be chosen by such algorithms via the VRA software [42].
The performances of a predictor that relies on a linear approximation model, as the one in Equation (1), are very poor when it is applied to data sequences in real environments. They often have noisy and chaotic components that force one to wisely choose the embedding parameters, as well as the function approximation model by using appropriate procedures [43]. In this context, the underlying, unknown system is observed through S ( t ) , and its state-space evolution is obtained by the trajectories of embedded vectors.
Overall, the estimated sample S ˜ ( t + m ) predicted at a distance m will be:
S ˜ ( t + m ) = f x ̲ t
where f ( · ) is the regression model to be determined. Thus, the function f ( · ) will approximate the link from the reconstructed state x ̲ t to the corresponding output y t [41]. It is important to note that f ( · ) must be non-linear since the considered system has a complex behavior. In the following, we will consider the usual case of a “one-step-ahead” prediction, that is m = 1 .
By driving the function approximation model with embedding data, neural networks and fuzzy logic are very suited to solving a prediction problem dealing with real-world sequences. In this paper, three models of this kind are proposed for the study of the considered photovoltaic time series. They are introduced in the following, considering as a baseline for benchmarking in the successive tests: the linear predictor reported in Equation (1), whose parameters are estimated through a common least-squares estimator (LSE) [44]; the well-known ARIMA model, which handles non-stationarity and seasonality and whose parameters are determined by a maximum likelihood approach [45]. Both of these models will be estimated by using lagged samples at the input, as determined by the embedding procedure, similarly to the adopted neural networks.

2.1. RBF

An RBF neural network is used to build up a function approximation model having the following structure:
f ( x ̲ ) = i = 1 M λ i ϕ ( x ̲ c ̲ i )
where x ̲ R N is the input vector, ϕ ( · ) is a radial basis function centered in c ̲ i and weighted by an appropriate coefficient λ i . The choice of ϕ ( · ) and c ̲ i must be considered for the approximation capability of the network. Commonly-used types of radial basis functions include:
  • Gaussian:
    ϕ ( r ) = e ( ϵ r ) 2
  • Multiquadric:
    ϕ ( r ) = 1 + ( ϵ r ) 2
  • Inverse quadratic:
    ϕ ( r ) = 1 1 + ( ϵ r ) 2
Several methods can be used to minimize the error between desired output and model output and, hence, to identify the parameters c ̲ i and λ i [46].

2.2. ANFIS

An ANFIS neural network implements a fuzzy inference system to approximate the function y = f ( x ̲ ) , f : I R N I R . It is composed of M rules of the Sugeno first-order type, where the k-th rule, k = 1 M , is:
If x 1 is B 1 ( k ) , , and x N is B N ( k ) then y ( k ) = j = 1 N a j ( k ) x j + a 0 ( k )
where x ̲ = x 1 x 2 x N is the input to the network and y ( k ) is the output of the rule. The antecedent part of the rule depends on the Membership Functions (MFs) μ B j ( k ) ( x j ) of the fuzzy input variables B j ( k ) , j = 1 N ; the consequent part is determined by the coefficients a j ( k ) , j = 0 N , of the crisp output y ( k ) . By using standard options for composing the input MFs and combining the rule outputs [40], the output of the ANFIS network is represented by:
y ˜ = k = 1 M μ B ̲ ( k ) ( x ̲ ) y ( k ) k = 1 M μ B ̲ ( k ) ( x ̲ )
where y ˜ is the estimation of y and μ B ̲ ( k ) ( x ̲ ) is the composed MF of the k-th rule.
The ANFIS learning of network coefficients is based on a classical back-propagation technique associated with a least-squares estimation [47].

2.3. HONFIS

The HONFIS model adopted for data regression is a Takagi–Sugeno (TS) fuzzy inference system that is a generalization of ANFIS where the consequent part in Equation (9) is a polynomial of order greater than one combining the input values x j , j = 1 N . It is made of M different fuzzy rules, and the coefficients are obtained through the use of a clustering procedure in the joint input-output space [48].
Let Γ = Γ 1 , Γ 2 , , Γ M be a set of M clusters (each associated with a rule output), and let every pattern of the training set be assigned randomly to one of these clusters. Then, the clustering procedure is fundamentally an alternating optimization technique that aims at identifying the M cluster prototypes [44]. At the end of the iterations, a label q, 1 q M , is associated with each pattern, representing the rule it has been assigned to during the last iteration. In this way, a classification model able to assign a fuzzy label L ( x ̲ ) to any pattern x ̲ of the input space is obtained:
L ( x ̲ ) = μ B ̲ ( 1 ) ( x ̲ ) μ B ̲ ( 2 ) ( x ̲ ) μ B ̲ ( M ) ( x ̲ )
where the k-th element of L ( x ̲ ) represents the fuzzy membership of the pattern to the k-th class, that is the firing strength μ B ̲ ( k ) ( x ̲ ) in Equation (10).
In the following, we will adopt a K-nearest neighbor (K-NN) strategy for classification of the fuzzy label during the prediction tests. This is done by using the input value x ̲ only, and hence, the fuzzy label of x ̲ will be:
L ( x ̲ ) = 1 K q = 1 K L ( x ̲ t q )
where x ̲ t 1 , x ̲ t 2 , , x ̲ t K are the K patterns of the training set that score the smallest Euclidean distance from x ̲ , with their fuzzy labels determined during the training phase.
The output y ˜ is calculated through Equation (10) by means of the firing strengths contained in the fuzzy label L ( x ̲ ) and the output consequents whose parameters have been previously determined by clustering in the joint input-output space. For instance, in the case of a 3rd order polynomial as adopted in the following, the k-th rule’s consequent, k = 1 M , is:
y ( k ) = a 31 ( k ) x 1 3 + + a 3 N ( k ) x N 3 + a 21 ( k ) x 1 2 + + a 2 N ( k ) x N 2 + a 11 ( k ) x 1 + + a 1 N ( k ) x N + a 0 ( k )
Regarding the generalization capability, the optimal number of rules is automatically determined on the basis of learning theory [49]. The best model is chosen by evaluating a cost function based on network complexity and approximation error [50]. As a measure of the network performance on the training set, the mean squared error (MSE) is adopted:
E = 1 N T t = 1 N T y t y ˜ t 2
where N T is the number of samples in the training set to be predicted and y ˜ t is the output generated by HONFIS in correspondence with the t-th input pattern x ̲ t . The optimal network is selected by minimizing the following cost function that depends on the maximum allowed number of the HONFIS rules:
F ( M ) = ( 1 λ ) E ( M ) E min E max E min + λ M N T
where E min and E max are the extreme values of the performance E that are encountered during the analysis of the different models, while λ is a non-critical weight, 0 λ 1 , set to 0.5 by default.

3. Results

The prediction performances of the considered models were investigated by means of several simulation tests carried out using data from the De Nittis Power Plant in the Apulia Region, Italy (latitude ϕ = 41 26 16 , longitude λ = 15 45 47 ). Data were relative to a single photovoltaic plant, organized with eight cabins with two inverters each. The output current was used as the variable to be predicted. The value was sampled at the source with a 5-min sample interval, and it was collected from 6:00 a.m. to 10:55 p.m., resulting in 204 samples per day. The used data stream comes from a single inverter of a single cabin, and it is relative to the year 2012. In order to provide a statistical characterization of the handled time series, whose samples are measured in Amperes, we have computed the first four statistical moments of the whole 2012 dataset, whose histogram is also reported in Figure 1: mean 85.1291; variance 1.0244 × 10 4 ; skewness 0.8151; kurtosis 2.1082. Successively, the whole time series has been normalized linearly between zero and one in order to cope with the numerical requirements of learning algorithms, especially for neural network models.
All of the computational models were trained using time series subsampled at an interval of 1 h, thus resulting in 17 samples per day. Four different kinds of training conditions were considered: 1-, 3-, 7- and 30-day, associated with training sets composed of 17, 51, 119 and 510 samples, respectively. Every computational model estimated by using one of these training sets was applied to test the day following the latest one in the training set, that is on a test set of 17 samples. The successive experiments would consider for testing one day for each month of 2012, and we have chosen the 15th of each month for the sake of uniformity (the last 15 samples of 2011 are also used for the 30-day training of 15 January). For the same reason, after the preliminary analysis we carried out for determining the embedding parameters, as the latter did not show a considerable sensitivity to seasonality and to the length of the training set, which was relatively small with respect to the usual length of time series processed by the AMI and FNN methods, we had adopted as the optimal choice the values T = 5 and D = 3 for every training set and the related experiment. We remark that all of the test sets had different irradiation and meteorological conditions, and thus, they could be considered as a good ensemble for representing the typical behaviors that might be encountered.
All of the experiments had been performed using MATLAB® 2016b, running on a 3.1 -GHz Intel Core i7 platform equipped with 16 GB of memory. As previously mentioned, in addition to the three proposed neural and fuzzy neural models, the linear (LSE) and ARIMA predictors were adopted for benchmarking. LSE did not have parameters to be set in advance, while for ARIMA, RBF and ANFIS, we had adopted the default options provided by the software platform for training and model regularization (ARIMA, RBF and ANFIS models were trained by using the supported functions in the econometrics toolbox, neural network toolbox and fuzzy logic toolbox of MATLAB, respectively). Regarding HONFIS, a preliminary analysis had been performed in order to evaluate the best model order for the rule consequents. In the following, the results of the third-order model are reported, as it was able to obtain the best performance in almost all of the considered cases. The input space classification was obtained by a three-NN classifier (i.e., K = 3 ). The optimal number of rules was determined by Equation (15), using λ = 0.5 and varying M from 1–50% of N T .
The prediction performances were measured by two metrics commonly adopted for energy time series, which were independent of the said procedure for data normalization:
  • Normalized Mean Square Error (NMSE):
    NMSE = t = 1 N T y t y ˜ t 2 t = 1 N T y t y ¯ 2
    where y ¯ is the average value of samples y t in the test set;
  • Mean Absolute Range Error (MARE):
    MARE = 1 N T t = 1 N T y t y ˜ t y max y min
    where y max and y min are the maximum and minimum value of samples y t in the test set, respectively.
The numerical results for each tested day are reported in Table 1, Table 2, Table 3, Table 4, Table 5, Table 6, Table 7, Table 8, Table 9, Table 10, Table 11 and Table 12, where the performance of the best model is marked in bold font. As per the following discussion, being HONFIS the model that assures the best performance in most of the situations, the graphical illustration of the actual time series (blue line) and the one predicted by HONFIS (red line), over the four training conditions, is reported in Figure 2, Figure 3, Figure 4, Figure 5, Figure 6, Figure 7, Figure 8, Figure 9, Figure 10, Figure 11, Figure 12 and Figure 13, respectively. In the x-axis is reported the cumulative index of samples of the considered day, starting from Index 1 for the first sample of 1 January 2012 (which is a leap year) and considering 17 samples per day. The output current reported in the y-axis is a dimensionless value between zero and one, as the whole dataset has been normalized before the model processing. Some negative values may occur because of possible numerical issues of a trained model when its performance is inadequate.
It is important to remark that the number of samples that should be predicted in a test set is strictly related to the computational cost of the learning procedure, as well as to the desired horizon of predictability, for which it is necessary to follow a given strategy for the energy management and to broadcast the relevant information to the involved operators. For instance, as the learning times of the previous models were in the order of some minutes and the considered time series are subsampled at 1 h, any model could be re-trained on a hourly basis for a new sample to be predicted (i.e., using a test set with one sample only), thus incorporating the new information given by the the latest available sample that was measured.
However, in order to ensure a wide range of action, it is worth forecasting more than one sample at a time, as previously done considering an entire day. Although longer test sets might not be so useful and accurate, we consider in the following also weekly test sets, so as to report the prediction results for more days in a year and to evaluate the sensitivity of the proposed approach not only to different training, but also to different test conditions. The numerical results for each tested week are reported in Table 13, Table 14, Table 15, Table 16, Table 17, Table 18, Table 19, Table 20, Table 21, Table 22, Table 23 and Table 24. In such a case, we have considered seven-day or 30-day training sets only, that is longer than the test set or of the same length. As the starting day in the test set is still the 15th of each month, the training results are the same as the ones reported in Table 1, Table 2, Table 3, Table 4, Table 5, Table 6, Table 7, Table 8, Table 9, Table 10, Table 11 and Table 12, respectively.

4. Discussion

Considering the experimental results reported in this paper, it can be noted that all of the proposed approaches are suitable for the prediction of the considered time series, depending on the chosen training set. The numerical results in terms of either NMSE or MARE basically agree, and they are coherent with the graphical behaviors, reported in the figures, between actual and predicted time series. Note that a perfect match is obtained when the NMSE is smaller than about 5 × 10 4 .
In the days with the most stable meteorological conditions (i.e., 15 March or 15 June), the prediction results for daily test sets are better than the others. Furthermore, there are differences among the results associated with diverse training procedures. The tests with one-day training sets have a high variability for all of the algorithms because the prediction is highly affected by the difference in irradiation between the day of the training and the day of the test. In fact, when the meteorological condition varies greatly from the training day to the sequent test day, the algorithms are trained on a set that is almost not correlated with the test set. It can also be noted that when the irradiation is very similar (no clouds, good weather), the performance is very good also for the one-day training set.
A similar discussion can be made, with smaller variability, for the three-day training sets. It has to be noted that we have inserted the one-day and three-day training sets mainly to show the sensitivity of the models to different training conditions. The results for seven-day training sets are much more stable for all of the algorithms and the days. The results for 30-day training sets are stable as well, but the sequences are much longer; thus, if the days are variable in terms of meteorological conditions, the model has a worse performance because the training process is more difficult. Training sets longer than 30 days were also considered in some preliminary tests, and the results confirm such a trend for which the performance decreases progressively as much as the length of the training set increases.
Overall, a one-day training is based on a very small number of samples in the training set, which is interesting for the computational cost of the learning procedure but it is an issue when many model parameters must be estimated, which is the well-known curse of dimensionality for neural networks. In fact, the third-order polynomial adopted for HONFIS makes it the most complex model among the ones considered in this paper, and its performances improve with respect to the other models as much as the number of samples increases in the training set, although a larger training set may not be the optimal choice.
Although of relative usefulness, the numerical results for weekly test sets are similar to the previous ones. In the case of seven-day training sets, all of the neural models suffer from the curse of dimensionality, as the information given by a training set of the same length of the test set is too small to ensure a robust estimation of the model parameters. In such cases, the LSE predictor yields better performances than the others models albeit too shallow in absolute terms. In the case of 30-day training sets, HONFIS is able to obtain the best results for weekly test sets, with a numerical score in terms of both NMSE and MARE that is stable enough with respect to shorter test sets of one day only.
The performance of the prediction is highly affected by the intrinsic seasonality of the time series considered. Anyway, HONFIS achieves the best results for almost all of the training sets with respect to the other proposed neural models, and all of them outperform both the LSE benchmark and the ARIMA approach, which is not a feasible solution in this case mainly for the intrinsic chaotic properties of the sequence. This reinforces the fact that the prediction of photovoltaic production is a promising field for the application of neural and fuzzy neural approaches along with the use of a suitable embedding procedure.

5. Conclusions

In this paper, we presented new embedding approaches based on neural and fuzzy neural networks that have been properly tailored to be efficiently applied to PV time series prediction. The models were applied to the time series of sampled output current of a test PV plant, predicting a single day of operation. The models were trained using different training sets, namely 1-day, 3-day, 7-day or 30-day.
The obtained results demonstrate that in normal operation conditions, all of the solutions are suitable for the proposed goal, except for the shorter training sets, which result in less reliable performances. In particular, the approaches were shown to be very accurate when considering a seven-day training set to forecast the successive day, mostly because in a single-week window, the solar irradiation is more stable.
The results are very promising and suggest several opportunities for future work. We could make use of detrending techniques, such as mean-reverting approaches, in order to remove seasonal differences and spike outliers, as well as to improve the training accuracy. Additionally, it could be useful to test distributed learning approaches [51], by which the results could be improved sharing the data from different cabins of the same plant or from different nearby plants.

Author Contributions

Massimo Panella and Rodolfo Araneo conceived of the theoretical approach and designed the experiments. Antonello Rosato and Rosa Altilio performed the experiments and analyzed the data, after having organized the materials and set up the analysis tools. All of the authors wrote the paper.

Conflicts of Interest

The authors declare no conflict of interest.

Nomenclature

The following main nomenclature is adopted in the paper:
S ( t ) :actual sample of the energy time series (output current) at time t.
S ˜ ( t ) :predicted sample of the energy time series at time t.
f ( x ̲ ) :parametric model used to solve the prediction problem by data regression.
x ̲ t :embedded vector feeding the input of f ( x ̲ ) that estimates the predicted sample S ˜ ( t ) .
y t :output at f ( x ̲ ) representing the predicted sample S ˜ ( t ) .
D:embedding dimension, that is the number of samples of the time series in the embedded vector x ̲ t .
T:time lag between the samples of the time series that are embedded in x ̲ t .
LSE:linear regression model for f ( x ̲ ) .
ARIMA:Auto regressive integrated moving average regression model for f ( x ̲ ) .
RBF:Radial basis function regression model for f ( x ̲ ) .
ANFIS:Adaptive neuro-fuzzy inference system regression model for f ( x ̲ ) .
HONFIS:Higher-order neuro-fuzzy inference system regression model for f ( x ̲ ) .
NMSE:Normalized mean square error measure.
MARE:Mean absolute range error measure.

References

  1. Lopes, J.P.; Hatziargyriou, N.; Mutale, J.; Djapic, P.; Jenkins, N. Integrating distributed generation into electric power systems: A review of drivers, challenges and opportunities. Electr. Power Syst. Res. 2007, 77, 1189–1203. [Google Scholar] [CrossRef]
  2. Dio, V.D.; Favuzza, S.; Cascia, D.L.; Massaro, F.; Zizzo, G. Critical assessment of support for the evolution of photovoltaics and feed-in tariff(s) in Italy. Sustain. Energy Technol. Assess. 2015, 9, 95–104. [Google Scholar] [CrossRef]
  3. Du, Y.; Lu, D.D.C.; James, G.; Cornforth, D.J. Modeling and analysis of current harmonic distortion from grid connected PV inverters under different operating conditions. Sol. Energy 2013, 94, 182–194. [Google Scholar] [CrossRef]
  4. Samadi, A.; Eriksson, R.; der, L.S.O.; Rawn, B.G.; Boemer, J.C. Coordinated active power-dependent voltage regulation in distribution grids with PV systems. IEEE Trans. Power Del. 2014, 29, 1454–1464. [Google Scholar] [CrossRef]
  5. Han, X.; Liao, S.; Ai, X.; Yao, W.; Wen, J. Determining the minimal power capacity of energy storage to accommodate renewable generation. Energies 2017, 10, 468. [Google Scholar] [CrossRef]
  6. Inman, R.H.; Pedro, H.T.; Coimbra, C.F. Solar forecasting methods for renewable energy integration. Prog. Energy Combust. Sci. 2013, 39, 535–576. [Google Scholar] [CrossRef]
  7. Garrigle, E.M.; Leahy, P. Quantifying the value of improved wind energy forecasts in a pool-based electricity market. Renew. Energy 2015, 80, 517–524. [Google Scholar] [CrossRef]
  8. Liu, N.; Tang, Q.; Zhang, J.; Fan, W.; Liu, J. A hybrid forecasting model with parameter optimization for short-term load forecasting of micro-grids. Appl. Energy 2014, 129, 336–345. [Google Scholar] [CrossRef]
  9. Kardakos, E.G.; Alexiadis, M.C.; Vagropoulos, S.I.; Simoglou, C.K.; Biskas, P.N.; Bakirtzis, A.G. Application of time series and artificial neural network models in short-term forecasting of PV power generation. In Proceedings of the 2013 48th International Universities’ Power Engineering Conference (UPEC), Dublin, Ireland, 2–5 September 2013; pp. 1–6. [Google Scholar]
  10. Clements, A.; Hurn, A.; Li, Z. Strategic bidding and rebidding in electricity markets. Energy Econ. 2016, 59, 24–36. [Google Scholar] [CrossRef]
  11. Raza, M.Q.; Nadarajah, M.; Ekanayake, C. On recent advances in PV output power forecast. Sol. Energy 2016, 136, 125–144. [Google Scholar] [CrossRef]
  12. Antonanzas, J.; Osorio, N.; Escobar, R.; Urraca, R.; de Pison, F.M.; Antonanzas-Torres, F. Review of photovoltaic power forecasting. Sol. Energy 2016, 136, 78–111. [Google Scholar] [CrossRef]
  13. Al-Messabi, N.; Goh, C.; Li, Y. Heuristic grey-box modelling for photovoltaic power systems. Syst. Sci. Control Eng. 2016, 4, 235–246. [Google Scholar] [CrossRef]
  14. Coyle, D.; Prasad, G.; McGinnity, T.M. A time-series prediction approach for feature extraction in a brain-computer interface. IEEE Trans. Neural Syst. Rehabil. Eng. 2005, 13, 461–467. [Google Scholar] [CrossRef] [PubMed]
  15. Bennett, C.; Stewart, R.A.; Lu, J. Autoregressive with exogenous variables and Neural Network short-term load forecast models for residential low voltage distribution networks. Energies 2014, 7, 2938–2960. [Google Scholar] [CrossRef]
  16. Tzafestas, S.; Tzafestas, E. Computational intelligence techniques for short-term electric load forecasting. J. Intell. Robot. Syst. 2001, 31, 7–68. [Google Scholar] [CrossRef]
  17. Proietti, A.; Panella, M.; Leccese, F.; Svezia, E. Dust detection and analysis in museum environment based on pattern recognition. Measurement 2015, 66, 62–72. [Google Scholar] [CrossRef]
  18. Safi, S.; Zeroual, A.; Hassani, M. Prediction of global daily solar radiation using higher order statistics. Renew. Energy 2002, 27, 647–666. [Google Scholar] [CrossRef]
  19. Kaplanis, S.; Kaplani, E. Stochastic prediction of hourly global solar radiation for Patra, Greece. Appl. Energy 2010, 87, 3748–3758. [Google Scholar] [CrossRef]
  20. Iqdour, R.; Zeroual, A. Prediction of daily global solar radiation using fuzzy systems. Int. J. Sustain. Energy 2007, 26, 19–29. [Google Scholar] [CrossRef]
  21. Voyant, C.; Muselli, M.; Paoli, C.; Nivet, M.L. Optimization of an artificial neural network dedicated to the multivariate forecasting of daily global radiation. Energy 2011, 36, 348–359. [Google Scholar] [CrossRef]
  22. Wang, F.; Mi, Z.; Su, S.; Zhao, H. Short-term solar irradiance forecasting model based on artificial Neural Network using statistical feature parameters. Energies 2012, 5, 1355–1370. [Google Scholar] [CrossRef]
  23. Mellit, A.; Kalogirou, S.A. Artificial intelligence techniques for photovoltaic applications: A review. Prog. Energy Combust. Sci. 2008, 34, 574–632. [Google Scholar] [CrossRef]
  24. Hocaoğlu, F.O.; Gerek, O.N.; Kurban, M. Hourly solar radiation forecasting using optimal coefficient 2-D linear filters and feed-forward neural networks. Sol. Energy 2008, 82, 714–726. [Google Scholar] [CrossRef]
  25. Chen, C.; Duan, S.; Cai, T.; Liu, B. Online 24-h solar power forecasting based on weather type classification using artificial neural network. Sol. Energy 2011, 85, 2856–2870. [Google Scholar] [CrossRef]
  26. Monteiro, C.; Fernandez-Jimenez, L.A.; Ramirez-Rosado, I.J.; Muñoz-Jimenez, A.; Lara-Santillan, P.M. Short-term forecasting models for photovoltaic plants: analytical versus soft-computing techniques. Math. Probl. Eng. 2013, 2013, 767284. [Google Scholar] [CrossRef]
  27. Izgi, E.; Öztopal, A.; Yerli, B.; Kaymak, M.K.; Şahin, A.D. Short-mid-term solar power prediction by using artificial neural networks. Sol. Energy 2012, 86, 725–733. [Google Scholar] [CrossRef]
  28. Fernandez-Jimenez, L.A.; Munoz-Jimenez, A.; Falces, A.; Mendoza-Villena, M.; Garcia-Garrido, E.; Lara-Santillan, P.M.; Zorzano-Alba, E.; Zorzano-Santamaria, P.J. Short-term power forecasting system for photovoltaic plants. Renew. Energy 2012, 44, 311–317. [Google Scholar] [CrossRef]
  29. Barbieri, F.; Rajakaruna, S.; Ghosh, A. Very short-term photovoltaic power forecasting with cloud modeling: A review. Renew. Sustain. Energy Rev. 2017, 75, 242–263. [Google Scholar] [CrossRef]
  30. Pierro, M.; Bucci, F.; Felice, M.D.; Maggioni, E.; Moser, D.; Perotto, A.; Spada, F.; Cornaro, C. Multi-Model Ensemble for day ahead prediction of photovoltaic power generation. Sol. Energy 2016, 134, 132–146. [Google Scholar] [CrossRef]
  31. Chu, Y.; Urquhart, B.; Gohari, S.M.; Pedro, H.T.; Kleissl, J.; Coimbra, C.F. Short-term reforecasting of power output from a 48 MWe solar PV plant. Sol. Energy 2015, 112, 68–77. [Google Scholar] [CrossRef]
  32. Takeda, H. Short-term ensemble forecast for purchased photovoltaic generation. Sol. Energy 2017, 149, 176–187. [Google Scholar] [CrossRef]
  33. Panella, M. Advances in biological time series prediction by neural networks. Biomed. Signal Process. Control 2011, 6, 112–120. [Google Scholar] [CrossRef]
  34. Mirowski, P.; Chen, S.; Ho, T.K.; Yu, C.N. Demand forecasting in smart grids. Bell Labs Tech. J. 2014, 18, 135–158. [Google Scholar] [CrossRef]
  35. Rosato, A.; Altilio, R.; Araneo, R.; Panella, M. Embedding of time series for the prediction in photovoltaic power plants. In Proceedings of the IEEE 16th International Conference on Environment and Electrical Engineering (EEEIC), Florence, Italy, 7–10 June 2016; pp. 1–4. [Google Scholar]
  36. Ko, C.N.; Lee, C.M. Short-term load forecasting using SVR (support vector regression)-based radial basis function neural network with dual extended Kalman filter. Energy 2013, 49, 413–422. [Google Scholar] [CrossRef]
  37. Osório, G.; Matias, J.; Ao, J.C. Short-term wind power forecasting using adaptive neuro-fuzzy inference system combined with evolutionary particle swarm optimization, wavelet transform and mutual information. Renew. Energy 2015, 75, 301–307. [Google Scholar] [CrossRef]
  38. Castro, J.R.; Castillo, O.; Sanchez, M.A.; Mendoza, O.; Rodríguez-Diaz, A.; Melin, P. Method for higher order polynomial Sugeno Fuzzy inference systems. Inf. Sci. 2016, 351, 76–89. [Google Scholar] [CrossRef]
  39. Masulli, F.; Studer, L. Time Series Forecasting and Neural Networks. In Proceedings of the International Joint Conference on Neural Networks, Washington, DC, USA, 10–16 July 1999. [Google Scholar]
  40. Jang, J.; Sun, C.; Mizutani, E. Neuro-Fuzzy and Soft Computing: A Computational Approach to Learning and Machine Intelligence; Prentice-Hall: Upper Saddle River, NJ, USA, 1997. [Google Scholar]
  41. Abarbanel, H. Analysis of Observed Chaotic Data; Springer: New York, NY, USA, 1996. [Google Scholar]
  42. Kononov, E. Visual Recurrence Analysis (VRA 4.9). Available online: http://visual-recurrence-analysis.software.informer.com/4.9/ (accessed on 15 July 2017).
  43. Panella, M.; Rizzi, A.; Mascioli, F.M.F.; Martinelli, G. ANFIS synthesis by hyperplane clustering. In Proceedings of the Joint 9th IFSA World Congress and 20th NAFIPS International Conference, Vancouver, BC, Canada, 25–28 July 2001; Volume 1, pp. 340–345. [Google Scholar]
  44. Panella, M.; Liparulo, L.; Proietti, A. A Higher-Order Fuzzy Neural Network for Modeling Financial Time Series. In Proceedings of the IEEE International Joint Conference on Neural Networks (IJCNN), Beijing, China, 6–11 July 2014; pp. 3066–3073. [Google Scholar]
  45. Box, G.; Jenkins, G.; Reinsel, G. Time Series Analysis: Forecasting and Control, 3rd ed.; Englewood Cliffs: Prentice Hall, NJ, USA, 1994. [Google Scholar]
  46. Broomhead, D.S.; Lowe, D. Multivariable Functional Interpolation and Adaptive Networks. Complex Syst. 1988, 2, 321–355. [Google Scholar]
  47. Jang, J.S. ANFIS: Adaptive-Network-Based Fuzzy Inference System. IEEE Trans. Syst. Man Cybern. 1993, 23, 665–685. [Google Scholar] [CrossRef]
  48. Panella, M. A Hierarchical Procedure for the Synthesis of ANFIS Networks. Adv. Fuzzy Syst. 2012, 2012, 1–12. [Google Scholar] [CrossRef]
  49. Sharma, A.; Podolsky, R.; Zhao, J.; McIndoe, R.A. A modified hyperplane clustering algorithm allows for efficient and accurate clustering of extremely large datasets. Bioinformatics 2009, 25, 1152–1157. [Google Scholar] [CrossRef] [PubMed]
  50. Haykin, S. Neural Networks, a Comprehensive Foundation, 2nd ed.; Prentice-Hall: Englewood Cliffs, NJ, USA, 1999. [Google Scholar]
  51. Scardapane, S.; Fierimonte, R.; Wang, D.; Panella, M.; Uncini, A. Distributed music classification using Random Vector Functional-Link nets. In Proceedings of the 2015 International Joint Conference on Neural Networks (IJCNN), Killarney, Ireland, 12–17 July 2015; pp. 1–8. [Google Scholar]
Figure 1. Histogram of the output current.
Figure 1. Histogram of the output current.
Energies 10 01003 g001
Figure 2. Prediction using the best model (HONFIS) for 15 January.
Figure 2. Prediction using the best model (HONFIS) for 15 January.
Energies 10 01003 g002
Figure 3. Prediction using the best model (HONFIS) for 15 February.
Figure 3. Prediction using the best model (HONFIS) for 15 February.
Energies 10 01003 g003
Figure 4. Prediction using the best model (HONFIS) for 15 March.
Figure 4. Prediction using the best model (HONFIS) for 15 March.
Energies 10 01003 g004
Figure 5. Prediction using the best model (HONFIS) for 15 April.
Figure 5. Prediction using the best model (HONFIS) for 15 April.
Energies 10 01003 g005
Figure 6. Prediction using the best model (HONFIS) for 15 May.
Figure 6. Prediction using the best model (HONFIS) for 15 May.
Energies 10 01003 g006
Figure 7. Prediction using the best model (HONFIS) for 15 June.
Figure 7. Prediction using the best model (HONFIS) for 15 June.
Energies 10 01003 g007
Figure 8. Prediction using the best model (HONFIS) for 15 July.
Figure 8. Prediction using the best model (HONFIS) for 15 July.
Energies 10 01003 g008
Figure 9. Prediction using the best model (HONFIS) for 15 August.
Figure 9. Prediction using the best model (HONFIS) for 15 August.
Energies 10 01003 g009
Figure 10. Prediction using the best model (HONFIS) for 15 September.
Figure 10. Prediction using the best model (HONFIS) for 15 September.
Energies 10 01003 g010
Figure 11. Prediction using the best model (HONFIS) for 15 October.
Figure 11. Prediction using the best model (HONFIS) for 15 October.
Energies 10 01003 g011
Figure 12. Prediction using the best model (HONFIS) for 15 November.
Figure 12. Prediction using the best model (HONFIS) for 15 November.
Energies 10 01003 g012
Figure 13. Prediction using the best model (HONFIS) for 15 December.
Figure 13. Prediction using the best model (HONFIS) for 15 December.
Energies 10 01003 g013
Table 1. Prediction results for 15 January .
Table 1. Prediction results for 15 January .
Prediction ModelNMSEMARE
1-Day3-Day7-Day30-Day1-Day3-Day7-Day30-Day
LSE  training0.46900.44290.75910.35230.09210.13120.04570.0884
LSE  test0.82300.43940.54630.43470.18430.15470.15420.1544
ARIMA  training1.99290.88430.66050.98520.39680.29100.22320.7345
ARIMA  test1.88520.51940.48380.26910.58220.20100.22220.1597
RBF  training0.11370.13930.18470.29010.05120.06870.06750.0801
RBF  test1.34720.27390.29590.43990.95860.11460.11340.1461
ANFIS  training0.07450.00050.26980.33830.05220.00400.08880.0910
ANFIS  test2.23930.00040.44150.52270.38950.00460.15970.1723
HONFIS (3rd ord.)  training0.01370.00050.00320.06500.01810.00270.00400.0174
HONFIS (3rd ord.)  test1.48200.00080.00090.05110.23730.00410.00440.0212
Table 2. Prediction results for 15 February.
Table 2. Prediction results for 15 February.
Prediction ModelNMSEMARE
1-Day3-Day7-Day30-Day1-Day3-Day7-Day30-Day
LSE  training0.09780.21670.34800.26310.09060.11230.13550.1213
LSE  test0.40720.36900.36530.37860.16810.14110.14260.1527
ARIMA  training0.33540.75231.02691.25550.51940.22540.17860.1139
ARIMA  test0.44800.66531.44070.53480.21170.22570.36120.2266
RBF  training0.00600.01470.12390.14790.02530.02850.07840.0768
RBF  test0.32120.00560.08600.26600.15080.02190.07970.1129
ANFIS  training0.00010.00670.08940.01560.00030.00970.05810.0786
ANFIS  test0.52200.00710.02950.26020.14800.01130.04720.1147
HONFIS (3rd ord.)  training0.00810.00810.00670.02690.00020.01340.00650.0157
HONFIS (3rd ord.)  test0.79610.00520.00680.12350.17330.00940.01060.0644
Table 3. Prediction results for 15 March.
Table 3. Prediction results for 15 March.
Prediction ModelNMSEMARE
1-Day3-Day7-Day30-Day1-Day3-Day7-Day30-Day
LSE  training0.03990.04280.12480.20930.06630.06820.09460.1061
LSE  test0.03930.03880.04120.04630.06570.06510.06730.0720
ARIMA  training0.60480.34982.37120.40850.24170.17520.06920.0688
ARIMA  test0.00910.01900.65370.50650.02920.04370.04260.0687
RBF  training0.00410.00020.03230.12250.01680.00370.04760.0715
RBF  test0.00630.00020.01530.01280.15080.00400.03550.0310
ANFIS  training0.00010.00050.01340.12680.00020.00110.02690.0720
ANFIS  test0.00210.00030.00860.02160.01200.00100.02080.0399
HONFIS (3rd ord.)  training0.00010.00010.00390.00120.00020.00120.00900.0035
HONFIS (3rd ord.)  test0.94370.00020.00620.00030.19850.00130.01290.0035
Table 4. Prediction results for 15 April.
Table 4. Prediction results for 15 April.
Prediction ModelNMSEMARE
1-Day3-Day7-Day30-Day1-Day3-Day7-Day30-Day
LSE  training0.43070.41170.40590.25020.11930.14820.13220.1126
LSE  test0.43560.42050.41150.42520.18820.18510.19160.1981
ARIMA  training1.23683.81050.81700.69610.48450.38250.19610.2525
ARIMA  test1.26580.98810.23920.16910.31130.27670.15430.1216
RBF  training0.15330.13090.11450.12250.09000.08280.06470.0763
RBF  test1.26850.07540.15000.26970.25950.07200.10180.1506
ANFIS  training0.00020.00660.14690.14440.01840.00110.08030.0801
ANFIS  test2.31680.00360.24570.19910.01200.01710.14370.1252
HONFIS (3rd ord.)  training0.00250.00060.00390.00030.00960.00390.00680.0030
HONFIS (3rd ord.)  test1.62480.00040.02030.00030.32500.00370.01720.0016
Table 5. Prediction results for 15 May.
Table 5. Prediction results for 15 May.
Prediction ModelNMSEMARE
1-Day3-Day7-Day30-Day1-Day3-Day7-Day30-Day
LSE  training0.78470.52050.34690.29980.14990.16710.14420.1290
LSE  test0.76780.54370.55860.57140.24790.18960.20060.2058
ARIMA  training6.90471.28110.40890.98080.56940.23330.17650.2757
ARIMA  test1.92631.59201.24290.36560.37160.31310.26940.2795
RBF  training0.21550.19400.13470.14040.07280.10660.09460.0854
RBF  test7.90040.19390.22500.29002.06100.10770.13270.1386
ANFIS  training0.00010.00800.07620.17180.00040.00980.05640.0972
ANFIS  test1.13880.02250.09570.32530.34580.01810.06050.1483
HONFIS (3rd ord.)  training0.01190.02180.00170.00050.01230.00410.00300.0027
HONFIS (3rd ord.)  test1.85550.00020.00030.00020.34270.00090.00160.0029
Table 6. Prediction results for 15 June.
Table 6. Prediction results for 15 June.
Prediction ModelNMSEMARE
1-Day3-Day7-Day30-Day1-Day3-Day7-Day30-Day
LSE  training0.01080.01210.00960.09570.03630.03830.03180.0579
LSE  test0.01190.01180.01180.01370.03760.03750.03720.0361
ARIMA  training0.73541.39230.30830.32400.29340.39220.13140.1667
ARIMA  test0.88850.01150.06700.02050.33970.03270.08310.0473
RBF  training0.00030.00020.00020.03670.00540.00490.00460.0326
RBF  test0.00050.00020.00020.00750.00730.00490.00480.0266
ANFIS  training0.00010.00050.00030.04230.00050.00170.00530.0353
ANFIS  test0.00100.00060.00040.01180.00990.00190.00640.0303
HONFIS (3rd ord.)  training0.00010.00060.00010.00530.00010.00100.00220.0068
HONFIS (3rd ord.)  test0.17630.00040.00060.00040.04820.00080.00120.0045
Table 7. Prediction results for 15 July.
Table 7. Prediction results for 15 July.
Prediction ModelNMSEMARE
1-Day3-Day7-Day30-Day1-Day3-Day7-Day30-Day
LSE  training0.01120.43190.19810.12260.03400.15760.09490.0700
LSE  test0.24710.30710.24320.23990.09050.15970.11170.0960
ARIMA  training1.00522.11910.21110.86710.56240.44360.13720.2310
ARIMA  test0.99540.10420.15410.10410.58200.07860.11700.0788
RBF  training0.00070.17330.04710.05380.00800.08480.03280.0438
RBF  test0.15190.12490.06980.08400.09090.08600.04520.0690
ANFIS  training0.00010.00040.02140.07070.00030.03150.00370.0501
ANFIS  test0.24380.00050.03970.08910.00020.03990.00540.0621
HONFIS (3rd ord.)  training0.00010.00080.00050.00230.00090.00530.00370.0044
HONFIS (3rd ord.)  test2.33580.00050.00110.00040.40130.00590.00540.0035
Table 8. Prediction results for 15 August.
Table 8. Prediction results for 15 August.
Prediction ModelNMSEMARE
1-Day3-Day7-Day30-Day1-Day3-Day7-Day30-Day
LSE  training0.02950.08610.04060.12260.05170.06140.04420.0853
LSE  test0.19730.19410.19630.19270.09420.08680.08460.0984
ARIMA  training0.88430.29230.22070.29900.66980.11630.11790.1566
ARIMA  test0.91240.29680.13160.10810.75460.17860.11550.0747
RBF  training0.00200.01170.00710.07580.01430.02440.01220.0529
RBF  test0.11740.12490.01570.08410.06130.02970.04520.0614
ANFIS  training0.00010.00040.02140.07070.00030.03150.00370.0501
ANFIS  test0.24380.00050.03070.08910.00020.03990.02130.0621
HONFIS (3rd ord.)  training0.00010.00850.00750.09290.01470.00530.01020.0525
HONFIS (3rd ord.)  test0.01070.00050.03500.07010.01770.00590.02360.0448
Table 9. Prediction results for 15 September.
Table 9. Prediction results for 15 September.
Prediction ModelNMSEMARE
1-Day3-Day7-Day30-Day1-Day3-Day7-Day30-Day
LSE  training0.75110.31410.29000.29710.18450.12800.12810.1192
LSE  test0.75260.72630.72650.72620.21810.20880.20990.2114
ARIMA  training1.33691.38750.97250.35710.44240.24830.24230.1741
ARIMA  test0.42370.66390.52730.34270.33540.23340.18880.1604
RBF  training0.20250.12430.10760.18400.10160.08550.07080.0883
RBF  test1.24120.21180.32010.34450.30140.11550.13790.1362
ANFIS  training0.00010.00010.04980.19530.00010.00010.04670.0903
ANFIS  test0.47660.00010.06330.46680.17380.00080.05290.1582
HONFIS (3rd ord.)  training0.00010.00010.01220.00320.00010.00100.01380.0050
HONFIS (3rd ord.)  test1.70320.00020.00680.00920.31680.00110.01360.0125
Table 10. Prediction results for 15 October.
Table 10. Prediction results for 15 October.
Prediction ModelNMSEMARE
1-Day3-Day7-Day30-Day1-Day3-Day7-Day30-Day
LSE  training0.18890.11980.09240.27200.09470.09830.09420.1047
LSE  test0.20710.19010.18990.19860.11080.11800.11940.1121
ARIMA  training1.54751.05690.58700.63010.36160.23110.17720.1965
ARIMA  test0.89570.67341.33920.37580.24390.20970.31190.1545
RBF  training0.01820.02310.00570.15830.02910.04260.01840.0677
RBF  test0.54440.03330.01260.10060.17720.04910.02730.0862
ANFIS  training0.00020.00010.00480.16920.00020.00030.01600.0702
ANFIS  test1.03720.00030.00860.15740.22170.00040.01880.1019
HONFIS (3rd ord.)  training0.00290.00360.00100.01450.00770.00780.00480.0112
HONFIS (3rd ord.)  test1.64750.00650.00040.00080.31240.01320.00300.0036
Table 11. Prediction results for 15 November.
Table 11. Prediction results for 15 November.
Prediction ModelNMSEMARE
1-Day3-Day7-Day30-Day1-Day3-Day7-Day30-Day
LSE  training0.82240.44290.75910.35230.11390.13120.04570.0884
LSE  test0.58960.43940.54630.43470.15750.15470.15420.1544
ARIMA  training1.05690.78700.66441.17590.23110.26980.19370.2855
ARIMA  test0.67340.94034.46490.95540.20970.27761.22470.2504
RBF  training0.07250.16620.48940.21600.03350.08680.03790.0638
RBF  test8.51470.18440.98323.15270.63480.09130.21680.3780
ANFIS  training0.00010.02470.04800.23630.00010.03230.06500.0687
ANFIS  test5.58500.01970.05992.68630.49800.02940.06880.3969
HONFIS (3rd ord.)  training0.00090.05660.16590.09110.00070.02190.00480.0279
HONFIS (3rd ord.)  test2.33140.04330.64800.37820.55520.03490.10450.1206
Table 12. Prediction results for 15 December.
Table 12. Prediction results for 15 December.
Prediction ModelNMSEMARE
1-Day3-Day7-Day30-Day1-Day3-Day7-Day30-Day
LSE  training0.63350.51070.55130.38140.11090.11100.09700.0984
LSE  test0.43260.39220.40510.45120.14360.15520.15260.1518
ARIMA  training1.04071.15361.93331.17660.30000.20670.21940.2016
ARIMA  test0.99980.88860.55630.42080.22720.20900.15570.1769
RBF  training0.07540.03790.05400.21060.04120.04810.03830.0665
RBF  test0.94420.07930.10610.12700.22070.07680.06270.0684
ANFIS  training0.06090.00430.05560.22720.04010.00640.04100.0719
ANFIS  test1.28900.00570.10050.14430.25990.00740.06390.0643
HONFIS (3rd ord.)  training0.00600.00390.01980.10120.00830.00690.00940.0316
HONFIS (3rd ord.)  test1.12320.00790.05440.10170.19510.00830.02980.0416
Table 13. Prediction results from 15 to 21 January.
Table 13. Prediction results from 15 to 21 January.
Prediction ModelNMSEMARE
7-Day30-Day7-Day30-Day
LSE  test0.63040.64640.07200.0739
ARIMA   test1.21301.22520.11920.0957
RBF  test5.91960.58630.15000.0734
ANFIS  test1.04520.58630.12390.0800
HONFIS (3rd ord.)  test3.48690.22960.16510.0185
Table 14. Prediction results from 15 to 21 February.
Table 14. Prediction results from 15 to 21 February.
Prediction ModelNMSEMARE
7-Day30-Day7-Day30-Day
LSE  test0.24900.23900.12420.1239
ARIMA   test1.43241.09980.30480.2697
RBF  test1.47710.12240.23590.0758
ANFIS  test0.66640.12810.18950.0755
HONFIS (3rd ord.)  test1.01240.01970.22360.0147
Table 15. Prediction results from 15 to 21 March.
Table 15. Prediction results from 15 to 21 March.
Prediction ModelNMSEMARE
7-Day30-Day7-Day30-Day
LSE  test0.14020.14140.09980.1007
ARIMA   test0.45670.29350.19340.1669
RBF  test0.28060.08010.12980.0657
ANFIS  test0.27320.09450.10910.0667
HONFIS (3rd ord.)  test3.49780.00080.35700.0039
Table 16. Prediction results from 15 to 21 April.
Table 16. Prediction results from 15 to 21 April.
Prediction ModelNMSEMARE
7-Day30-Day7-Day30-Day
LSE  test0.40620.44180.14930.1627
ARIMA   test0.77510.60390.26670.2392
RBF  test1.16340.17800.19380.1004
ANFIS  test0.23110.10440.31020.1223
HONFIS (3rd ord.)  test1.84550.00040.33430.0019
Table 17. Prediction results from 15 to 21 May.
Table 17. Prediction results from 15 to 21 May.
Prediction ModelNMSEMARE
7-Day30-Day7-Day30-Day
LSE  test0.40620.38820.14930.1299
ARIMA   test0.68850.43440.24920.1996
RBF  test4.43470.15710.12980.0773
ANFIS  test0.29620.10440.11090.0778
HONFIS (3rd ord.)  test1.82190.00030.35920.0033
Table 18. Prediction results from 15 to 21 June.
Table 18. Prediction results from 15 to 21 June.
Prediction ModelNMSEMARE
7-Day30-Day7-Day30-Day
LSE  test0.23340.22950.10310.0981
ARIMA   test0.35030.38250.11240.1502
RBF  test0.93450.12380.20620.0688
ANFIS  test0.59720.16550.16500.0818
HONFIS (3rd ord.)  test0.95600.012130.19350.0108
Table 19. Prediction results from 15 to 21 July.
Table 19. Prediction results from 15 to 21 July.
Prediction ModelNMSEMARE
7-Day30-Day7-Day30-Day
LSE  test0.19320.18860.07120.0725
ARIMA   test0.41670.40250.19180.1839
RBF  test9.81150.06950.49560.0400
ANFIS  test0.45920.08540.12290.045
HONFIS (3rd ord.)  test1.07010.00570.19050.0053
Table 20. Prediction results from 15 to 21 August.
Table 20. Prediction results from 15 to 21 August.
Prediction ModelNMSEMARE
7-Day30-Day7-Day30-Day
LSE  test0.14450.14080.07500.0815
ARIMA   test0.12830.14870.06200.0763
RBF  test6.23380.05211.29800.0473
ANFIS  test0.15230.08490.06910.0478
HONFIS (3rd ord.)  test0.98750.00450.18840.0065
Table 21. Prediction results from 15 to 21 September.
Table 21. Prediction results from 15 to 21 September.
Prediction modelNMSEMARE
7-Day30-Day7-Day30-Day
LSE  test0.18030.18060.10320.1023
ARIMA   test0.72500.37810.22160.1668
RBF  test0.47510.08980.16710.0670
ANFIS  test1.73610.10310.28230.0678
HONFIS (3rd ord.)  test1.81850.00170.31050.0048
Table 22. Prediction results from 15 to 21 October.
Table 22. Prediction results from 15 to 21 October.
Prediction ModelNMSEMARE
7-Day30-Day7-Day30-Day
LSE  test0.22660.21380.11690.1098
ARIMA   test0.83760.38080.20210.1626
RBF  test1.68180.08980.22370.0670
ANFIS  test1.02120.12250.17820.0650
HONFIS (3rd ord.)  test0.88920.01790.18840.0169
Table 23. Prediction results from 15 to 21 November.
Table 23. Prediction results from 15 to 21 November.
Prediction ModelNMSEMARE
7-Day30-Day7-Day30-Day
LSE  test0.41260.30830.12740.1166
ARIMA   test1.24141.36390.21930.2285
RBF  test2.65640.15090.22370.0686
ANFIS  test1.35670.14450.17660.1630
HONFIS (3rd ord.)  test1.64370.11590.23200.0347
Table 24. Prediction results from 15 to 21 December.
Table 24. Prediction results from 15 to 21 December.
Prediction ModelNMSEMARE
7-Day30-Day7-Day30-Day
LSE  test0.38000.37900.13470.1348
ARIMA   test1.32091.14320.23220.2059
RBF  test3.24320.21430.26490.0823
ANFIS  test0.47640.21680.12530.0886
HONFIS (3rd ord.)  test3.60650.06840.25190.0364

Share and Cite

MDPI and ACS Style

Rosato, A.; Altilio, R.; Araneo, R.; Panella, M. Prediction in Photovoltaic Power by Neural Networks. Energies 2017, 10, 1003. https://doi.org/10.3390/en10071003

AMA Style

Rosato A, Altilio R, Araneo R, Panella M. Prediction in Photovoltaic Power by Neural Networks. Energies. 2017; 10(7):1003. https://doi.org/10.3390/en10071003

Chicago/Turabian Style

Rosato, Antonello, Rosa Altilio, Rodolfo Araneo, and Massimo Panella. 2017. "Prediction in Photovoltaic Power by Neural Networks" Energies 10, no. 7: 1003. https://doi.org/10.3390/en10071003

APA Style

Rosato, A., Altilio, R., Araneo, R., & Panella, M. (2017). Prediction in Photovoltaic Power by Neural Networks. Energies, 10(7), 1003. https://doi.org/10.3390/en10071003

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop