Next Article in Journal
Development of an FEM for the Combined Electromagnetic and Hydraulic Forming Process Based on Experimental Data
Previous Article in Journal
The Saturation Calculation of NMR Logging Based on Constructing Water Spectrum Function
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhancing the Predictability of Wintertime Energy Demand in The Netherlands Using Ensemble Model Prophet-LSTM

by
Susan N. P. van de Sande
,
Ali M. M. Alsahag
* and
Seyed Sahand Mohammadi Ziabari
Faculty of Science, Informatics Institute, University of Amsterdam, 1012 WP Amsterdam, The Netherlands
*
Author to whom correspondence should be addressed.
Processes 2024, 12(11), 2519; https://doi.org/10.3390/pr12112519
Submission received: 20 September 2024 / Revised: 1 November 2024 / Accepted: 6 November 2024 / Published: 12 November 2024
(This article belongs to the Section Energy Systems)

Abstract

:
Energy demand forecasting is crucial for maintaining stable and affordable energy supplies, especially for vulnerable populations most affected by shortages and high costs. In the Netherlands, transmission system operator TenneT has raised concerns about potential electricity shortages by 2030. Rising energy prices and the impact of climate change on the energy demand further complicate today’s energy market. Policymakers lack clear insights into demand patterns, which complicates the optimization of energy use and the protection of at-risk communities. Accurate and timely forecasts are essential for addressing these issues and supporting sustainable energy management. This research focuses on enhancing the accuracy and lead time of wintertime energy demand forecasts in the Netherlands using advanced machine learning. The ensemble model Prophet-LSTM is trained on hourly load consumption data combined with climate change-related and energy price predictors. The results demonstrate significant improvements over baseline models, achieving a Pearson correlation coefficient of r = 0.93 compared to r = 0.50 in prior studies, as well as accurate forecasts up to 180 days ahead, compared to 2 months. Incorporating climate change-related predictors is challenging due to multicollinearity, highlighting the importance of careful predictor selection. Including energy price predictors yielded modest yet hopeful results, suggesting their ability to optimize energy demand forecasting.

1. Introduction

Wintertime energy demand forecasting (EDF) plays a crucial role in ensuring a reliable energy supply and managing available resources. It also helps mitigate the impact of extreme weather events caused by climate change on vulnerable populations, particularly in the Netherlands. In 2023, Dutch transmission system operator TenneT expressed concerns about potential electricity shortages in the Netherlands by 2030 [1]. From a global perspective, the current security of the electricity supply in the Netherlands is extremely high at 99.99963 % . However, the number of controllable coal and gas power stations in the Netherlands is set to decrease, while the country becomes increasingly dependent on sustainable electricity, which is highly dependent on weather conditions. This makes it uncertain whether sufficient electricity can be generated at any time to meet demands [2,3]. Therefore, accurate and timely EDF is crucial.
Important influences on the Dutch energy market are energy prices and climate change. Energy prices have been quickly increasing in the Netherlands, resulting in very high costs for Dutch households compared to other European Union (EU) countries [4]. The average gas price in the Netherlands is EUR 0.182 per kWh, significantly higher than the EU average of EUR 0.110 per kWh [5]. In 2022, high energy prices caused inflation to reach 5.2 % , decreasing purchasing power by an average of 2.7 % . The Dutch government has taken measures to help low- and middle-income households facing rising costs [6], but many continued to struggle financially [4]. In addition, gas consumption decreased by 15 % due to significantly higher gas prices in 2022, with people saving energy by turning down heating, taking shorter showers, or better insulating their homes [7]. Climate-wise, it is notable that climate change is causing shifts in the weather in the Netherlands, leading to more rain and heavier storms [8]. The wintertime energy demand fluctuates yearly and is closely tied to the variability of weather conditions [9]. Thus, changing weather patterns can lead to shifts in the energy demand, highlighting the importance of capturing these changes to enable people to meet their energy demands. Moreover, the share of renewable energy of the total energy consumption in the Netherlands was 15 % in 2022 [10]. The Climate Agreement aims for 70 % renewable electricity by 2030, and the EU aims for at least 32 % sustainable energy by 2030 [11]. However, climate change is causing less certainty in weather patterns in Europe, which has a negative long-term effect on the production of renewable energy [12]. For people to still meet their energy demands, less available renewable energy requires them to use more nonrenewable energy.
Ultimately, enhancing the accuracy and lead time of EDF in the Netherlands is crucial to protect the energy supply against evolving challenges such as increasing energy prices and the impact of climate change. In particular, TenneTs’ concerns highlight the importance of accurately forecasting the energy demand at any time to mitigate the effects of potential shortages and thereby protect vulnerable populations.
This study addresses the research gap in the unique dynamics of the Dutch energy market by introducing an advanced forecasting approach tailored to this specific context. While EDF has been extensively studied, examining the Dutch context provides novel insights. High energy prices negatively impact the energy demand [7], while the Netherlands is shifting toward renewable energy due to national climate goals and the phasing out of coal and gas power stations [1,11]. However, both nonrenewable and renewable energy consumption are reliant on weather conditions, which are impacted by climate change [8]. These factors highlight the complexity of the Dutch energy market and the need for an appropriate approach to EDF in the Netherlands.
The primary contribution of this study lies in extending current EDF research by applying advanced machine learning (AML) techniques specifically to Dutch energy data, building upon the research of [9]. In emphasizing the critical interaction between environmental, economic, and energy dynamics, this research aimed to improve forecasting accuracy and lead time—the length of time into the future for which predictions are made. By fitting this approach to Dutch energy data and addressing climate change impacts and energy price fluctuations, this study improves academic understanding and offers practical insights for energy planning strategies. This promises a more adaptive and resilient forecasting framework for the unique dynamics of the Netherlands. A novel approach is introduced that combines predictors closely related to the Dutch energy market to demonstrate its effectiveness in improving Dutch EDF performance.

2. Related Work

EDF is an extensively researched topic with research directions varying in both predictor and model selection. This section discusses the relevant predictors and models for EDF.

2.1. Baseline Approach

Challenges in EDF are two-fold: (1) rising energy prices are influencing the energy demand [7] and (2) climate change is shifting weather patterns, which influences energy demand patterns [8].
One of the most dominant approaches is proposed in [9], using a multiple linear regression (MLR) model to predict winter seasonal energy consumption two months in advance in the United Kingdom (UK), using climate change-related predictors. Leave-one-out cross-validation was used to evaluate the model using the Pearson correlation coefficient (PCC), yielding a significant correlation of r = 0.50 between the predicted and observed time series. The study explored the relationship between surface climate conditions and daily UK wintertime gas and electricity consumption, revealing that the prediction skill of energy consumption arises from the predictability of the North Atlantic oscillation (NAO) variability and the joint estimation of surface climate condition variables temperature, dew point depression, and wind speed. One main limitation of this work was the lack of validation of the model’s generalizability.
However, recent developments regarding rising energy prices and climate change, which affect the energy demand, require a more advanced approach to EDF.

2.2. Impact of Climate Change on Energy Demand

Currently, the EDF challenge is addressed by focusing mostly on climate change. Based on recent papers, it is notable that climate change-related features are frequently used as predictors, emphasizing the importance of these predictors in today’s energy market.
In 2022, the residential gas consumption in the Netherlands decreased by 22 % compared to 2021, depicting the lowest energy usage since 1990. Warmer weather conditions contributed to this reduction, demonstrating how changing weather conditions influence the energy demand. In the agricultural sector, a  31 % decrease in gas consumption was observed, mostly driven by high gas prices, but also influenced by warmer weather [13]. These trends emphasize the relevance of climate change-related predictors in EDF, such as temperature, which play an important role in understanding fluctuations in the energy demand.
Firstly, relevant climate change-related predictors are the sea ice concentration (SIC), stratospheric circulation (SC), sea surface temperature (SST), and the NAO variability index. The NAO is the dominant mode of winter variability over the North Atlantic and strongly influences surface weather conditions of the winter period such as average temperature, wind speed, and storminess, over much of the European continent and the United States [9,14,15]. Autumn conditions in the Arctic SIC, SC (especially the geopotential at 70 pHa), and SST influence the NAO variability [9,16]. SIC measures the amount of sea ice in a given area, typically expressed as a percentage [17]. SC is the air movement in the stratosphere [18]. SST indicates the surface temperatures in a broad swath of the tropical North Atlantic Ocean [19]. These are all important predictors as they influence NAO variability, which in turn affects energy consumption.
Secondly, relevant climate change-related predictors are weather-related such as surface air temperature, dew point depression, and wind speed, all showing significant correlations with gas consumption [9]. In the UK, winter means of temperature, the NAO, a UK-centered north–south pressure difference, and an additional predictor based on the frequency of high-demand weather types over the winter period are used to predict the gas demand and the number of extreme gas demand days over the winter period [15], demonstrating that seasonal weather forecasts can skillfully predict the weather-driven component of the winter gas demand and the number of extreme gas demand days in the UK. Similar research has also been performed outside of the UK, namely in Hanoi where temperature, absolute humidity, and wind speed are used to forecast the short-term electricity load [20].

2.3. Impact of Energy Price and Renewable Energy on Energy Demand

Furthermore, considering energy prices and renewable energy in EDF is also crucial. The relationship between the renewable and nonrenewable energy demand is complex. Intuitively, an increase in renewable energy demand results in a decrease in nonrenewable energy demand, given that only so much is needed. However, [21] suggests that it is unlikely that renewable and nonrenewable energy are perfect substitutes, suggesting some level of complementarity between these two forms of energy. The relationship between the nonrenewable and renewable energy demand was further investigated in [22], studying the elasticity of renewable and nonrenewable energy in countries that are part of the Organisation for Economic Co-operation and Development (OECD), which includes the Netherlands. The study showed that nonrenewable and renewable energy are substitutes in some industries and complements in others.
Additionally, the impact of nonrenewable energy prices on US renewable energy consumption was studied, showing a positive relationship [21]. This indicates that increases in nonrenewable energy prices cause increases in the renewable energy demand, resulting in decreases in the nonrenewable energy demand. This finding is supported by the authors of [23], who investigated the response of the electricity demand to price changes in Portugal, showing that increases in price lead to decreases in electricity consumption, with lower-income groups being most affected.
As briefly mentioned in Section 2.2, a  31 % decrease in gas consumption was observed in the Dutch agricultural sector, primarily driven by high gas prices [13]. This shows how significantly price can influence the energy demand, illustrating the importance of incorporating energy prices into EDF.

2.4. Energy Demand Forecasting Models

The consideration of modeling approaches is necessary to tackle the challenging problem of forecasting energy demand accurately.

2.4.1. More Traditional Models

Traditional statistical models are foundational in EDF. For instance, in [15], an LR model was utilized to predict Britain’s gas demand and the number of extreme gas demand days over the winter period, showing skillful predictions. Furthermore, an LR model and a long short-term memory (LSTM) model—a type of recurrent neural network that is effective for tasks such as time series forecasting due to its ability to learn and model long-term dependencies in sequential data [24]—were compared for predicting the daily and monthly El Niño–Southern Oscillation [25], showing that monthly performances from LSTM and LR are similar, whereas daily results show that the LSTM has some advantage over LR regarding correlation coefficient. This suggests that LSTM is superior at capturing nonlinear relationships compared to LR, although at a computational cost.

2.4.2. AML Models

State-of-the-art (SOTA) EDF research focuses on utilizing AML models, often comparing them with more traditional models. For example, a combined model based on LSTM and LightGBM was used to forecast power load in [26]. This model decomposes historical power load data with Empirical Mode Decomposition and combines this with historical weather data to establish LSTM and LightGBM prediction models, showing improved forecasting accuracy and application prospects for power load forecasting compared to traditional methods and standard LSTM and LightGBM methods. Moreover, the ensemble model NeuralProphet-LightGBM was used in [27] to forecast power load. By first decoupling the original sequence using NeuralProphet and then using LightGBM for further feature extraction, their model shows significantly better performance in comparison with comparative models.
Overall, the finding that LSTM models are better at capturing nonlinear relationships compared to LR models [25], in combination with the success of ensemble models in EDF [26,27], motivates the exploration of similar approaches for tackling the challenging task of EDF, combined with the use of the important predictors of climate change and energy price.

3. Methodology

This research concerns a multivariate time series forecasting task. This necessitates five key steps: data collection, exploratory data analysis (EDA), data cleaning and preparation, and model implementation and evaluation.

3.1. Data Collection

Multiple open source data sets were integrated to create a data set comprising predictors and a target variable, intended for model training and evaluation. The data set included energy consumption as the target variable, together with predictors related to climate change and energy prices.

3.1.1. Target Variable

Energy consumption in the Netherlands served as the target variable. For this purpose, load consumption data from the European Network of Transmission System Operators for Electricity (ENTSO-E) were utilized. ENTSO-E defines load as “an end-use device or customer that receives power from the electric system” [28]. The data set, ranging from 2009 to 2019, included hourly load information in megawatts from the Netherlands.

3.1.2. Predictors

The predictors in this study were subdivided into two categories: climate change-related and energy prices.
Climate change-related: Daily temperature, precipitation, and wind speed were used—previously utilized in studies such as [9,15,20]—obtained from the Royal Dutch Meteorological Institute [29]. The data, ranging from 1901 to 2024, originated from a weather station located in De Bilt, the Netherlands, which is representative of the average Dutch weather due to its central location [30].
Moreover, the NAO variability index was included, similar to [9,15]. Monthly data from 1950 to 2023 were sourced from the National Oceanic and Atmospheric Administration (NOAA) [31].
Furthermore, predictors that influence the NAO index—stratospheric circulation (geopotential height at 70 pHa), SIC and SST—were included [9,16]. The data were accessed from the Copernicus Climate Data Store and the NOAA on daily, every-three-day, and monthly levels [17,18,19], ranging from 2009 to 2019. The geopotential height and SIC data have high dimensions due to the numerous longitude and latitude values. Principal Component Analysis was applied to reduce the dimensionality to a single daily value, as previously discussed in [16].
In addition, the yearly ratio of renewable energy consumption relative to total energy consumption was incorporated as a predictor, sourced from the Central Bureau of Statistics (CBS) [32], ranging from 1990 to 2023.
Energy price: Finally, Dutch energy prices were included. Half-yearly average energy prices for consumers in the Netherlands were used, obtained from CBS and ranging from 2009 to 2023 [33].

3.2. EDA

EDA was conducted on all data sets individually and collectively, revealing daily, monthly, and yearly seasonality patterns in the load consumption data, as illustrated in Figure 1 and Figure 2. Daily patterns show higher load consumption on weekdays compared to weekends, with peaks typically at 7:00 and 18:00, and lower load consumption during nighttime and afternoon hours, as depicted in Figure 3. Figure 4 shows that load consumption peaks in the winter months (December, January, and February) and decreases in the summer.

3.2.1. Seasonal Decomposition

Seasonal decomposition was used to improve understanding of the underlying patterns in the load consumption data by decomposing it into its seasonal, trend-cycle, and remainder components [34,35], as displayed in Figure 5. This revealed an upward trend over the years and a seasonal pattern showing that consumption increases in the winter and decreases in the summer.

3.2.2. Autocorrelation Analysis

The seasonality in the data was further investigated using the autocorrelation function (ACF) to measure the linear relationship between lagged values of a time series [36]. The ACF was used to identify the appropriate lags of the load consumption target variable to use as predictors in the model.
Figure 6, Figure 7 and Figure 8 depict the ACF of hourly, daily, and monthly load consumption data, respectively. The plots reveal significant peaks at 24 h, 7 days, and 12 months. In Figure 6, spikes at both 24 h and 48 h indicate daily seasonality. Figure 7 shows spikes every 7 days, indicating weekly seasonality. Figure 8 displays a spike at 12 months, indicating yearly seasonality. The daily, weekly, and yearly seasonality in the data advocates for the use of lagged features of 24 h, 7 days, and 12 months, corresponding to these periods.

3.3. Data Cleaning and Preparation

Preprocessing steps are applied to each individual data set, after which they are combined into a unified data set that includes all predictors and the target variable.

3.3.1. Handling Outliers

Based on the EDA findings, the data exhibit outliers. This is addressed using the interquartile range ( I Q R ) method as a fitting outlier detection method [37,38]. The I Q R is robust against outliers because it relies on the middle 50 % of the data distribution. The IQR method considers a data point an outlier if it is more than 1.5 × I Q R below the first quartile q 1 or above the third quartile q 3 , where I Q R = q 3 q 1 . Outliers are only cleared from the train folds to prevent overfitting.

3.3.2. Shifting Data

The aim is to forecast the load consumption (at least) two months in advance. Hence, at the point of forecasting, only the reality of two months ago is known. The data are shifted for this purpose, such that day t has the predictors of day t 60 and the lags of load consumption of 24 h, 7 days, and 12 months of day t 60 . The 12-month lag results in the loss of the first year of the data set, 2009, since this requires a year of data available before a day t. Hence, with a data set that starts 1 January 2009, the first moment possible to take a lag of 60 days and an additional 12 months is 2 March 2010.

3.3.3. Handling Missing Values

Outlier removal resulted in some missing values in the data set. The missing values were imputed using Multivariate Imputation by Chained Equations (MICE), similar to in [39]. MICE involves a series of linear regressions run in sequence and repeated until the estimated values for missing values have sufficiently converged. To prevent data leakage, which is the case when data outside of the training data set is used to train the model, MICE is solely fitted to the training fold and is exclusively used to transform the test fold.

3.3.4. Predictor Selection

The model has 10 available predictors. The forecasting model was tested both including and excluding predictors. The impact of energy price and climate change-related predictors was tested by either including or excluding them from the model. Furthermore, the seasonal lags (as mentioned in Section 3.2.2) of the target variable load consumption were included in each of the predictor selections.

3.3.5. MinMaxScaling

MinMaxScaling was used to scale all of the features of the LSTM model, similar to in [40]. An advantage of this scaler is that it provides a uniform [ 0 , 1 ] scale across all features [41]. The scaling was carried out using the following formula:
x s c a l e d = x x m i n x m a x x m i n
where x is the original value of a feature, and x m i n and x m a x are the minimum value and the maximum value, respectively, of that feature in the data set. To prevent data leakage, the MinMaxScaler was solely fitted to the training data, and the fitted scaler was used to transform both the training and testing data sets.

3.4. Model Implementation: Baseline Model

For a baseline model, the traditional statistical forecasting model MLR was used. MLR has been shown to perform competitively with sophisticated methods [9,25] and offers several advantages, including its ease of implementation and interoperability. MLR allows for the use of multiple predictors, which makes the model suitable for capturing complex relationships between the predictors and target variables, aligning well with the overall goal of this research.

3.5. Model Implementation: AML Model

3.5.1. Ensemble Model

The ensemble model of Prophet and LSTM was used as the AML model, using Prophet for capturing seasonal patterns and holiday effects in the time series data and using LSTM to model the more complex and nonlinear relationships in the data, which might not be captured yet by Prophet. The use of an LSTM model for EDF was suggested by [9], supported by the performances of LSTM models in other studies [25,26,42]. The use of ensemble models is based on [26,27,42].

3.5.2. Prophet

Prophet is an additive time series forecasting tool that was developed by Facebook [43]. It is a nonlinear regression model and is widely recognized for its effectiveness in estimating seasonal fluctuations, performing well for time series with strong seasonality and several seasons of historical data [44]. Prophet automatically fits weekly and yearly seasonalities for time series more than two cycles long [44]. Considering that the data also show daily seasonality, automatic detection for daily seasonalities was enabled. Furthermore, predictors were added based on the predictor selection. Prophet was trained on the entire training data set, and this model was used for predictions on the entire training data set and test set. Prophet’s predictions y h a t and its trends for daily, weekly, and yearly seasonal components were combined with the selected predictors as input for the LSTM model [27].

3.5.3. LSTM

LSTM, a type of recurrent neural network, was first proposed in [24]. It is designed to capture and utilize long-term dependencies in sequential data, making it effective for tasks such as time series forecasting. Two LSTM setups were deployed: one including dropout layers (Model 1) and one excluding dropout layers (Model 2), visualized in Figure 9 and Figure 10.
Model 1 comprises an input layer, followed by two LSTM layers, and concludes with a dense layer. The first LSTM layer contains 30 nodes, and the second has 90 nodes. This architecture is similar to that introduced in [25]. However, the referenced study used 50 nodes in the first LSTM layer, and 150 nodes in the second. Due to the smaller data set in this research compared to in the study, the number of nodes was reduced while maintaining the same node ratio. Model 2 comprises two LSTM layers with dropout layers in between, and concludes with a dense layer. The first LSTM layer contains 30 nodes, and the second has 90 nodes. This architecture is similar to that introduced in [42], the only difference being that the study used 50 nodes in the first LSTM layer, and 100 nodes in the second. Again, node numbers were reduced while maintaining the node ratio. Both models were trained using 50 epochs, determined by the number of epochs needed for convergence, and a batch size of 128. Early stopping was implemented to prevent overfitting, allowing the model to stop training if the validation loss does not decrease by 0.0001 for 3 epochs, starting from epoch 10 [45].

3.6. Model Evaluation

3.6.1. Rolling Origin Cross-Validation

The model was trained and evaluated using rolling origin cross-validation (ROCV). In ROCV, the model is estimated on the training set, and forecasts are created for the test set. For every iteration, another year is added to the end of the training set, and the test set is advanced by the same period. The process stops when there are no more data left [46]. The minimum length of the training set was set at two years. To illustrate, the observations of 2010–2011 were used to train the model, which was then used to predict the load consumption in 2012. Next, the observations of 2010–2012 were used to train the model, which was used to predict for 2013, etcetera. This was repeated every year from 2010 to 2019 to ensure that the model was evaluated on 8 training–test folds.
ROCV was chosen over other methods, such as k-fold cross-validation (KCV), because of its suitability for time series data, where the order of observations matters. Where KCV would randomize the order of the observations, ROCV keeps their temporal order. This ensures that the model is only trained on past observations and predictions are only made on future observations.
To ensure the model’s generalizability, the test fold was subdivided into a wintertime test set, which includes the months of December, January, and February. Similar performance in the winter months and the general test set should give insights into the model’s ability to generalize and robustness.

3.6.2. PCC

The PCC measures the strength of the relationship between two variables, returning a value (r) between 1 and 1, where 1 indicates a strong positive relationship; 1 , a strong negative relationship; and 0, no relationship. It requires the following assumptions to be met: (1) the two variables must be measured at the interval or ratio scale; (2) there is a linear relationship between the two variables; (3) there should be no significant outliers; and (4) the data should be approximately normally distributed [47]. PCC was used to measure the strength of the relationship between the target variable and forecasted value and was appropriate for this work as the load consumption was measured at an interval scale, there was a linear relationship between the forecasted and actual value of the load consumption, outliers were removed, and the data were approximately normal distributed. Furthermore, the relevance of using the PCC is being able to compare this work to reference works.

3.6.3. RMSE

The RMSE is the distance of the prediction from the actual value and is calculated as R M S E = m e a n ( e t 2 ) , where e t = y t y ^ t . A disadvantage of the RMSE is its sensitivity to outliers due to its nature of being scale-dependent. Given that this metric takes the square of errors, the RMSE value has the potential to become enormous for extreme errors, as then the forecasted value and actual value are no longer on the same scale [48]. Especially considering that the average RMSE over the ROCV folds was calculated in this work, this has the potential to return extreme values if just a single fold shows bad performance, even though the rest might not.

3.6.4. MAPE

The MAPE is the accuracy of the prediction relative to the actual value and is calculated as M A P E = 1 n i = 1 n | Y i Y ^ i Y i | . It gives the percentage error and is commonly used in forecasting to compare performances. A limitation of the MAPE is that it can be infinite or undefined if Y i = 0 and that it can have extreme values if any Y i is close to zero [48]. In this research, the values of Y i were significantly larger than zero, so this disadvantage was neglectable.

4. Results

This section presents the results of training and evaluating the baseline MLR model and the Prophet-LSTM model with different predictor selections. The impacts of including energy price and climate change-related predictors and using a longer forecasting horizon are also investigated.

4.1. Performance of Baseline MLR

To investigate how the wintertime performance differs between the baseline model and the AML model (SRQ1), first, a baseline MLR model was trained, including both including and excluding predictors. The model was tested against multiple test folds consisting of wintertime months.
The results are displayed in Table 1. The results show similar performances for the MLR models, where the inclusion of the predictors ( R M S E a v g = 1874.75 , M A P E a v g = 11.11 , P C C a v g = 0.70 ) has a slightly improved RMSE, yet a slightly worse MAPE and PCC compared to the exclusion of predictors ( R M S E a v g = 1908.81 , M A P E a v g = 10.17 , P C C a v g = 0.69 ). Compared to the PCC baseline of r = 0.50 as stated in the RQ, the MLR shows promising improvement.

4.2. Wintertime Performance of Prophet-LSTM

To further continue the investigation of SRQ1, the Prophet-LSTM model was trained using ROCV and evaluated in the wintertime months.
Table 2 shows the performance of the Prophet-LSTM models during the wintertime period. In comparing the predictor selections for Model 1, it is notable that the exclusion of predictors ( R M S E a v g = 934.13 , M A P E a v g = 5.23 , P C C a v g = 0.93 ) outperforms the inclusion of predictors ( R M S E a v g = 1248.63 , M A P E a v g = 7.32 , P C C a v g = 0.92 ), especially showing improvements in the RMSE and MAPE. Excluding predictors gives more stable results across the different folds than including predictors.
In comparing the predictor selections for Model 2, it is notable that the exclusion of predictors ( R M S E a v g = 1060.95 , M A P E a v g = 6.07 , P C C a v g = 0.91 ) outperforms the inclusion of predictors ( R M S E a v g = 1366.50 , M A P E a v g = 7.63 , P C C a v g = 0.90 ), again especially showing large improvements in terms of the RMSE and MAPE. Again, performance across the different folds is more stable when excluding predictors.
When comparing the performances of Model 1 and Model 2, the results show that Model 1 consistently outperforms Model 2 across all average metrics. Moreover, Model 1 shows more stable results across the different folds when excluding predictors in comparison to Model 2. In contrast, both models show variability in performance metrics when including predictors. Overall, Model 1 outperforms Model 2 in terms of average metrics and demonstrates better stability across the different folds, especially when predictors are excluded.
The results show that for the models excluding predictors, Model 1 ( R M S E a v g = 934.13 , M A P E a v g = 5.23 , P C C a v g = 0.93 ) outperforms Model 2 ( R M S E a v g = 1060.95 , M A P E a v g = 6.07 , P C C a v g = 0.91 ) in terms of all average metrics. For the models including predictors, the performances of Model 1 ( R M S E a v g = 1441.66 , M A P E a v g = 8.56 , P C C a v g = 0.91 ) and Model 2 ( R M S E a v g = 1452.50 , M A P E a v g = 8.06 , P C C a v g = 0.90 ) are very similar. Model 1 has a slightly improved RMSE and PCC, yet a lower MAPE than Model 2.

MLR vs. Wintertime Prophet-LSTM

To answer SRQ1, a comparison is made between the obtained results from the MLR and Prophet-LSTM models on the wintertime test set. Observable from the results is that the Prophet-LSTM model significantly outperforms the MLR model with both model setups and with all predictor selections, showing improved RMSEs, MAPEs, and PCCs compared to the best-performing MLR model ( R M S E a v g = 1874.75 , M A P E a v g = 11.11 , P C C a v g = 0.70 ). The Prophet-LSTM model performs best when excluding predictors, whereas the MLR model performs slightly better when including them. Again, compared to the PCC baseline of r = 0.50 , as stated in the research question, both the MLR and especially the Prophet-LSTM model show significant improvement in PCC.

4.3. General Performance of Prophet-LSTM

To further assess the predictive power and generalization ability of the Prophet-LSTM models, their performances were examined on the entire test set.
The results in Table 3 show the performances of the Prophet-LSTM models for both predictor selections on the entire test set. Firstly, Model 1 performs better when excluding predictors ( R M S E a v g = 910.22 , M A P E a v g = 5.55 , P C C a v g = 0.92 ) rather than including them ( R M S E a v g = 1417.15 , M A P E a v g = 9.27 , P C C a v g = 0.89 ). A similar pattern is seen for Model 2, showing improved performance when excluding predictors ( R M S E a v g = 1297.00 , M A P E a v g = 8.50 , P C C a v g = 0.86 ) contrary to including them ( R M S E a v g = 1379.03 , M A P E a v g = 8.58 , P C C a v g = 0.86 ). Moreover, the results show that Model 1 outperforms Model 2 when excluding predictors. However, the evaluation of the entire test set shows that Model 2 performs slightly better than Model 1 when including predictors.

Best Model Selection

For computational reasons, the decision was made to continue the rest of the evaluation with a single model, which was the one considered the best Prophet-LSTM model. This model was chosen based on its performance and its ability to generalize.
Across all the Prophet-LSTM models, Model 1 excluding predictors outperformed all the other models on both the general ( R M S E a v g = 910.22 , M A P E a v g = 5.55 , P C C a v g = 0.92 ) and the wintertime test sets ( R M S E a v g = 934.13 , M A P E a v g = 5.23 , P C C a v g = 0.93 ). Additionally, it had the best ability to generalize, showing very stable performances across the different folds as well as very similar performances on the wintertime and general test sets.

4.4. Impact of Predictors

4.4.1. Climate Change-Related Predictors

The impact of climate change-related predictors (SRQ2) was investigated by integrating them into the model, after which the performance was compared to the model excluding them.
Figure 11 shows the performance of including the climate change-related predictors, on both the entire test set and wintertime subset. The detailed performances are noted in Table 4.
The results show that including climate change-related predictors negatively impacts the forecasting ability of the model, showing lower performance across all evaluation metrics on both the entire ( R M S E = 1182.93 , M A P E = 6.98 , P C C = 0.90 ) and wintertime test data sets ( R M S E = 1304.87 , M A P E = 7.32 , P C C = 0.92 ).
Subsequently, further research was conducted on the impact of the inclusion of individual climate change-related predictors on the performance of the model. Due to computational considerations, the data set was split into seven years of training data and two years of testing data (roughly a training–test split of 75–25%), rather than applying cross-validation. Of these two years of testing data, about 25% were considered wintertime.
Table 5 shows the impact of evaluating the Prophet-LSTM model trained with the individual climate change-related predictors. In comparing the performance metrics obtained during evaluation from the model including all climate change-related features ( R M S E g e n e r a l = 1182.93 , M A P E g e n e r a l = 6.98 , P C C g e n e r a l = 0.90 , R M S E w i n t e r = 1304.87 , M A P E w i n t e r = 7.32 , P C C w i n t e r = 0.92 ), it is notable that the performances of the models with the individual predictors are now much closer to the model excluding any climate change-related predictors ( R M S E g e n e r a l = 934.13 , M A P E g e n e r a l = 5.23 , P C C g e n e r a l = 0.93 , R M S E w i n t e r = 910.22 , M A P E w i n t e r = 5.55 , P C C w i n t e r = 0.92 ).
To understand why using individual predictors improves performance, multicollinearity amongst the climate change-related variables was considered. Multicollinearity occurs when strongly correlated predictors are collectively used in a model, making it difficult for the model to determine the individual effects of the predictors when incorporating them collectively. This leads to overfitting of the model due to possibly inaccurate estimates of the predictor’s coefficients [49]. The Variance Inflation Factor (VIF) was used to measure the multicollinearity between the climate change-related predictors. The results indicate that the predictors dailyPrecipitation, NAO, and SIC exhibit nearly no multicollinearity with VIF scores of V I F = ± 1 , whereas the predictors dailyMeanWindspeed and dailyMeanTemperature display moderate to high multicollinearity with VIF scores of V I F = ± 10 . The predictors SST, geopotential, and renewableEnergyRatio show extremely high multicollinearity with VIF scores of V I F > 42 [50]. The VIF scores are depicted in Figure 12.

4.4.2. Energy Price Predictors

To investigate the impact of using energy price predictors (SRQ3), the best model was tested by including the energy price predictors.
Figure 13 depicts the performance of including the energy price predictors in the Prophet-LSTM model, on both the entire test set and the wintertime subset. The detailed performances are listed in Table 6. The results show that the addition of energy price predictors is very similar to that of excluding them. A modest improvement in the RMSE is seen for the entire test set including price predictors ( R M S E a v g = 895.74 , M A P E a v g = 5.40 , P C C a v g = 0.92 ) compared to that excluding them ( R M S E a v g = 934.13 , M A P E a v g = 5.23 , P C C a v g = 0.93 ). Alternatively, the wintertime forecasts including price predictors ( R M S E a v g = 932.58 , M A P E a v g = 5.21 , P C C a v g = 0.93 ) show slightly worse results compared to excluding the price predictors ( R M S E a v g = 910.22 , M A P E a v g = 5.55 , P C C a v g = 0.92 ), but the differences are minor.

4.5. Longer Horizon Forecasts

The feasibility of forecasting the energy demand for more than two months in advance (RQ) was investigated using an iterative process, iterating through lags of 60 to 180 days (with 15-day increments) to evaluate its impact on all performance metrics. Due to computational considerations, the same trainingtest split of 75–25% was used as in Section 4.4.1 to conduct further research.
Figure 14 illustrates the trends in the metrics across different lags. The detailed performance values of each lag can be found in Table 7. The results show that overall, the performance generally decreases with increasing forecast horizons, exhibiting a slight increase again for lag 180. The wintertime forecasts perform competitively with and even slightly outperform the general forecasts. The best-performing lag is observed at the 90-day lag ( R M S E g e n e r a l = 898.11 , M A P E g e n e r a l = 5.18 , P C C g e n e r a l = 0.90 , R M S E g e n e r a l = 696.31 , M A P E g e n e r a l = 3.73 , P C C g e n e r a l = 0.95 ), showing improved performance metrics compared to all other lags.

5. Discussion

This section reflects on the obtained results, compares them to SOTA approaches, and provides limitations for this study.

5.1. Baseline MLR vs. Prophet-LSTM

The results demonstrate that Prophet-LSTM consistently outperforms the MLR model across all predictor selections and model setups. Interestingly, the Prophet-LSTM model performs best when predictors are excluded, whereas the MLR model benefits from their inclusion. This difference arises from the models’ natures: Prophet-LSTM combines Prophet’s ability to capture seasonality and trends with LSTM’s ability to model complex patterns. The additional predictors lead Prophet-LSTM to overfit, capturing noise rather than underlying patterns. In contrast, MLR benefits from the predictors since it is not able to fit such complex patterns and is thus less likely to overfit. Instead, including the predictors improves MLR’s forecast accuracy. This finding aligns with [25], which emphasized LSTMs’ strength in capturing nonlinearities compared to LR models.
Another important finding is that Prophet-LSTM Model 1 outperforms Model 2’s forecasts for both predictor selections in the wintertime forecasts. Contrarily, for general forecasts, Model 2 shows better performance compared to Model 1 when predictors are included. This difference can be explained by the use of dropout layers in Model 2. These likely mitigate overfitting caused by the large dimensionality of the data due to the numerous predictors. Likely, the overfitting is not an issue when forecasting on the wintertime test set given its lower data variability.
Unlike the model in [9], the Prophet-LSTM model in this work was extensively validated for its ability to generalize, showing stable performance across the different folds, especially when predictors are excluded. The general and wintertime forecasts of Prophet-LSTM are quite similar, indicating that the models can generalize well given their similar performances on different test sets.
Comparing the wintertime performances of the MLR and Prophet-LSTM to SOTA approaches using a similar MLR setup in [9] shows that both models outperform the paper’s approach consistently, comparing PCCs of 0.69 and 0.70 for the MLR model and PCCs ranging from 0.90 to 0.93 for Prophet-LSTM, compared to the obtained PCC of 0.50 in the reference paper.

5.2. Impact of Adding Climate Change-Related and Energy Price Predictors

An important finding is that incorporating energy price predictors in the Prophet-LSTM model results in a slight improvement in performance. This finding aligns with past research [51], and considering that energy prices have been increasing over the past years, the expectation is that incorporating energy prices when modeling with more recent data will show even more impact from adding this predictor.
Furthermore, it was observed that climate change-related predictors negatively impact the performance of the Prophet-LSTM model, as shown in Table 4 in Section 4.4.1. Upon further investigation, it was found that including the predictors collectively results in poorer performance compared to including the predictors individually. Further analysis using the VIF revealed multicollinearity amongst some of the climate change-related predictors. This makes it difficult for the model to determine the individual effects of the predictors when incorporating them collectively. In turn, this leads to overfitting of the model due to possibly inaccurate estimates of the predictor’s coefficients [49]. Overall, this illustrates the importance of careful predictor selection.

5.3. Lead Time Improvement

Noteworthy is the performance of the 60-day lag in the lead time experiment conducted on a 75–25% training–test split, which compares well with previously obtained results through ROCV. This indicates that the model generalizes well, as forecasting on a larger test set with more years and data variance yields similar results to forecasting on a smaller test set. Moreover, the results reveal that longer forecasting horizons exhibit lower performance, except for the 90-day lag, which demonstrates increased performance compared to the 60-day lag as consistently used in this paper. Although performance tends to decrease for longer lags, a slight performance improvement is observed at the 180-day lag. The performance at the 90-day lag is noteworthy given the insignificant autocorrelation at the 3-month lag (roughly 90 days) in the monthly ACF in Figure 8 and might be caused by uncaptured seasonalities in the data. In contrast, the improvement at the 180-day lag corresponds with the significant autocorrelation observed at the 6-month lag in the monthly ACF in Figure 8. Overall, the findings show improved performance on the 60-day lag compared to similar works [9].

5.4. Limitations

Some limitations of this work are important to consider. Firstly, one limitation arises from the different data granularities. While the target variable was at an hourly level, not all data were available at the same granularity. During ROCV, some training folds contained predictors with minimal variance, such as half-yearly values, which might not fully represent reality. However, performance did not vary much when comparing ROCV results and 75–25% test split results, in which case the training set resembles reality more closely. This weakens the impact of this limitation to some extent, though it should still be considered. Furthermore, when comparing the findings to baseline work [9], which uses hourly energy consumption data as opposed to the hourly data used in this paper, it may not be entirely fair to directly compare correlation factors. While the differences in granularity could account for some variability, the magnitude of the observed differences suggests that other factors very likely contribute as well. Secondly, the primary focus on wintertime forecasts limited the amount of available data for training and testing. Since a year only contains three winter months, it is hard to form a test set large enough while also retaining sufficient training data. The limited amount of data remains a concern but was mitigated using ROCV to validate the model on different winters across different years, as well as validation on a general test set to assess how well the model generalizes. Thirdly, multicollinearity among predictors is a potential limitation. Model performance may have varied if the predictors had not displayed this behaviour, potentially showing enhanced performance. Finally, while the performance of the Prophet-LSTM model has been extensively researched within the context of the Dutch energy market, its generalizability may be limited. Despite testing the model’s performance on both wintertime and a general test set, its applicability to other countries remains untested. This limitation arises from the model’s specific tailoring to the unique characteristics of the Dutch energy market, in addition to differing energy demand factors and challenges faced in other countries.

6. Conclusions

In conclusion, this study sheds light on accurate long-horizon EDF in the Netherlands. EDF is crucial for maintaining stable and affordable energy supplies, especially for vulnerable populations most affected by shortages and high costs, as emphasized by multiple studies [9,15]. With potential energy shortages such as those Tennet is concerned with, rising energy prices, and the impact of climate change on the energy demand complicating today’s energy market, the importance of forecasts being very accurate and readily available upfront to mitigate the effects of electricity shortages, rising energy prices, and climate change is highlighted. The results demonstrate that the accuracy and lead time of energy demand forecasts in the Netherlands can be enhanced to a PCC of r = 0.93 and lead time of 180 days, compared to the PCC of r = 0.50 and lead time of 2 months observed in reference work [9]. Additionally, the Prophet-LSTM model shows significant improvement compared to a baseline MLR model in terms of RMSE, MAPE, and PCC. It consistently outperforms the MLR model, demonstrating its ability to capture complex patterns in the data. Furthermore, collectively incorporating climate change-related features illustrated poorer performance as opposed to incorporating them individually. This finding highlights the importance of careful prediction selection, trying to avoid including predictors that exhibit multicollinearity. Conversely, including energy price predictors yielded slightly improved energy demand forecasts. Aligning with existing research [23], this finding gives valuable insight for enhancing forecasting models in the future.
Despite the aforementioned limitations, this paper provides valuable insights into enhancing the accuracy and lead time of energy demand forecasts, especially in the context of the Netherlands. The societal impact of these findings is important to note. Accurately forecasting energy demand during the winter months is crucial for ensuring reliable energy supply, managing available resources, and mitigating the impact of extreme weather events caused by climate change on vulnerable populations. This is especially important with the prospect of potential future energy shortages and in times when climate change and energy prices play such an important role.

Future Work

For future work, a suggestion is to explore increasing the forecasting horizon, taking bigger lags such as 180 days, where it is especially important that lags are in line with the seasonal patterns observed in the data. Moreover, it would be interesting to see what the impact of higher granularity predictors would be on the forecasting ability of the models. Moreover, combining predictors that show multicollinearity into a single variable could enhance the forecasting ability of the model, as this allows the model to accurately estimate the predictors’ coefficients. Finally, the model’s generalizability could be improved by adapting it to fit a wider range of countries. This would allow for a better evaluation of the model’s ability to generalize across different energy environments.

Author Contributions

Conceptualization, S.N.P.v.d.S. and A.M.M.A.; methodology, S.N.P.v.d.S. and A.M.M.A.; software, S.N.P.v.d.S.; validation, A.M.M.A. and S.S.M.Z.; formal analysis, S.N.P.v.d.S.; investigation, S.N.P.v.d.S.; resources, S.N.P.v.d.S. and A.M.M.A.; data curation, S.N.P.v.d.S.; writing—original draft preparation, S.N.P.v.d.S. and A.M.M.A.; writing—review and editing, S.N.P.v.d.S. and A.M.M.A.; visualization, S.N.P.v.d.S.; supervision, A.M.M.A. and S.S.M.Z.; project administration, A.M.M.A.; funding acquisition, A.M.M.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding. The APC was funded by the University of Amsterdam.

Data Availability Statement

The data presented in this study are openly available in https://github.com/snpsande/edf-prophet-lstm (accessed on 20 February 2024).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. TenneT TSO B.V. Rapport Monitor Leveringszekerheid 2024; TenneT: Arnhem, The Netherlands, 2024. [Google Scholar]
  2. Koster, R. Netbeheerder Waarschuwt Voor Stroomtekort in 2030; NOS: Hilversum, The Netherlands, 2023. [Google Scholar]
  3. Telegraaf. TenneT: Stroomtekort Dreigt Na 2030 Door Afhankelijkheid Groene Energie; Telegraaf: Arnhem, The Netherlands, 2023. [Google Scholar]
  4. NOS. Nederland Heeft Hoogste Gasprijs van EU, FNV wil Prijsplafond en Hogere Lonen; NOS: Hilversum, The Netherlands, 2022. [Google Scholar]
  5. Koenraadt, B. Energieprijzen Europa Verschillen Sterk: Nederlandse Gasprijs Bij Hoogste Drie; Energievergelijk: Barendrecht, The Netherlands, 2024. [Google Scholar]
  6. Ministerie van Algemene Zaken. Maatregelenpakket om Gevolgen Stijgende Energieprijzen en Aanhoudende Inflatie te Verzachten; Rijksoverheid: The Hague, The Netherlands, 2022.
  7. RTL, Z. Laagste Gasverbruik in 50 Jaar: ‘Bezuinigen Vaak Noodzakelijk’; RTL Z: Hilversum, The Netherlands, 2023. [Google Scholar]
  8. Ministerie van Infrastructuur en Waterstaat. Klimaatverandering en Gevolgen; Ministerie van Infrastructuur en Waterstaat: The Hague, The Netherlands, 2023.
  9. Li, S.; Sriver, R.; Miller, D. Skillful prediction of UK seasonal energy consumption based on surface climate information. Environ. Res. Lett. 2023, 18, 064007. [Google Scholar] [CrossRef]
  10. Centraal Bureau voor de Statistiek. Renewable Energy Share Rose to 15 Percent in 2022; Centraal Bureau voor de Statistiek: The Hague, The Netherlands, 2023.
  11. Ministerie van Algemene Zaken. Rijksoverheid Stimuleert Duurzame Energie; Rijksoverheid: The Hague, The Netherlands, 2023.
  12. European Commission. Gevolgen van Klimaatverandering. Rijksoverheid. Available online: https://www.rijksoverheid.nl/onderwerpen/klimaatverandering/gevolgen-klimaatverandering (accessed on 29 June 2024).
  13. voor de Statistiek, C.B. Laagste Energieverbruik in Nederland Sinds 1990; Centraal Bureau voor de Statistiek: The Hague, The Netherlands, 2023. [Google Scholar]
  14. Hurrell, J.W. Decadal Trends in the North Atlantic Oscillation: Regional Temperatures and Precipitation. Science 1995, 269, 676–679. [Google Scholar] [CrossRef]
  15. Thornton, H.; Scaife, A.; Hoskins, B.; Brayshaw, D.; Smith, D.; Dunstone, N.; Stringer, N.; Bett, P. Skilful seasonal prediction of winter gas demand. Environ. Res. Lett. 2019, 14, 024009. [Google Scholar] [CrossRef]
  16. Wang, L.; Ting, M.; Kushner, P. A robust empirical seasonal prediction of winter NAO and surface climate. Sci. Rep. 2017, 7, 279. [Google Scholar] [CrossRef]
  17. Copernicus Climate Change Service (C3S). Sea Ice Concentration Daily Gridded Data from 1979 to Present Derived from Satellite Observations. Available online: https://cds.climate.copernicus.eu/datasets/satellite-sea-ice-concentration?tab=overview (accessed on 29 June 2024).
  18. Hersbach, H.; Bell, B.; Berrisford, P.; Biavati, G.; Horányi, A.; Muñoz Sabater, J.; Nicolas, J.; Peubey, C.; Radu, R.; Rozum, I.; et al. ERA5 Hourly Data on Pressure Levels from 1940 to Present. Copernicus Climate Change Service (C3S). 2023. Available online: https://cds.climate.copernicus.eu/datasets/reanalysis-era5-pressure-levels?tab=overview (accessed on 29 June 2024).
  19. Reynolds, R.; Rayner, N.; Smith, T.; Stokes, D.; Wang, W. An improved in situ and satellite SST analysis for climate. J. Clim. 2002, 15, 1609–1625. [Google Scholar] [CrossRef]
  20. Huy, P.; Minh, N.; Tien, N.; Anh, T. Short-Term Electricity Load Forecasting Based on Temporal Fusion Transformer Model. IEEE Access 2022, 10, 106296–106304. [Google Scholar] [CrossRef]
  21. Atems, B.; Mette, J.; Lin, G.; Madraki, G. Estimating and forecasting the impact of nonrenewable energy prices on US renewable energy consumption. Energy Policy 2023, 173, 113374. [Google Scholar] [CrossRef]
  22. Kumar, S.; Fujii, H. Substitute or complement? Assessing renewable and nonrenewable energy in OECD countries. Appl. Econ. 2015, 47, 1438–1459. [Google Scholar] [CrossRef]
  23. Silva, S.; Soares, I.; Pinho, C. Electricity demand response to price changes: The Portuguese case taking into account income differences. Energy Econ. 2017, 65, 335–342. [Google Scholar] [CrossRef]
  24. Hochreiter, S.; Schmidhuber, J. Long Short-Term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  25. Huang, A.; Vega-Westhoff, B.; Sriver, R. Analyzing El Niño—Southern Oscillation Predictability Using Long-Short-Term-Memory Models. Earth Space Sci. 2019, 6, 212–221. [Google Scholar] [CrossRef]
  26. Zhou, Y.; Lin, Q.; Xiao, D. Application of LSTM-LightGBM Nonlinear Combined Model to Power Load Forecasting. J. Phys. Conf. Ser. 2022, 2294, 012035. [Google Scholar] [CrossRef]
  27. Zhao, Y.; Guo, N.; Chen, W.; Zhang, H.; Guo, B.; Shen, J.; Tian, Z. Multi-step ahead forecasting for electric power load using an ensemble model. Expert Syst. Appl. 2022, 211, 118649. [Google Scholar] [CrossRef]
  28. European Network of Transmission System Operators for Electricity (ENTSO-E). Glossary; European Network of Transmission System Operators for Electricity (ENTSO-E): Brussels, Belgium, 2004. [Google Scholar]
  29. Koninklijk Nederlands Meteorologisch Instituut. KNMI—Daggegevens van het Weer in The Nederlands; Koninklijk Nederlands Meteorologisch Instituut: De Bilt, The Netherlands, 2024.
  30. Koninklijk Nederlands Meteorologisch Instituut. KNMI—Automatische Weerstations; Koninklijk Nederlands Meteorologisch Instituut: De Bilt, The Netherlands, 2024.
  31. North Atlantic Oscillation (NAO). Available online: https://www.ncei.noaa.gov/access/monitoring/nao/ (accessed on 29 June 2024).
  32. CBS Statline. Hernieuwbare Energie; Verbruik Naar Energiebron, Techniek en Toepassing; Centraal Bureau voor de Statistiek: The Hague, The Netherlands, 2024. [Google Scholar]
  33. CBS Statline. Eindverbruikersprijzen Aardgas en Elektriciteit; Centraal Bureau voor de Statistiek: The Hague, The Netherlands, 2024. [Google Scholar]
  34. Hyndman, R.J.; Athanasopoulos, G. 6.1 Time series components. In Forecasting: Principles and Practice, 2nd ed.; OTexts: Melbourne, Australia, 2018. [Google Scholar]
  35. Statsmodels. Seasonal Decompose, 0.15.0. 2024. Available online: https://www.statsmodels.org/dev/generated/statsmodels.tsa.seasonal.seasonal_decompose.html (accessed on 29 June 2024).
  36. Hyndman, R.J.; Athanasopoulos, G. 2.8 Autocorrelation. In Forecasting: Principles and Practice, 2nd ed.; OTexts: Melbourne, Australia, 2018. [Google Scholar]
  37. Shaikh, A.K.; Nazir, A.; Khan, I.; Shah, A.S. Short term energy consumption forecasting using neural basis expansion analysis for interpretable time series. Sci. Rep. 2022, 12, 22562. [Google Scholar] [CrossRef]
  38. Al-Saudi, K.; Degeler, V.; Medema, M. Energy Consumption Patterns and Load Forecasting with Profiled CNN-LSTM Networks. Processes 2021, 9, 1870. [Google Scholar] [CrossRef]
  39. Ruggles, T.H.; Farnham, D.J.; Tong, D.; Caldeira, K. Developing reliable hourly electricity demand data through screening and imputation. Sci. Data 2020, 7, 155. [Google Scholar] [CrossRef]
  40. Aidoni, A.; Kofidis, K.; Cocianu, C.L.; Avram, L. Deep learning models for natural gas demand forecasting: A comparative study of MLP, CNN and LSTM. Rom. J. Pet. Gas Technol. 2023, 4, 133–148. [Google Scholar] [CrossRef]
  41. Genov, E.; Petridis, S.; Iliadis, P.; Nikopoulos, N.; Coosemans, T.; Messagie, M.; Camargo, L.R. Short-Term Load Forecasting in a microgrid environment: Investigating the series-specific and cross-learning forecasting methods. J. Phys. Conf. Ser. 2021, 2042, 012035. [Google Scholar] [CrossRef]
  42. Tan, M.; Yuan, S.; Li, S.; Su, Y.; Li, H.; He, F. Ultra-Short-Term Industrial Power Demand Forecasting Using LSTM Based Hybrid Ensemble Learning. IEEE Trans. Power Syst. 2019, 35, 2937–2948. [Google Scholar] [CrossRef]
  43. Taylor, S.; Letham, B. Forecasting at Scale; PeerJ, Inc.: London, UK, 2017. [Google Scholar] [CrossRef]
  44. Hyndman, R.J.; Athanasopoulos, G. 12.2 Prophet model. In Forecasting: Principles and Practice, 2nd ed.; OTexts: Melbourne, Australia, 2018. [Google Scholar]
  45. Nazir, S.; Aziz, A.A.; Hosen, J.; Aziz, N.A.; Murthy, G.R. Forecast Energy Consumption Time-Series Dataset using Multistep LSTM Models. J. Phys. Conf. Ser. 2021, 1933, 012054. [Google Scholar] [CrossRef]
  46. Svetunkov, I. 2.4 Rolling Origin. In Forecasting and Analytics with the Augmented Dynamic Adaptive Model (ADAM); CRC Press: Boca Raton, FL, USA, 2024. [Google Scholar]
  47. Obilor, E.I.; Amadi, E.C. Test for Significance of Pearson’s Correlation Coefficient. Int. J. Innov. Math. Stat. Energy Policies 2018, 6, 11–23. [Google Scholar]
  48. Hyndman, R.J.; Athanasopoulos, G. 3.4 Evaluating forecast accuracy. In Forecasting: Principles and Practice, 2nd ed.; OTexts: Melbourne, Australia, 2018. [Google Scholar]
  49. Yoo, W.; Mayberry, R.; Bae, S.; Singh, K.; He, Q.; Lillard, J. A Study of Effects of MultiCollinearity in the Multivariable Analysis. Int. J. Appl. Sci. Technol. 2014, 4, 9–19. [Google Scholar] [PubMed]
  50. Shrestha, N. Detecting multicollinearity in regression analysis. Am. J. Appl. Math. Stat. 2020, 8, 39–42. [Google Scholar] [CrossRef]
  51. Suganthi, L.; Samuel, A.A. Energy models for demand forecasting—A review. Renew. Sustain. Energy Rev. 2012, 16, 1223–1240. [Google Scholar] [CrossRef]
Figure 1. Hourly and monthly average load consumption from 2009 to 2019.
Figure 1. Hourly and monthly average load consumption from 2009 to 2019.
Processes 12 02519 g001
Figure 2. One month of daily load consumption (January 2010).
Figure 2. One month of daily load consumption (January 2010).
Processes 12 02519 g002
Figure 3. Average hourly load consumption.
Figure 3. Average hourly load consumption.
Processes 12 02519 g003
Figure 4. Monthly load consumption of a single year (2015).
Figure 4. Monthly load consumption of a single year (2015).
Processes 12 02519 g004
Figure 5. Seasonal decomposition of the load consumption time series.
Figure 5. Seasonal decomposition of the load consumption time series.
Processes 12 02519 g005
Figure 6. Hourly load consumption ACF.
Figure 6. Hourly load consumption ACF.
Processes 12 02519 g006
Figure 7. Daily load consumption ACF.
Figure 7. Daily load consumption ACF.
Processes 12 02519 g007
Figure 8. Monthly load consumption ACF.
Figure 8. Monthly load consumption ACF.
Processes 12 02519 g008
Figure 9. Visualization of first LSTM setup (Model 1), consisting of an input layer, followed by two LSTM layers, and concluding with a dense layer.
Figure 9. Visualization of first LSTM setup (Model 1), consisting of an input layer, followed by two LSTM layers, and concluding with a dense layer.
Processes 12 02519 g009
Figure 10. Visualization of second LSTM setup (Model 2), consisting of an input layer, followed by two LSTM layers that are each followed by a dropout layer, and concluding with a dense layer.
Figure 10. Visualization of second LSTM setup (Model 2), consisting of an input layer, followed by two LSTM layers that are each followed by a dropout layer, and concluding with a dense layer.
Processes 12 02519 g010
Figure 11. Performance of Prophet-LSTM model trained with climate-change related predictors in RMSE, MAPE, and PCC, in comparison with average performance of best Prophet-LSTM model.
Figure 11. Performance of Prophet-LSTM model trained with climate-change related predictors in RMSE, MAPE, and PCC, in comparison with average performance of best Prophet-LSTM model.
Processes 12 02519 g011
Figure 12. VIF scores of the individual climate-change related predictors.
Figure 12. VIF scores of the individual climate-change related predictors.
Processes 12 02519 g012
Figure 13. Performance of Prophet-LSTM model trained with energy price predictors in RMSE, MAPE, and PCC, in comparison with average performance of best Prophet-LSTM model.
Figure 13. Performance of Prophet-LSTM model trained with energy price predictors in RMSE, MAPE, and PCC, in comparison with average performance of best Prophet-LSTM model.
Processes 12 02519 g013
Figure 14. Performance of Prophet-LSTM trained with varying lags, evaluated on general and wintertime test set using RMSE, MAPE, and PCC per lag.
Figure 14. Performance of Prophet-LSTM trained with varying lags, evaluated on general and wintertime test set using RMSE, MAPE, and PCC per lag.
Processes 12 02519 g014
Table 1. Performance of baseline MLR model on wintertime months across different folds in RMSE, MAPE, and PCC.
Table 1. Performance of baseline MLR model on wintertime months across different folds in RMSE, MAPE, and PCC.
Pred.FoldRMSEMAPEPCC
excl.11910.5311.450.71
22207.3911.850.72
31856.9010.550.72
41887.119.470.71
51743.459.290.70
61826.489.510.69
71959.859.970.63
81878.809.290.62
Avg.1908.8110.170.69
incl.12199.5412.800.71
21958.1312.620.66
31896.9612.640.73
41967.4410.420.73
51753.8010.040.70
61693.4410.260.70
71757.519.430.65
81777.619.110.65
Avg.1874.7511.110.70
Table 2. Performance of two Prophet-LSTM model setups with different predictor selections on wintertime months across different folds in RMSE, MAPE, and PCC.
Table 2. Performance of two Prophet-LSTM model setups with different predictor selections on wintertime months across different folds in RMSE, MAPE, and PCC.
Pred.FoldModel 1Model 2
RMSEMAPEPCCRMSEMAPEPCC
excl.1824.564.910.951158.007.700.92
21291.127.470.921340.388.050.89
3847.104.500.95920.694.990.93
4853.984.640.95989.015.680.91
5956.525.500.921045.816.050.89
6838.514.970.94930.094.990.92
7938.364.960.911113.215.890.90
8922.934.940.91990.385.240.90
Avg.934.135.230.931060.956.070.91
incl.11054.207.130.951217.587.450.92
22211.3913.590.902026.1111.990.85
31269.567.230.94935.214.900.94
41501.639.280.931550.039.200.88
51109.645.890.901525.658.160.87
6948.575.080.921138.565.890.91
7971.955.300.901413.267.590.90
8922.145.050.911125.585.870.90
Avg.1248.637.320.921366.507.630.90
Table 3. Performance of two Prophet-LSTM model setups with different predictor selections on entire test fold across different folds in RMSE, MAPE, and PCC.
Table 3. Performance of two Prophet-LSTM model setups with different predictor selections on entire test fold across different folds in RMSE, MAPE, and PCC.
Pred.FoldModel 1Model 2
RMSEMAPEPCCRMSEMAPEPCC
excl.1701.184.410.961604.7012.010.87
21077.496.830.951452.229.940.86
3674.923.970.961094.257.120.93
4704.014.310.961450.019.640.86
51215.837.780.861584.5810.390.76
6962.015.810.891105.996.840.85
7957.635.540.881063.146.250.84
8988.725.750.911021.105.810.88
Avg.910.225.550.921297.008.500.86
incl.11478.7410.980.931168.608.150.89
21878.1812.370.931596.429.910.90
31258.067.820.95770.804.550.95
42044.2413.940.902041.1014.020.88
51266.637.970.821722.6511.260.71
61045.576.320.861126.896.320.85
71192.927.590.841342.907.520.85
81172.867.210.861262.916.880.87
Avg.1417.159.270.891379.038.580.86
Table 4. Performance of Prophet-LSTM model trained with climate-change related predictors in RMSE, MAPE, and PCC.
Table 4. Performance of Prophet-LSTM model trained with climate-change related predictors in RMSE, MAPE, and PCC.
FoldGeneralWintertime
RMSEMAPEPCCRMSEMAPEPCC
11819.0710.910.922560.0115.660.94
21569.339.570.931448.728.110.91
3851.304.930.931069.905.900.95
4919.165.350.911181.796.230.93
51029.535.850.87958.655.230.91
61165.027.350.87858.014.840.93
71077.905.980.861265.336.650.89
81032.165.940.901096.535.970.89
Avg.1182.936.980.901304.877.320.92
Table 5. Performance of Prophet-LSTM trained with individual climate-change predictors in RMSE, MAPE, and PCC.
Table 5. Performance of Prophet-LSTM trained with individual climate-change predictors in RMSE, MAPE, and PCC.
PredictorTest SetRMSEMAPEPCC
dailyPrecipitationEntire1020.05.920.86
Winter947.455.090.90
dailyMeanTemperatureEntire1035.795.740.87
Winter991.355.120.90
dailyMeanWindspeedEntire988.065.840.87
Winter938.965.130.91
NAOEntire1079.086.010.86
Winter993.435.380.89
SICEntire1053.746.690.88
Winter1009.515.890.91
SSTEntire999.135.750.87
Winter902.044.820.91
geopotentialEntire1113.947.020.87
Winter1042.455.940.91
renewableEnergyRatioEntire1215.777.960.86
Winter1069.896.300.90
Table 6. Performance of Prophet-LSTM model trained with energy price predictors in RMSE, MAPE, and PCC.
Table 6. Performance of Prophet-LSTM model trained with energy price predictors in RMSE, MAPE, and PCC.
FoldGeneralWintertime
RMSEMAPEPCCRMSEMAPEPCC
1704.704.560.96812.674.530.95
21114.927.130.951248.887.280.92
3697.763.970.96777.213.810.95
4710.904.110.95873.964.790.95
51065.196.230.86987.405.790.91
6976.606.060.89873.035.310.94
7947.215.600.88927.275.000.91
8948.675.540.90960.235.170.91
Avg.895.745.400.92932.585.210.93
Table 7. Performance of Prophet-LSTM model with varying lags in evaluation metrics RMSE, MAPE and PCC.
Table 7. Performance of Prophet-LSTM model with varying lags in evaluation metrics RMSE, MAPE and PCC.
Days AheadGeneralWintertime
RMSEMAPEPCCRMSEMAPEPCC
60986.695.720.87926.115.060.91
75980.675.760.88895.734.900.91
90898.115.180.90696.313.730.95
105939.555.520.89771.914.180.94
1201529.968.760.821260.116.540.86
1351301.767.290.841182.486.270.86
1501112.056.250.851078.165.910.87
1651053.466.060.851043.455.740.88
1801033.695.980.86949.455.100.90
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

van de Sande, S.N.P.; Alsahag, A.M.M.; Mohammadi Ziabari, S.S. Enhancing the Predictability of Wintertime Energy Demand in The Netherlands Using Ensemble Model Prophet-LSTM. Processes 2024, 12, 2519. https://doi.org/10.3390/pr12112519

AMA Style

van de Sande SNP, Alsahag AMM, Mohammadi Ziabari SS. Enhancing the Predictability of Wintertime Energy Demand in The Netherlands Using Ensemble Model Prophet-LSTM. Processes. 2024; 12(11):2519. https://doi.org/10.3390/pr12112519

Chicago/Turabian Style

van de Sande, Susan N. P., Ali M. M. Alsahag, and Seyed Sahand Mohammadi Ziabari. 2024. "Enhancing the Predictability of Wintertime Energy Demand in The Netherlands Using Ensemble Model Prophet-LSTM" Processes 12, no. 11: 2519. https://doi.org/10.3390/pr12112519

APA Style

van de Sande, S. N. P., Alsahag, A. M. M., & Mohammadi Ziabari, S. S. (2024). Enhancing the Predictability of Wintertime Energy Demand in The Netherlands Using Ensemble Model Prophet-LSTM. Processes, 12(11), 2519. https://doi.org/10.3390/pr12112519

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop