Next Article in Journal
Local Deformation Analysis and Optimization of Steel Box Girder during Incremental Launching
Previous Article in Journal
Cyclic Behavior of Concrete-Filled Tube Columns with Bidirectional Moment Connections Considering the Local Slenderness Effect
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Short-Term Energy Forecasting to Improve the Estimation of Demand Response Baselines in Residential Neighborhoods: Deep Learning vs. Machine Learning

by
Abdo Abdullah Ahmed Gassar
Laboratory of Engineering Sciences for the Environment (LaSIE, UMR CNRS 7356), La Rochelle University, 17000 La Rochelle, France
Buildings 2024, 14(7), 2242; https://doi.org/10.3390/buildings14072242
Submission received: 27 May 2024 / Revised: 11 July 2024 / Accepted: 18 July 2024 / Published: 21 July 2024
(This article belongs to the Section Building Energy, Physics, Environment, and Systems)

Abstract

:
Promoting flexible energy demand through response programs in residential neighborhoods would play a vital role in addressing the issues associated with increasing the share of distributed solar systems and balancing supply and demand in energy networks. However, accurately identifying baseline-related energy measurements when activating energy demand response events remains challenging. In response, this study presents a deep learning-based, data-driven framework to improve short-term estimates of demand response baselines during the activation of response events. This framework includes bidirectional long-term memory (BiLSTM), long-term memory (LSTM), gated recurrent unit (GRU), convolutional neural networks (CNN), deep neural networks (DNN), and recurrent neural networks (RNN). Their performance is evaluated by considering different aggregation levels of the demand response baseline profile for 337 dwellings in the city of La Rochelle, France, over different time horizons, not exceeding 24 h. It is also compared with fifteen traditional statistical and machine learning methods in terms of forecasting accuracy. The results demonstrated that deep learning-based models, compared to others, significantly succeeded in minimizing the gap between the actual and forecasted values of demand response baselines at all different aggregation levels of dwelling units over the considered time-horizons. BiLSTM models, followed by GRU and LSTM, consistently demonstrated the lowest mean absolute percentage error (MAPE) in most comparison experiments, with values up to 9.08%, 8.71%, and 9.42%, respectively. Compared to traditional statistical and machine learning models, extreme gradient boosting (XGBoost) was among the best, with a value up to 11.56% of MAPE, but could not achieve the same level of forecasting accuracy in all comparison experiments. Such high performance reveals the potential of the proposed deep learning approach and highlights its importance for improving short-term estimates of future baselines when implementing demand response programs in residential neighborhood contexts.

1. Introduction

The European residential sector accounts for approximately 75% of European buildings and is solely responsible for over 25% of the final energy demand in the European Union (EU), making it the second-largest consumer after transport [1]. With this in mind, increasing the energy efficiency of residential and non-residential buildings is one of the main objectives of EU strategies to achieve the ambitious endeavor of decarbonizing European buildings [2]. Given that residential buildings are significant contributors to global carbon emissions, there is great interest in creating a low-carbon residential sector in Europe [3]. The widespread deployment of advanced demand-side management strategies through demand response programs, enabling flexible energy demand in European residential buildings, is seen as a promising direction to maximize energy efficiency while meeting comfort requirements. Demand response programs focus significantly on the increased integration of low-carbon energy generation systems, such as distributed solar photovoltaics systems [4], and the modification of natural energy usages for residential buildings in response to fluctuations in supply and demand when the reliability and security of energy network systems are compromised [5]. At the same time, end-use customers become participators in demand response programs by modulating their natural consumption patterns according to electricity prices or through corresponding payment incentives in response to control signals issued by energy network operators or aggregators [6,7].
Thus, various types of demand response scenarios/programs, including load-shifting, valley-filling, peak-clipping, and the shaping of flexible loads, have been introduced [6,8] to accurately operate and manage the available local energy resources considering outdoor weather conditions, consumption patterns, and network security. In parallel, demand response strategies that reduce a building’s energy demands during stressful times of the energy network are seen as feasible approaches to harness flexible energy demand without the need for substantial investments [9]. Despite the transition towards greater energy efficiency in residential buildings, driven by the widespread adoption of demand response programs, incomplete and significant improvements are still ongoing in this direction. However, there is a strong need to understand the baseline demand trends (i.e., so-called demand response baselines) for energy in residential buildings on the side of end-use customers/occupants and the interaction with the energy network. This requires providing an accurate estimation of demand response baselines, which would be consumed by end-use customers in the absence of demand response programs [10]. Demand response baselines serve as a fundamental reference point for measuring, optimizing, and assessing the potential reduction in energy demand during response events [11]. In contrast, demand response baselines in buildings are distinguished by highly fluctuating with non-linear characteristics due to the nature of consumption, which depends on occupancy, the culture of a particular building, the working schedule of each building, and outdoor weather conditions [12]. Thus, developing a data-driven learning framework to characterize baseline demand patterns and provide accurate estimates of demand response baselines, enabling the calculation of energy reductions in the context of residential buildings, is crucial [13,14].
In response, energy demand forecasts are a pivotal component of demand response programs to investigate the effectiveness of demand response scenarios and maximize their benefits. Specifically, accurate forecasts of energy demand in the short term and very short term can be employed to address different types of challenges at both the building level and the energy network level [15]. Common building-level challenges that can be addressed include tracking progress in energy efficiency improvements and defining abnormal behaviors and deviations in expected energy usage patterns, which enable the detection of potential energy losses, breakdowns, and inefficiency within the building’s systems [16,17]. Energy network-level challenges are short-term optimal scheduling and identification of the optimal energy flow to meet expected demand [18], facilitating increased integration of low-carbon energy sources into the energy network [19], and demand response flexibility optimization by tracking the improvement progress in energy reductions [20]. In demand response contexts, accurate short-term forecasts of demand response baselines would be utilized by aggregators (intermediaries between end-use consumers and energy utility suppliers) to support the fair compensation of participating households in demand response programs [6]. Accurate demand response baselines can also serve as essential information for resource planners and energy system operators interested in implementing demand response programs with high effectiveness [14]. However, how to develop an accurate and reliable data-driven framework that can fulfill the above-gap remains a difficult task [21]. Therefore, this study aims to address this research need by developing a short-term forecasting framework for demand response baselines based on time-series data. This is essential for any optimal demand response strategy, particularly in large-scale residential neighborhoods that would interact with intermittent renewables (i.e., low-carbon energy sources, such as distributed solar photovoltaic systems).

2. Literature Review

Over the past few years, several research studies have been devoted to accurately estimating demand response baselines at both the individual customer-level and the aggregate-level, as summarized in Table 1. Various methods have been employed, which can be categorized into statistical and traditional machine learning methods. Concerning statistical methods, Ghasemi et al. [22] and Wijaya et al. [23] introduced the averaging-based method (XofY method), which is based on historical datasets to investigate their effectiveness in providing accurate estimates of the demand response baselines for 32 industrial and 782 residential customers in Iran and Switzerland, respectively. Despite the importance of this work, the main limitation of these methods is that they tend to provide a simplified and less accurate representation of historical energy demand data and may not accurately capture the nuances and variations in demand patterns, which can lead to sub-optimal estimates for demand response baselines. Similarly, Zhang et al. [24] and Wang et al. [25] proposed using residential consumption of non-participants in demand response programs to estimate baselines of demand response participants. The problem with such a practical approach is that there must be reference buildings with similar characteristics that do not participate in demand response actions. Furthermore, this approach becomes problematic under frequent demand response actions. In the same context, the authors in [10,26,27,28,29,30] presented statistical regression with external inputs, such as weather variables as predictive factors, to perform predictions of demand response baselines. The results revealed that statistical regression models have significant potential to provide accurate estimates of demand response baselines. However, a drawback of statistical regression-based models is inadequately quantified uncertainty in predicting energy demand baselines due to their inability to measure non-linear relationships between energy demand and relevant influencing factors such as consumer behaviors and ambient weather conditions [31].
As a potential procedure to overcome the problems associated with statistical methods, several researchers have proposed a diverse combination of traditional machine learning methods to construct accurate demand response baselines. In this context, Chen et al. [14] proposed a support vector regression (SVR) method to estimate demand response baselines for office buildings, using factors such as weather and building working schedules as inputs to the SVR models. Similarly, Li et al. [32] proposed an SVR method to estimate customer demand response baselines in the presence of integrated distributed photovoltaic systems. Srivastav et al. [33] and Zhang et al. [34] proposed the development of predictive models based on the Gaussian Mixture Regression (GMR) method to characterize demand response baselines of building clusters. However, the GMR method has difficulties in processing time series data and requires the use of complex algorithmic models to calculate demand response baselines. Bampoulas et al. [35] compared the performance of RF (random forests), MNN (multilayer neural networks), SVR, and XGBoost (extreme gradient boosting) methods in providing accurate estimates of residential energy demand response baselines. Similarly, Sha et al. [36] developed six types of predictive models based on multiple linear regression (MLR), SVR, RF, CatBoost, LightGBM (light gradient boosting machine), and ANN (artificial neural network) to improve the calculation of demand response baselines for commercial buildings over the next 24 h. Tao et al. [37] proposed a graph convolutional network (GCN) method to improve the estimation of aggregated demand response baselines, as its performance was compared with that of their counterparts from SVR, MLR, and averaging methods. This is in addition to the other methods proposed in [38,39,40] to improve the estimation of demand response baselines in buildings.
Notwithstanding the effectiveness of some data-driven machine learning methods in estimating demand response baselines, as mentioned above, these methods require substantial improvement by considering more external factors, such as occupancy and indoor environmental conditions, which can be difficult to acquire in the context of large-scale neighborhoods and district buildings. In addition, implementation strategies for demand response in buildings necessitate high-resolution forecasts (from hourly to daily), leading to the need for developing accurate forecasting models [12,41]. As demand patterns fluctuate randomly, inaccurate estimates of demand response baselines can lead to significant errors when aggregated to determine the total energy reductions caused by the activation of response events [37]. In the face of such challenges, deep learning methods have brought the issue of reliable and accurate estimates in short-term energy forecasting studies in the building sector back into the spotlight and have received considerable attention in recent years. Researchers have pointed out the great potential of these methods in providing accurate results for building energy demand forecasting [42,43]. Accordingly, this study introduces the deep learning approach as a potential candidate to improve the accuracy of residential demand response baseline estimates over a short-term forecast horizon. The aim is to develop a robust and reliable deep learning-based data-driven framework and evaluate its performance, considering different residential energy demand profiles in neighborhood buildings, in order to provide accurate estimates for aggregated demand response baselines over multiple time forecast horizons, not exceeding 24 h. To the author’s knowledge, no previous studies have applied bidirectional long-short-term memory (BiLSTM) and gated recurrent unit neural network (GRU) to estimate aggregated demand response baselines in a neighborhood context.

Contribution of the Study

In light of the study’s objective and considering the strengths and weaknesses identified in the literature, the main contribution of this work is as follows:
  • A data-driven framework is proposed to identify the most effective deep learning methods in providing accurate estimates of residential demand response baselines over various time-horizons, not exceeding 24 h. This provides a novel insight for a deeper understanding of the forecasting characteristics exhibited by different data-driven models.
  • Verify the change in model performance during the evaluation phase by considering the demand response baseline profile according to different aggregation levels of residential units and other input features. This investigation is essential for understanding the different behaviors of forecasting models and the importance level of input features.
  • The performance of the deep learning models is compared with that of the traditional statistical and machine learning models developed in this work, considering both the type of forecasting model and the expected margin of error. This comparison helps to identify the strengths and limitations of each model and method in the context of short-term demand forecasts for residential demand response baselines.
The rest of this paper is organized into four main sections as follows: Section 3 presents the deep learning methods proposed in this work. Section 4 describes the methodology of this work developed from the previously mentioned methods. Next, the findings obtained from this work are presented and discussed in Section 5, and finally, the main conclusions and potential future developments of this work are drawn in Section 6.

3. Proposed Forecasting Methods

Basic concepts of the deep learning methods proposed in this work are outlined as follows.

3.1. Deep Neural Networks

Deep neural networks (DNNs) are one of the most used and popular neural network architectures in energy demand forecasting, commonly known as multilayer perceptron neural network models, due to the inclusion of more hidden layers. They are widely regarded as a robust and effective tool for solving complex problems, including classifications and forecasting tasks, due to their capacity to learn and represent intricate non-linear relationships between input and output data [44]. To achieve this, the basic structure of DNNs consists of three types of successive layers: input layers, hidden layers, and output layers, as shown in Figure 1a. The back-propagation algorithm is used to train DNNs [45]. As shown in Figure 1a, the input layers receive input signals μ ( t τ 1 ) , μ ( t τ 2 ) , μ ( t τ 3 ) , …, μ ( t τ n ) , where τ 1 , τ 2 , τ 3 , , τ n are constants. The summation of control signals and the system’s outputs at time t are represented by u ( t ) . The weights that connect the input layers to the hidden layers are represented by w 11 1 , w 12 1 , , w 1 n 1 for the first neuron, w 21 1 , w 22 1 , , w 2 n 1 for the second neuron and w 31 1 , w 32 1 , , w 3 n 1 for the third neuron. The weights associated with hidden layer neurons with neuron q are denoted by w h 1 1 , w h 2 1 , , w q n 1 , where q denotes the number of neurons. The weights that connect hidden layers to an output layer are represented by w 21 , w 22 , , w 2 q [46,47]. These weights and connections between different layers enable the DNN to learn and make reliable and accurate predictions based on the input data. Thus, DNNs have gained significant attention in time-series forecasting of building energy demand, with residential buildings receiving a substantial part of this attention [48,49].

3.2. Convolutional Neural Networks

Convolutional neural networks (CNNs) are a unique class of advanced neural network methods that hierarchically perform convolutional operations on input time-series data. In recent years, CNNs have received increasing attention in building energy demand forecasting [48,50] due to their ability to capture time-series dependencies. The typical structure of CNNs consists of five layers, namely input, convolutional, pooling, fully connected, and output layers [51]. They are characterized by their capability to process and transform time-series datasets utilizing three building blocks. These include (1) convolutional layer, which implements two types of operations. (a) The first of these operations are convolutional operations, which require two components, called the kernel and the time-series data. The kernel implements convolution on the time series by moving from the beginning to the end of a series (i.e., in one direction), and the dot product between the kernel and corresponding parts of the series is computed. (b) The other operations involve the non-linear activation being applied to the final output of convolutional operations. The other building blocks are (2) a pooling layer to maintain the stability and prevent overfitting of the model and (3) a fully connected layer to perform the same duties in the conventional neural networks [52].
In this work, a 1-D convolutional neural network (ConvID) was utilized to extract the features from the time series data for demand response baselines in residential buildings. This network applies sliding convolutional operations along the sequence of one-dimensional time series data [53]. The proposed ConvID consists of five foundational layers: the convolutional layer, the pooling layer, the fully connected layer, the dropout layer, and the ReLU (relu) correction layer. The CNN network performance depends on the parameters of these layers. They consist of several components, including the number of filters in the layers, filter size, padding, stride, and batch size. Figure 1b displays the CNN architecture used in this work.

3.3. Recurrent Neural Networks

Recurrent neural networks (RNNs) are an advanced class of neural networks designed to overcome the disadvantages of a traditional neural network in accounting for temporal correlations and dependencies in data sequences [54]. They are also distinguished from other deep learning neural network architectures by their recurrent connections, which enable them to memorize the information from previous outputs and incorporate it into the computation of the current result [55]. Thus, the impact of recurrent neural network models has been remarkable in many disciplines, including energy demand forecasting, where the temporal order in the dataset is a fundamental feature in model design. RNNs are typically networks composed of standard recurrent connection cells, hidden states, and input and output layers. The input nodes in RNN models have no incoming connections, and the output nodes have no outgoing connections, while the hidden state nodes have both incoming and outgoing connections [52]. Each time, the RNN model updates the information in its memory according to the following Equation (1).
h t = f c ( W h h t 1 , x t )
where h t is the current hidden state at time t ; f c is the activation function, typically the hyperbolic tangent (tanh) or the rectified linear unit (relu); W h is the weight matrix for the recurrent connections; h t 1 is the previous hidden state at time t 1 ; and x t is the input at time t .
The architecture of the RNN is exhibited in Figure 2a. Each node represents a neuron for a single timestep. W1 represents the connection weight for inputs, W2 signifies the self-connection weight for each neuron, and W3 denotes the connection weight for outputs. The input data sequence is processed sequentially within the network based on time steps, and the weight coefficients are reused in a recycling fashion. The training process of an RNN model includes a forward pass and a backward pass. The forward pass of an RNN model mirrors that of a single-hidden-layer multilayer perceptron, except that the hidden layer in an RNN model receives activations from both the current external input and the hidden layer activations from the previous timestep [56]. The process of computing weight derivatives for an RNN during the backward pass is referred to as “backpropagation” through time. The advantage of RNN models in time-series forecasting is that they can predict not only the next time step but also multiple future time steps, making them versatile for different forecasting horizons.

3.4. Long Short-Term Memory

Long short-term memory (LSTM) is an upgraded variant for RNN architectures designed to overcome the vanishing gradient and gradient explosion problems in the long-sequence training process. For this reason, LSTMs can be trained utilizing the time-series dataset to make predictions for the future energy demand of buildings, each time utilizing the historical dataset processed by the LSTM cells. Typically, the architecture of LSTM is composed of frequently interconnected subnetworks called memory blocks, as shown in Figure 2b. Each block consists of a forget gate, an input gate, an output gate, and one or more self-connected memory cells [57]. During the training of LSTM models, the LSTM gates facilitate the long-term storage and retrieval of information within the memory cells, effectively addressing the problem of vanishing gradients [58]. For instance, if the input gate remains closed (i.e., with an activation close to 0), the cell’s activation persists and is not overwritten by new inputs entering the network. Consequently, this information can be retained and made accessible to the network at a later point in the sequence by simply opening the output gate.
Significantly, the values of the forget gate and input gate are influenced by both the previous hidden state and the current input. The operation of these units in LSTM is parameterized as follows.
I t = σ ( W i x x t + W i h h t 1 + b i )
f t = σ ( W f x x t + W f h 1 h t 1 + b f )
G t = t a n h ( W g h x t + W g h h t 1 + b g )
C t = ( f t × C t 1 ) + ( I t × G t )
O t = σ ( W o x x t + W o x h t 1 + b o )
H t = O t × t a n h ( C t )
where I t , f t , G t , and O t are the input, forget, update, and output gate activations at time t. The notations σ and tanh represent non-linear activation functions that take values in the range the [0, 1] and [−1, 1], respectively. W * and b * are weight matrices and bias vectors specific to each gate, while C t is the cell state at time t.

3.5. Bidirectional Long Short-Term Memory

Bidirectional long short-term memory (BiLSTM) is an extension of the LSTM network to capture dependencies in both the past and future contexts of a given data point. With respect to prediction based on time series data, the multiple sequences of energy loads are highly time-dependent, and loads at any point in time are significantly correlated with loads at its previous and subsequent points in time, requiring a deeper temporal-feature-extractor [59]. Compared to the unidirectional-state-transmission in LSTMs, BiLSTM consists of two LSTM layers, namely the forward LSTM and the backward LSTM, as shown in Figure 3a, and the output is jointly identified by the states of these two LSTMs [12,60]. This configuration achieves bidirectional time series feature extraction, which can fully exploit the temporal correlations of energy load sequences. By fusing two bidirectional LSTM layers, the output outcome is computed as follows.
h t = L S T M ( x t , h t 1 )
h t = L S T M ( x t , h t + 1 )
O t = σ ( W o h   h t + W o h   h t + b o )
Here, L S T M represents the LSTM function, while W and W are the weight matrices of the forward and backward LSTMs for the computed output outcome.

3.6. Gated Recurrent Units

Gated recurrent units (GRUs) are a simplified alternative to LSTMs, designed to alleviate the computational burden associated with the large number of parameters in the LSTM network [61]. Compared with LSTMs, the GRU architecture has only two gates, namely the update gate and the reset gate (see Figure 3b) [62]. Typically, the update gate dictates what to discard from the memory of the previous unit, and the reset gate guides the fusion of the new input with the last memory [63]. Based on both the previous output h t 1 and the current input x t , the functioning principle of the GRU cell is listed as follows.
Z t = σ ( W z h h t 1 + W z x x t + b z )
r t = σ ( W r h h t 1 + W r x x t + b r )
h t ¯ = t a n h ( W h h r t h t 1 + W h x x t + b h )
h t ~ = 1 Z t ) h t 1 + Z t   h t ¯
As exhibited in Equations (11)–(14), Z t , r t , h t ¯ , and h t ~ are the update gate, reset gate, output candidate vector, and the shared memory at time t, respectively, while is an element-wise product. As demonstrated in Equations (11)–(14), the shared memory h t ~ passes through various time steps to encode new information while discarding memories that are no longer relevant in terms of timing. Hence, the shared memory stores significant information over an extended period. r t defines how much information h t 1 should be retained. A smaller value for r t (close to 1) denotes that more information h t 1 is retained in h t ¯ . Then, the update gate Z t defines how much information h t 1 should be discarded. A bigger Z t indicates that more information h t 1 was not discarded, whereas a smaller Z t means that a significant amount of information h t 1 was ignored.

4. Methodology

As mentioned above, the primary objective of this study is to investigate the potential of deep learning methods in providing an accurate estimation of residential demand response baselines. This includes data preprocessing, input feature selection, hyperparameters, development of baseline models, forecasting, and evaluations. Therefore, the methodology of this study includes several steps consisting of the following:
  • Obtaining a representative profile of the baseline residential energy demand (i.e., demand response baselines) in building clusters, which vary in terms of household space size and occupant behavior in the absence of response events.
  • Pre-processing to understand how the demand response baselines of residential buildings correlate with potential input features, obtaining an input feature selection based on the Pearson Correlation Coefficient (PCC) technique [64,65].
  • Training the models on a dataset composed of all the input features processed during the feature engineering stage and those determined as significant inputs by PCC in the input feature selection stage, with the error measured by performance indicators in the validation and evaluation stages.
  • Finally, the trained models and the specified input features are used to predict the future demand response baselines of residential units over various time horizons, not exceeding 24 h, and the results are utilized to assess the performance of each forecasting model.
The methodology can also be divided into the steps described as data acquisition (i.e., residential building and energy demand data inventory) and data preprocessing and input feature selection, as shown in Figure 4. The subsequent sections provide further details on the model training process, which involves the utilization of a rolling window, the hyperparameter selection, and performance evaluations.

4.1. Residential Building and Energy Demand Data Inventory

To investigate the performance of the proposed forecasting methods, 337 dwellings in the Atlantech of La Rochelle City, France, were selected as a typical case study. In order to secure training data for developing forecasting models, the baseline energy demand profile of 337 dwelling units was generated by simulating the real dwellings in DIMOSIM (District MOdeller and SIMulator), a Python-based urban building energy simulation platform (more details in [66,67]), the numeric datasets were saved in comma-separated values (CSV) files. The reason for this is that the measured data for the residential energy demand of all dwelling units, which must represent real demand response baselines in the absence of demand response events, are not currently available. Therefore, the DIMOSIM models were used to generate meticulously prepared datasets that accurately represent the behavior of existing dwellings exposed to the typical climatic conditions of Atlantech in the city of La Rochelle.
These datasets cover all aspects of dwelling units, with a particular focus on space heating, lighting, and electric appliances to represent the smart-meter readings of dwellings. They consist of 674 columns representing the most important information on the thermal behaviors of the dwelling and the different energy demand patterns during the heating season. In terms of size, the dataset contains a considerable amount of information due to both the five-month simulation duration (from November to the end of March) and the 10 min sampling interval. In this study, energy demand simulations for dwelling units during the heating season were carried out because of the significance of demand response programs in reducing non-essential electricity consumption during peak heating hours while at the same time providing economic benefits and promoting energy efficiency. The results of the DIMOSIM models are also compared with external references (for more details, see Ref. [68]), with DIMOSIM showing good agreement with the other tools in terms of energy production. As depicted in Figure 5, the black solid lines represent the average demand response baseline of all dwelling units, while both the red and blue solid ones are the metered heating loads and lighting and appliances during the heating months.
In terms of dwelling characteristics, the floor area of the dwelling units ranges from 38 m2 up to 225 m2, with an average size of 71 m2. The annual energy demand of each dwelling unit is approximately 105.85 kWh/m2. For space heating, the heating system consists of air-to-water heat pumps installed in each dwelling unit. In addition, each heat pump is equipped with its dedicated thermostat, enabling control over individual zones. The heat pump equipment is variable speed, and its coefficient of performance (COP) depends on the polynomial regression of the nominal performance coefficient to estimate the thermal power output as a function of both the radiator temperature (i.e., the sink) and the ambient temperature (i.e., the source). For the population and occupancy characteristics, each dwelling unit is occupied by either a couple with or without children or a single adult living alone. The dwelling occupancy characteristics include people who are employed, unemployed, students, or retired, as shown in Table 2.

4.2. Data Preprocessing

Since the baseline energy demand of the dwelling units involves separate profiles for space heating, lighting, and electric appliances in each dwelling, the initial step of data preprocessing was to appropriately integrate the dataset of demand response baselines for all dwelling units to create a unified dataset for analysis and model training. In the first stage, the data for space heating, lighting, and electric appliances are aggregated together for each dwelling. The next step is to determine the total and hourly energy demand for all dwelling units, as shown in Figure 6. On the other hand, the measured outdoor weather data collected at 1 h intervals in this study is not aligned with the generated dataset of demand response baselines, which had a 10 min sampling interval. Therefore, the data are reorganized using the technique of resampling and indexing time series to form unified readings. The reorganized dataset includes the data of demand response baselines, a total of 52,555 samples. This step is of great importance in order to facilitate the estimation of the correlation between demand response baselines for dwelling units and other related input features, as explained in Section 4.3. In this context, other new features (e.g., hour of the day or day of the week) are derived using a time-based feature engineering/temporal feature engineering technique. Feature engineering is used in this work due to its importance in the development of forecasting models based on time series data, as it directly affects the performance and accuracy of these models.
The final datasets, which included the hourly dataset for demand response baselines, weather factors, and other input features, were obtained. These datasets have not been subjected to the normalization process [0, 1] to ensure that all models work with the original data. They go through the input selection process to select the final inputs to the forecasting models and to understand how the future baseline energy demand of residential buildings correlates with the possible input features, as explained in Section 4.3.

4.3. Selection of the Input Features

Using the feature engineering technique results in datasets that have multiple attributes associated with the considered demand response baseline profile. Since there are multiple associated characteristics, it is essential to use the appropriate method to determine the importance of each of them to the demand response baselines of the dwelling units. Along with the measured outdoor weather factors (e.g., the outdoor temperature and direct solar radiation), the impact of sixteen (16) independent input features is also considered. Nine input features (9) capture previous energy demand patterns, encompassing the past 1 h, the past 2 hours, the past 3 hours, the past 24 h, the past 48 h, the past 72 h, and so on. The other seven input features (7) relate to the working schedule factors of the dwelling, such as the hour of the day, the day of the week, the day of the month, the day of the year, the week of the month, and the month of the year. In this context, both the Pearson Correlation Coefficient (PCC) and the Shapley Additive Explanation (SHAP) techniques are used to minimize redundant or useless input features and to identify the most important variables for the forecasting models.
PCC was used to estimate the correlation coefficients (R) between the average energy demand of all dwelling units in the district and each input feature, as described in Equation (15).
R = i = 1 n x i x ¯ + y i y ¯ i = 1 n x i x ¯ 2 i = 1 n y i y ¯ 2
where x i and y i represent the actual values, x ¯ and y ¯ donate the average values, and n is the number of observations. The correlation coefficient of PCC takes a value between (+1) and (−1), as the highest absolute value of PCC translates to a parameter closely related to the future energy demand of dwelling units. Figure 7a displays the PCC values between the baseline residential energy demand (i.e., the demand response baselines) and the most meaningful input features in descending order when considering the whole evaluation. As shown in Figure 7a, the factors associated with the previous energy demand patterns show strong positive correlations with the demand response baselines of the dwelling units. Weather-related outdoor temperatures also show strong negative correlations, while the other factors have relatively different correlation values. However, these factors with low PCC values were selected as input features for the forecasting models because deep learning-based data-driven models have a high ability to detect non-linear and linear relationships between the energy demand baselines of the dwelling units and other relevant influencing factors. Based on this, the use of small correlation values of the PCC ensures that the low-degree correlation is not neglected in model training.
As mentioned above, SHAP is also used to identify the potential contribution of each input feature to the forecasting model. Figure 7b shows the contributions of each input feature in descending order. It is important to note the importance of factors related to past energy demand patterns, outdoor temperature, and the hour of the day in forecasting baseline values of residential demand response over very short-term periods, not exceeding 24 h. Together with PCCs, these findings demonstrate the importance of including previous input features, including those related to previous energy demand pattern-related ones, for accurate forecasting of demand response baselines. The selection of the three (3) historical energy demand-related input features is based on a comprehensive analysis of factors that significantly affect energy demand forecasting. These features have been chosen to capture the diverse aspects of historical energy consumption patterns, ensuring a robust and accurate forecasting model. Accordingly, these nine factors out of all input features were selected as the final input features for all forecasting models, as described in Table 3.

4.4. Forecast Model Development

The development of deep learning-based forecasting models explained in Section 3, involves two fundamental steps: hyperparameter tuning and post-training. It is important to note that the performance of these models is compared with that of classic and ensemble models based on traditional statistical and machine learning methods used in previous literature [69,70,71,72]. The classic models included support vector regression (SVR), Autoregressive integrated moving average (ARIMA), multiple linear regression (MLR), Lasso regression (Lasso), ridge regression (Ridge), polynomial regression (PolyR), Bayesian regression (Bayesian), kernel ridge regression (KernelR), and stochastic gradient descent regression (SGDReg) algorithms. The tree-based ensemble machine learning models included extreme gradient boosting (XGBoost), light gradient-boosting machine (LightGBM), gradient boosting (GB), random forest (RF), Adaptive Boosting (AdaBoost), Bagging (Bagging), and categorical gradient boosting (CatBoost) algorithms. Both classic and ensemble algorithms are among the most popular traditional statistical and machine learning methods in building energy analysis applications (see Refs. [69,70,71]), which have become specifically data-driven methods for predicting, benchmarking, and mapping baseline energy demand in buildings [72].

4.4.1. Hyperparameter Tuning

Tuning the appropriate hyperparameters is a critical step in the development of accurate data-driven forecasting models and can have a considerable impact on convergence speeds and generalizability. In this work, the controlled-variable method, relying on empirical expertise, is utilized in the experiments to optimize the choice of hyperparameters. To determine the optimal architecture of the proposed models, this work referred to references [73,74,75] and identified the practicable scopes and the optimization scope for hyperparameters, as shown in Table 4. In terms of deep learning models, forecasting based on BiLSTM and CNN is presented here as a typical example. First, with default BiLSTM parameters (activation function = relu, optimizer = Adam, loss function = RMSE, batch size = 32, verbose = 2, learning rate = 0.0001, epochs = 30) and the default parameters of CNN (activation function = relu, optimizer = Adam, loss function = RMSE, filters = 64, kernel size = 2, pool size = 2, learning rate = 0.0001, verbose = 2, epochs = 30), the number of hidden units in each hidden layer was fixed. Second, the learning rate, epochs, batch size, filters, and kernel size were adjusted. For example, the RMSE value of BiLSTM models decreases when the learning rate is set to 0.01, and the number of epochs is set to 200 or 300. For CNN models, the RMSE value decreases when the learning rate is set to 0.001, and the number of epochs is set to 300 or 500. Third, in cases where there was a tendency for overfitting to occur, dropout was implemented based on empirical expertise. Dropout is a useful regularization technique to control the proportion of randomly selected neurons during each training iteration. This helps to prevent co-adaptation of neurons, making the network more robust and reducing overfitting. Finally, to minimize redundancy in the process of defining hyperparameters for optimization, parameters of the same model type in the same case were kept consistent, and adjustments were only made to those parameters that prevented overfitting.
For both classic and ensemble models, additional hyperparameters were employed to optimize performance and prevent overfitting, as shown in Table 4. One of these hyperparameters is max_depth, which regulates the maximum depth of each tree in the ensemble, thus limiting the complexity of individual trees. The second is n_estimators, which allows controlling the size of the ensemble in terms of the number of base learners while training the model. The third parameter is subsample, which determines the proportion of randomly sampled training data utilized to grow each tree within the ensemble. Another technique is the kernel coefficient of SVR, which specifically controls the influence of individual training samples on the decision boundary. All models developed (as described in Section 4.4.2) in this work were implemented using the Python programming language and executed within the Scikit-learn [76] and TensorFlow [77] frameworks. The experimental hardware configuration included an Intel(R) Xeon(R) Bronze 3106 CPU, a 64-bit operating system, and 64 GB of RAM on an Intel processor.

4.4.2. Training and Validation

Once the final input features have been specified, as explained in Section 4.3, it is crucial to develop robust and accurate models for predicting the future energy demand (demand response baselines) of dwellings over multiple time horizons, not exceeding 24 h. To achieve this goal, the datasets were divided into training and validation sets, comprising 80% and 20%, respectively. Considering the training dataset, each model is trained to learn patterns and trends of energy demand, as well as relationships between data series for all dwelling units and different dwelling unit aggregation levels—not a specific pattern of a particular dwelling unit. The goal is to know the performance of each model in the context of all dwelling units and different aggregation levels of dwelling units when traditional machine learning and deep learning models are considered. Each model is trained to predict the demand response baselines of dwelling units, starting at 0:00 am and ending at 11:00 pm (23:00 pm) of the day, as this lasts for the entire period under consideration (five months). Secondly, the trained model, together with its learned information, will later be used to make predictions about the future energy demand of these dwelling units.
With optimal hyperparameters, each model of traditional classic, ensemble, and deep learning models is trained, and its performance is evaluated using three different performance indicators described in Section 4.5. More specifically, using the training dataset, each model is trained with all input features (nine factors) previously defined in Section 3.3. Then, each model is saved at the optimal time for each trained dataset, and its performance is evaluated and compared with its other counterparts. This performance evaluation step is always based on the optimal hyperparameters for each trained model. The purpose of this step is to obtain accurate forecasting models. Subsequently, these forecasting models (i.e., pre-trained models) are used to periodically predict the future demand response baseline of dwelling units over various time horizons: 6 h (00:00 am to 06:00 am), 12 h (00:00 am to 12:00 pm), 18 h (00:00 am to 06:00 pm), and 24 h (00:00 am to 11:00 pm), as shown in Figure 8. In the forecasting step, each model only includes the input features (9) defined previously in Section 4.3., while the demand response baseline profile of the dwelling units was excluded. The aim is to determine the ability of each model to predict the future energy demand (demand response basis) of dwellings by considering only those factors related to previous energy demand patterns, dwelling working schedules, and outdoor weather conditions. Ultimately, the performance of those forecasting models is evaluated and validated using the validation dataset. In addition, the discrepancy between the values of the actual and forecasted demand response baselines was used to express the overall performance of the deep learning and traditional machine learning-based forecasting models by considering performance indicators, as described in Section 4.5.
To further investigate the forecasting behavior of the proposed deep learning and machine learning models, this work is extended to incorporate various aggregation levels of the demand response baseline profile (i.e., 200, 150, 100, 50, 30, 20, and 10 dwelling units). Each time, these models are trained and validated and then evaluated and compared in their performance at each level of aggregation for the demand response baseline profile of dwelling units. The motivation behind this is to better understand the forecasting behaviors of these models in the case of fluctuations or changes occurring in the profile of the demand response baselines of dwelling units. This can be performed by forecasting the energy demand profile of dwelling units [48,78] over a short-term horizon. All forecasting models, whether based on deep learning or classic and ensemble machine learning methods, had the same input features as defined in Section 4.3 and were compared in terms of forecasting accuracy, as discussed in Section 5.

4.5. Performance Assessment

The primary goal of adopting deep learning models, as well as classic and ensemble, is to minimize the gap between the actual aggregated demand response baselines and their forecasted counterparts for the next day. Therefore, the forecasting accuracy of aggregated demand response baselines over various time horizons (6 h, 12 h, 18 h, and 24 h ahead) was measured utilizing three performance metrics. The most commonly used statistical metric to determine forecast accuracy in the literature is the mean absolute percentage error (MAPE) [78,79]. However, MAPE has a limitation concerning the actual values of demand response baselines being zero, which can occur in forecasting demand response baselines. MAPE might take extreme values when the actual values are close to zero in some cases, making it less reliable. To avoid this limitation, this study also uses mean absolute error (MAE) and root mean squared error (RMSE), which do not have the above-mentioned limitations. The three metrics are formulated as follows.
M A P E ( % ) = 1 N i = 0 N y a c t u a l ,   i y f o r e c a s t , i y a c t u a l ,   i × 100
M A E k W = 1 N     i = 0 N y a c t u a l , i y f o r e c a s t ,   i
R M S E ( k W ) = i = 0 N ( y a c t u a l ,   i y f o r e c a s t , i ) 2 N
where y a c t u a l , i is the actual aggregated demand response baselines per hour i, y f o r e c a s t , i is the forecasted aggregated demand response baseline per hour i, and N is the number of hours in the datasets. The MAE reflects the magnitude of deviation between forecasted and actual values by utilizing the absolute error, while the RMSE refers to the standard deviation of the residuals between forecasted and actual values of demand response baselines for dwelling units. In contrast, the MAPE measures the forecasting accuracy between forecasted and actual values of demand response baselines for dwelling units, expressed as a relative percentage of forecasting errors. Both RMSE and MAE are scale-dependent metrics and describe the forecasting errors at their original scale. MAPE is a scale-independent metric because the denominator of its equation includes actual values, making it suitable for comparing performance with other studies. The lower values of these metrics mean that the dispersion is more similar between the actual and the forecast demand response baselines. In this study, the MAPE was used as the primary performance measure, with the MAE and RMSE used only as linkage breakers when the MAPE did not show a significant difference between the forecasting models.

5. Results and Discussion

In this section, the performance of the forecasting models developed based on deep learning algorithms is analyzed (Section 5.1.) and compared with those based on classic and ensemble machine learning algorithms (Section 5.2 and Section 5.3) over various time horizons, as explained in the previous sections. Section 5.4 shows the forecasting behaviors of all models developed over different aggregation levels of demand response baseline profile for dwelling units. Note that the performance results represent multiple testing outcomes for the developed predictive model, reflecting different test periods. Section 5.5 provides an example to assess the ability of deep learning and classic and ensemble machine learning methods in forecasting energy reductions resulting from the activation of demand response events for 3 h on peak days of the heating season.

5.1. Performance of Deep Learning Models

Table 5 lists the MAPE, RMSE, and MAE values for the forecasting performance of the proposed deep learning methods over various time horizons, compared to their classic and ensemble counterparts, as a function of the input features and each forecasting method. Given the performance results, the authors of this study evaluated the effectiveness of each model over multiple test periods, surpassing five consecutive periods. As shown in Table 5, BiLSTM-based forecasting models could considerably reduce the gap between the measured demand response baselines of dwelling units and their forecasted counterparties, considering different forecast horizons, leading to better performance. As a result, BiLSTM models outperformed their deep learning counterparts, including traditional neural networks (ANNs), with MAPE values of 9.08% (6 h ahead), 11.14% (12 h ahead), 11.11% (18 h ahead), and 11.59% (24 h ahead), respectively. The error represented by RMSE and MAE was also lower than ANN, DNN, CNN, and RNN in all cases of forecasting demand response baselines of dwelling units. The best conditions yielded an RMSE of 7.07 kW and an MAE of 5.41 kW when forecasting 6 h demand response baselines. This performance is attributed to the bidirectional processing (in both forward and backward directions) of BiLSTM, which allows the neural network to efficiently learn and capture information from both past and future states. Since the primary objective of BiLSTM is to acquire further knowledge concerning a given context by capturing it from more than one perspective and then concatenating the two outputs into a single contextual representation.
Compared to other deep learning models, GRU showed considerable improvements in forecast accuracy over all time horizons considered, better than LSTM, RNN, DNN, and CNN. The performance values of the GRU forecasts were better than those methods by 8.92% and 11.22% in MAPE for the 6 h and 18 h ahead forecasts. The LSTM models performed reasonably well in terms of forecasting accuracy, with MAPE values of up to 9.23% (6 h) and 12.03% (24 h). However, these improvements in the forecasting accuracy were slightly lower than those achieved by BiLSTM. For all four time horizons, BiLSTM performed better than GRU and LSTM, indicating that BiLSTM has a higher learning potential over the next 24 h due to its bidirectional learning capability and produces fewer errors than GRU and LSTM. On the contrary, this potential was found to be lowest with a traditional ANN-based forecasting model, which, due to its shallow structure, was not able to effectively learn time series data and produce accurate predictions for demand response baselines of dwelling units over the four time horizons.
In Figure 9 and Figure 10, four time horizons have been illustrated to provide an intuitive representation of the forecast, each encompassing a single test period. Figure 9 illustrates a comparison of the efficiency of the deep learning methods in representing the closeness between the curves of forecasted and actual values for the demand response baselines of dwelling units over the four time horizons. As shown in Figure 9, these methods were able to produce a similar profile of demand response baselines for dwelling units, with observable differences in the efficiency of each. Figure 10 shows a comparison of the proposed deep learning methods in terms of the magnitude of the error at each hour, in kW, and at different time horizons. As depicted in Figure 10, it illustrates how the best model reduces the magnitude of an error when producing estimates of the demand response baselines over 6 h, 12 h, 18 h, and 24 h ahead forecasts. The magnitude of the forecast error decreased to some extent as the length of the input interval decreased. For each model, the 6 h input produced the lowest forecast error. However, in terms of conciseness, the BiLSTM, followed by the GRU and LSTM models, outperformed the others in all cases. As accuracy is very important in demand response program applications, BiLSTM, GRU, and LSTM models would be preferred over ANN and DNN models for short-term baseline energy forecasting in residential buildings. Therefore, BiLSTM, GRU, and LSTM, along with the RNN and CNN as alternative methods, should be one of the deep learning methods to consider when developing baseline demand response forecasting models for dwelling units.

5.2. Comparison of Deep Learning and Ensemble Models

As shown in Table 5, the performance of seven ensemble methods, based on Bagging and boosting techniques, was assessed over various time horizons. Among these seven methods, XGBoost showed the best performance, with the lowest values of MAPE, RMSE, and MAE, when forecasting the demand response baselines of dwelling units over the four time horizons. As a result, the magnitude of error was lower than its counterparts, with values (8.14 kW, 13.95 kW, 14.39 kW, and 15.78 kW) and (10.28 kW, 17.61 kW, 18.07 kW, 19.84 kW) for MAE and RMSE, respectively. In contrast, the performance of AdaBoost and CatBoost was the worst, with (39.70%, 27.74%, 25.31%, and 23.27%) and (33.79%, 27.17%, 26.36%, and 25.09%) of MAPE, (22.45 kW, 24.76 kW, 24.21 kW, and 24.88 kW) and (24.76 kW, 29.88 kW, 30.29 kW, and 31.61 kW) of RMSE, and (20.34 kW, 21.42 kW, 20.58 kW, and 20.85 kW) and (20.84 kW, 24.62 kW, 24.79 kW, and 25.73 kW) of MAE over 6 h, 12 h, 18 h and 24 h ahead forecasts, respectively. Otherwise, other ensemble methods, such as GB, LightGBM, and RF, showed reasonable performance with slightly lower error variability. This is attributed to two main reasons: (1) the strong non-linear mapping generalization and parallelization potential of XGBoost, which is derived from its boosted decision tree-based architecture [80], and (2) the limited ability of other models to effectively learn from a given time series dataset and generalize outcomes has negatively affected the forecast accuracy, despite the implementation of hyperparameter adjustments.
Compared to deep learning methods, all ensemble models could not achieve the same performance observed in deep learning-based forecasting models. This difference is primarily due to the essential structural characteristics of these approaches. As ensemble algorithms without inherent memory, such ensemble methods are unable to capture and preserve past information. As a result, they exhibit suboptimal performance in scenarios where the input time-series information is intricate and where a shorter output interval is required. As shown in Figure 11, ensemble-based forecasting models can produce the general pattern of demand response baselines for dwelling units. However, the magnitude of forecast errors remains considerable, as depicted in Figure 12. For example, the total magnitude error of CatBoost models was 29.88 kW, 30.29 kW, and 31.61 kW of RMSE, whereas that was 15.12 kW, 17.79 kW, and 18.27 kW of RMSE for traditional artificial neural network (ANN) of deep learning models for 12 h, 18 h, and 24 h ahead forecasts, respectively.
To be more concise, in terms of forecasting accuracy, deep learning-based baseline models have exhibited superiority in reducing the magnitude error between forecasted and measured values of demand response baselines, thus demonstrating their advantage in short-term forecasts of demand response baselines for dwelling units. However, the XGBoost method should be one of the alternative forecasting methods to be considered, along with deep learning methods, when estimating demand response baselines in residential neighborhood contexts over various time horizons, not exceeding 24 h.

5.3. Comparison of Deep Learning and Classic Models

Compared to deep learning methods, the nine (9) classic methods proposed in this study showed no observable improvements in forecast accuracy over the four time horizons considered. Although methods such as SVR exhibited reasonable performance, they were far from achieving the performance obtained by deep learning-based forecasting models. As shown in Table 4, the best performance conditions for the SVR resulted in an RMSE of 16.16 kW and an MAE of 14.19 kW for the 6 h ahead forecasts. In addition, the RMSE and MAE values for ARIMA and Lasso were (26.97 kW and 24.85 kW) and (21.32 kW and 21.02 kW) at 12 h and 24 h ahead forecasts, respectively. Among the classic models, KernelR and SGDReg performed the worst compared with the other classic models, including SVR, ARIMA, and Lasso, with errors of magnitude of up to 33.21 kW and 28.27 kW (RMSE and MSE) and 28.35 kW and 24.51 kW (RMSE and MSE), respectively. The KernelR and SGDReg models often suffered from instability when using relatively diverse and sparse data samples (because they have different very short-term and short-term forecast horizons), resulting in considerable differences in forecast accuracy.
Figure 13 shows the demand response baseline profiles produced by classical approach-based forecasting models. It can be seen that using the classic models would produce the general pattern of demand response baselines for dwelling units. However, when the classic models are used, more deviations from the measured values of the demand response baselines are likely to be seen. This is due to the presence of more non-linear values in the input features, leading to error magnitudes of up to 36.88% (KernelR) and 41.29% (SGDReg) of MAPE. One of the drawbacks of the classic methods’ linear nature is that they prevent high-quality forecasts from being obtained with the original input features and, therefore, fail to achieve performance levels similar to those of deep learning models at the same time horizons. This is due to their inability to capture complex non-linear relationships in the demand response baselines’ datasets. In situations where the underlying models are highly non-linear or involve interactions between input features, linear models may not perform as well as more sophisticated non-linear models.
Figure 14 shows a comparison of the forecasted and measured value curves for one test period, where the maximum error amplitudes of the classic models were between (−40.52 and 12.73 kW), (−40.52 and 16.22 kW), (−40.52 and 36.77 kW), and (−40.52 and 34.26 kW) for forecasts at 6 h, 12 h, 18 h, and 24 h, respectively. At the same time, the maximum error magnitude of the deep learning models was between (−9.07 and 3.16 kW), (−12.91 and 9.31 kW), (−11.73 and 10.88 kW), and (−9.59 and 11.76 kW) at 6 h, 12 h, 18 h, and 24 h ahead forecasts, respectively. Two observations in this context can be made, as follows: (1) Deep learning methods demonstrated the best performance, highlighting their advantages in short-term residential energy forecasting. (2) Classic methods with regularization terms exhibited larger error fluctuations when forecasting demand response baseline values compared to their deep learning counterparts. Therefore, it can be concluded that the classic method failed to significantly reduce the forecasting error when producing accurate estimates of demand response baselines for dwelling units.

5.4. Performance at Different Aggregation Levels

To better understand the forecasting behavior of each method, different demand response baseline profiles were used at different levels of dwelling unit aggregation, namely, 200, 150, 100, 50, 20, and 10, as mentioned above. Each method has the same time horizons and input features, allowing analysis of the difference in forecast performance between different models with different levels of forecastability. Figure 15, Figure 16 and Figure 17 show the overall performance in terms of MAPE values for each of the deep learning, classic, and ensemble models over different time horizons. As depicted in Figure 15, the variations in the demand response baseline profiles for dwelling units indicate that deep learning models typically avoid overestimating future demand response baselines for different dwelling units. It was observed that deep learning-based forecasting models could achieve higher accuracy compared to other models, with reduced errors of up to 8.71%, 10.59%, 11.53%, and 13.05% of MAPE for 6 h, 12 h, 18 h, and 24 h ahead forecasts, respectively. In this respect, BiLSTM-based forecasting models demonstrated superior performance, followed by GRU, LSTM, and RNN, over different time horizons (see Appendix A and Appendix B). This behavior in forecasting performance is due to the robust learning capabilities of these methods in mapping energy data of dwelling units, resulting in lower errors than their counterparts from other deep learning methods.
Considering both classic and ensemble methods, the evaluation results demonstrated that the forecasting models based on these methods often tend to overestimate future demand response baselines when using different profiles of demand response baselines, leading to inaccurate forecasts. As shown in Figure 16 and Figure 17, SGDReg, ARIMA, KernelR, AdaBoost, and RF have considerable error variability, with MAPE values of (39.96%, 31.67%, 30.66%, 29.02%), (25.91%, 23.58%, 28.44%, 29.85%), (25.12%, 20.98%, 20.96%, 20.21%), and (37.91%, 27.63%, 26.11%, 24.63%); these are (49.94%, 75.10%, 111.65%, 96.08%), (70.38%, 63.03%, 70.62%, 67.06%), (94.24%, 72.88%, 74.16%, 65.96%), and (97.03%, 77.48%, 80.41%, 71.25%) when using demand response baseline profiles for 200 and 10 dwelling units over 6 h, 12 h, 18 h, and 24 h ahead forecasts, respectively. These considerable variations in error for each method are due to changes in the dataset of demand response baseline profiles. Consequently, the margin of forecasting behavior between these models and deep learning models is more apparent in such sparse datasets of demand response baselines because, for example, several classic and ensemble models, including Bagging and RF models, suffer more from instability in sparse datasets [81].
Intuitively, the forecasting behavior of the XGBoost, GB, LightGBM, and SVR models was better in terms of forecasting accuracy compared to their counterparts of other classic and ensemble models, as these models enabled better learning from different datasets of demand response baseline profiles (see Table A1, Table A2, Table A3, Table A4, Table A5 and Table A6 in Appendix B). However, these models could not outperform the deep learning models in terms of improving forecasting accuracy. This result was expected due to (1) the stability of deep learning models when dealing with sparse datasets and (2) the ability of deep learning models to better learn and efficiently deal with the complexity of sequential datasets (time series data) based on training data. Concerning the values of the accuracy metrics, using the demand response baseline profile for different aggregated dwelling units, for example, 200 dwelling units, results in (10.10%, 10.59%, 11.53%, 13.05%) of MAPE with BiLSTM, in contrast to (19.73%, 18.53%, 20.03%, 19.91%) and (11.05%, 14.12%, 15.45%, 16.02%) with the SVR and XGBoost models over 6 h, 12 h, 18 h, and 24 h ahead forecasts, respectively.
Furthermore, the RMSE and MAE for the deep learning models were lower compared to the classic and ensemble models, demonstrating better forecasting performance than others, as shown in Figure A1, Figure A2, Figure A3 and Figure A4 from Appendix A and Table A1, Table A2, Table A3, Table A4, Table A5 and Table A6 from Appendix B. Therefore, it can be said that the proposed deep learning methods consistently outperform the comparative classic and ensemble methods for forecasting demand response baselines at different aggregation levels of dwelling units, indicating accurate forecasting behaviors. The significant improvement in MAE and RMSE values primarily demonstrates the ability of the deep learning methods to effectively learn the complexity of time series datasets, which enables the correct capture of the data points of demand response baselines for different aggregations of dwelling units. It can be seen that the proposed deep learning models outperform the other models, as shown in Table A1, Table A2, Table A3, Table A4, Table A5 and Table A6. The deep learning models can adequately capture the dynamic stochastic nature of the aggregated demand response baselines, caused by the outdoor weather conditions and the energy demand behavior of the occupants, represented by the working schedules (calendar factor) of the dwelling. As a result, the gap between measured and forecasted aggregated demand response baselines was minimized to some extent, and good forecasting accuracy levels were achieved.

5.5. Example for Demand Response Forecasts

In order to integrate the forecasting models developed so far for demand response baselines into the residential energy management systems, it is necessary to evaluate their performance not only for the baseline demand patterns but also during the activation of demand response events (i.e., assess their ability in capturing energy reductions when response events are triggered). Energy reductions are calculated based on the difference in energy values between the baselines and the energy profile while the response events are triggered. To do so, datasets of the energy demand profile of all dwelling units when demand response events are activated for 3 h per day (i.e., from 6:00 pm to 9:00 pm) during the coldest days are utilized (see Figure A5 in Appendix A). The proposed deep learning models are also used to provide an accurate estimate of the energy reductions resulting from demand response events during these three hours, and their performance is then evaluated against the classic and ensemble models by considering performance metrics. Table 6 presents the performance results of the deep learning, ensemble, and classic models as a function of the given time horizon, input features, and for each proposed forecasting method. As expected, the deep learning-based forecasting models demonstrated better performance in capturing the energy reductions of the dwelling units due to demand response events over a 3 h time horizon. The LSTM models, followed by ANN, GRU, and BiLSTM, showed a significant ability to improve the forecasting accuracies by minimizing the gap between forecasted and measured values, up to 8.02%, 9.56%, 9.59, and 9.89% of MAPE, respectively. Table 6 also shows that the accuracy of other forecasting models based on deep learning is acceptable. The highest error magnitude of CNN is 7.55 kW of RMSE and 7.15 kW of MAE, making it the least accurate of the deep learning models. However, the forecast outcomes are all close to the actual reductions in energy demand. In contrast, all classic and ensemble models, including SGDReg, KernelR, ARIMA, and LightGBM models, failed to improve forecasting accuracy, with errors up to 46.32%, 34.24%, 36.94%, and 23.15% of MAPE, respectively. Both RMSE and MAE values were also significant, around (29.33 kW, 25.96 kW, 26.12 kW, 16.35 kW) and (26.68 kW, 23.72 kW, 22.18 kW, 15.65 kW), respectively. Otherwise, XGBoost performed better than its other classic and ensemble counterparts, with an error of 13.12% of MAPE, 9.58 kW of RMSE, and 8.01 kW of MAE, which is the best. However, it was not able to achieve the same level of accuracy when forecasting the energy reduction of dwelling units during the next three hours (3 h ahead forecast).
For conciseness in this section, a notable observation was the satisfactory performance of the deep learning models compared to the classic and ensemble models in providing accurate estimates of energy reductions over a 3 h ahead. At the same time, the comparison of the actual profile of energy demand reductions with the forecasted profiles of each forecasting method showed that deep learning-based forecasting models can provide more accurate profiles of energy demand reductions resulting from demand response events over short-term and very short-term horizons. This result is due to the intrinsic characteristics of deep learning structures. XGBoost can also be used as an alternative method, along with deep learning methods to support demand response programs in providing accurate estimates of the energy reductions of buildings during the activation of residential neighborhood-level response events.

6. Conclusions

The paper presents the development of a deep learning-based, data-driven learning framework to provide accurate and reliable estimates of demand response baselines in a residential neighborhood context over short-term and very short-term time horizons. Several predictive models based on a deep learning approach, including ANN, DNN, CNN, RNN, LSTM, GRU, and BiLSTM, were developed to predict the future demand response baselines of 337 dwelling units and explore the influence of using different levels of aggregation (200, 150, 100, 50, 20, 10 dwelling units) for the demand response baseline profiles on the forecast accuracy. At the same time, all these methods are compared with fifteen different classic and ensemble methods to verify their potential to provide accurate and reliable estimates of demand response baselines over a time horizon not exceeding 24 h. The classic methods included MLR, Lasso, Ridge, PolyR, Bayesian, KernelR, SGDReg, and ARIMA, while the ensemble methods included XGBoost, LightGBM, GR, RF, Bagging, CatBoost, and AdaBoost.
In all these methods, firstly, the PCC technique is used to select the most significant variables (input features) that influence the energy demand baselines of dwelling units. Secondly, SHAP is used to identify the potential contribution of each input feature to the predictive model. This not only effectively reduces the dimensionality of the input parameters but also enhances the model’s running speed while ensuring the incorporation of scientifically and rationally chosen input features. Finally, the controlled-variable method, relying on empirical expertise, is utilized in the experiments to determine the best combination of hyperparameters for building a robust demand response baseline model. Several demand response baseline models were developed and then their performance was analyzed based on MAE, RMSE, and MAPE measured over energy demand baseline datasets from different dwelling aggregation levels to identify the most accurate models over multiple forecast horizons.
The results showed that deep learning-based forecasting models, in comparison with others, could significantly minimize the gap between the actual and forecasted values of demand response baselines at all the different dwelling unit aggregation levels over the time horizons considered. The ANN, DNN, CNN, RNN, LSTM, GRU, and BiLSTM models consistently showed the smallest MAE, RMSE, and MAPE in all comparison experiments, with values up to (6.49 kW, 5.89 kW, 6.92 kW, 7.85 kW, 5.47 kW, 5.81 kW, 5.41 kW), (7.30 kW, 7.17 kW, 7.35 kW, 8.03 kW, 8.93 kW, 7.78 kW, 7.07 kW), and (10.18%, 8.86%, 9.15%, 9.92%, 9.23%, 8.92%, 9.08%), respectively. The BiLSTM models, followed by the GRU and LSTM, had the highest forecasting accuracies, as demonstrated by their superiority in most demand response baseline forecasting experiments. Compared to the performance of classic and ensemble models, XGBoost-based models were among the best for demand response baseline forecasts at different dwelling aggregation levels over the four time horizons considered. Meanwhile, KernelR, SGDReg, ARIMA, CatBoost, and AdaBoost were among the worst models when forecasting demand response baselines of dwellings. The classic and ensemble models could not achieve the same level of forecast accuracy in all comparative experiments over the time horizons considered. The optimal combination of hyperparameters, such as hidden layers and hidden units, was sufficient to characterize the different underlying patterns of demand response baselines for dwelling units in datasets. In some cases, the ANN, DNN, and CNN models suffered from instability but were able to self-regulate and achieve high performance in reliably and accurately forecasting residential demand response baselines. This is due to the training techniques associated with the deep learning approach, including the use of ReLU as the activation function and the dropout method for model regularization, which help to improve the forecasting performance of the neural network.
This work, in general, contributes to the body of knowledge on two levels. First, this research not only presents a performance comparison of each proposed method but also highlights the importance of employing advanced neural network models to improve the short-term and very short-term estimation of demand response baselines, as well as the energy reductions resulting from the implementation of demand response programs in residential neighborhood contexts. The MAE, RMSE, and MAPE values clearly show that method structure and other features can significantly influence the production of accurate and reliable demand response baselines. With this comparison, the predictions of residential demand response baselines can be expanded and associated with other issues to promote the efficiency of implementing demand response programs. Second, this work provides important insights into the domain of advanced deep learning-based energy reduction estimation. The results demonstrated that the neural network model with optimal hyperparameters can serve as a useful tool for enhancing demand response programs by providing accurate and reliable estimates of baseline values in residential neighborhoods. In future work, the author will focus on advanced hybrid neural network techniques and address the associated limitations of real datasets representing the real-world complexity of occupant behaviors. In addition, other features related to energy prices, building typology, such as thermal insulations of dwellings, and occupant behavior, can be incorporated along with concurrent prediction intervals to investigate the effect of uncertainties in the forecasting processes and improve forecast accuracy in the context of residential neighborhoods.

Funding

This research no received external funding.

Data Availability Statement

The datasets and analyses used in this paper will be made available on reasonable request.

Acknowledgments

The author would like to gratefully acknowledge the Engineering Sciences for the Environment Laboratory (LaSIE UMR—7356 CNRS) at La Rochelle University for facilitating this research through the excellent research network, thus enabling fruitful studies like the present one. The author also extends sincere appreciation to Jérôme LE DREAU for his invaluable notes, which improved this study.

Conflicts of Interest

The author declares no conflicts of interest.

Appendix A

Figure A1. RMSE values for forecasting models at the aggregation level of 200 dwelling units for the demand response baseline profile over four time horizons.
Figure A1. RMSE values for forecasting models at the aggregation level of 200 dwelling units for the demand response baseline profile over four time horizons.
Buildings 14 02242 g0a1
Figure A2. RMSE values for forecasting models at the aggregation level of 10 dwelling units for the demand response baseline profile over four time horizons.
Figure A2. RMSE values for forecasting models at the aggregation level of 10 dwelling units for the demand response baseline profile over four time horizons.
Buildings 14 02242 g0a2
Figure A3. MAE values for forecasting models at the aggregation level of 200 dwelling units for the demand response baseline profile over four time horizons.
Figure A3. MAE values for forecasting models at the aggregation level of 200 dwelling units for the demand response baseline profile over four time horizons.
Buildings 14 02242 g0a3aBuildings 14 02242 g0a3b
Figure A4. MAE values for forecasting models at the aggregation level of 10 dwelling units for the demand response baseline profile over four time horizons.
Figure A4. MAE values for forecasting models at the aggregation level of 10 dwelling units for the demand response baseline profile over four time horizons.
Buildings 14 02242 g0a4
Figure A5. Examples of residential energy demand profiles considering energy reductions caused by the activation of demand response events and predictions for proposed models.
Figure A5. Examples of residential energy demand profiles considering energy reductions caused by the activation of demand response events and predictions for proposed models.
Buildings 14 02242 g0a5aBuildings 14 02242 g0a5b

Appendix B

Table A1. Average performance of forecasting models at the aggregation level of 200 dwelling units for the profile of demand response baselines over four time horizons.
Table A1. Average performance of forecasting models at the aggregation level of 200 dwelling units for the profile of demand response baselines over four time horizons.
Model6 h Ahead12 h Ahead18 h Ahead24 h Ahead
MAE (kW)RMSE (kW)MAPE (%)MAE (kW)RMSE (kW)MAPE (%)MAE (kW)RMSE (kW)MAPE (%)MAE (kW)RMSE (kW)MAPE (%)
ClassicMLR10.5911.3928.3612.7814.7723.8512.7915.0225.0112.8515.3623.52
Ridge10.5611.3628.0512.7114.6923.6312.6514.8724.6312.7315.2323.22
Lasso10.3811.2926.8012.3414.3322.7212.0714.2723.0412.2214.6721.77
PolyR10.5811.3828.3312.7714.7623.8312.7715.0124.9612.8415.3923.89
Bayesian10.5911.3827.6012.7114.6923.3612.6014.8324.3112.6915.1922.94
KernelR14.5615.2537.1315.5717.4729.4414.5816.7628.3114.6717.0326.88
SGDReg16.0816.9539.9016.8819.1331.6715.7618.3430.6616.1218.8429.02
ARIMA10.1813.0625.9113.0917.0223.5814.9219.2628.4417.1621.7929.85
SVR8.189.4719.7310.8513.1518.5310.9613.3020.0311.6714.2619.91
EnsembleXGBoost5.066.2611.058.1910.3914.128.3610.6415.459.3911.9616.02
LightGBM7.729.5418.539.6912.2417.199.6112.1318.0410.1512.7917.65
GB6.798.7115.619.5612.2116.029.6612.3017.539.9912.7416.94
RF9.9211.5625.3211.1113.5320.9810.6413.1020.9611.1813.9220.21
Bagging9.7111.1925.5210.7712.9520.7910.4112.6920.9311.0613.6620.29
AdaBoost12.6613.7637.9112.7114.6227.6311.8814.0726.1112.2914.8424.63
CatBoost13.0015.0833.3414.3717.2827.3013.6916.7426.7214.2217.5925.63
Deep LearningANN4.845.9010.438.079.4612.527.979.7914.788.9410.7414.75
DNN4.585.799.527.618.2711.497.829.6613.868.049.7713.63
CNN4.695.5510.347.909.5513.167.579.6213.838.289.8813.81
RNN4.465.799.777.229.5111.207.759.2512.008.1710.5314.38
LSTM4.315.289.427.359.8313.227.349.3912.837.759.2313.12
GRU3.815.128.717.999.2112.827.619.1312.108.3610.5515.96
BiLSTM4.025.2910.106.568.2310.596.908.6211.538.229.0313.05
Table A2. Average performance of forecasting models at the aggregation level of 150 dwelling units for the profile of demand response baselines over four time horizons.
Table A2. Average performance of forecasting models at the aggregation level of 150 dwelling units for the profile of demand response baselines over four time horizons.
Model6 h Ahead12 h Ahead18 h Ahead24 h Ahead
MAE (kW)RMSE (kW)MAPE (%)MAE (kW)RMSE (kW)MAPE (%)MAE (kW)RMSE (kW)MAPE (%)MAE (kW)RMSE (kW)MAPE (%)
ClassicMLR8.659.3935.3510.4512.0727.1010.5212.3825.9710.6212.6124.16
Ridge8.649.3734.9810.4312.0526.8910.4612.3125.6810.5812.5623.99
Lasso8.429.2133.5310.3512.0626.5310.2612.1725.0610.3812.3823.35
PolyR8.659.3835.3110.4512.0727.0710.5112.3725.9310.6112.5924.13
Bayesian8.649.3734.3910.4512.0726.6210.4512.2925.4010.5512.5223.68
KernelR10.8311.5342.1312.2413.7931.7711.7313.4729.0111.9513.8527.47
SGDReg10.6111.4637.2412.9615.1330.8312.8915.2130.7712.7715.1828.12
ARIMA7.8610.0628.0910.8914.0124.9411.8315.0826.7213.8317.6228.21
SVR6.447.5421.928.7310.6519.398.9610.9819.839.4611.5819.29
EnsembleXGBoost4.295.4113.187.849.9315.697.939.8916.127.9810.0816.54
LightGBM6.097.5520.638.1210.2418.468.3310.4719.098.5910.7918.16
GB5.417.0317.737.9610.2317.188.2010.5218.108.4710.8817.26
RF7.448.8326.519.4111.4522.509.2111.2821.959.4311.5820.68
Bagging7.228.4026.388.9810.8221.988.8810.7921.689.2311.2720.53
AdaBoost9.7310.7442.2110.4812.1530.2810.1711.9528.0710.2712.2025.76
CatBoost9.4411.4233.9311.4514.1428.0711.8414.5828.2012.2815.1926.75
Deep LearningANN4.095.1212.067.119.0414.336.478.1714.206.989.8815.54
DNN3.734.5012.646.799.0114.017.609.6415.687.459.4413.88
CNN3.644.4312.137.099.3414.747.259.3615.667.038.9215.03
RNN3.815.0211.995.587.4713.276.798.8714.657.529.4615.46
LSTM3.364.1111.296.218.2314.737.339.5215.886.508.3513.18
GRU3.284.0710.475.927.9413.307.379.3615.446.969.0415.16
BiLSTM3.674.3812.416.288.4414.927.179.1115.696.388.0814.54
Table A3. Average performance of forecasting models at the aggregation level of 100 dwelling units for the profile of demand response baselines over four time horizons.
Table A3. Average performance of forecasting models at the aggregation level of 100 dwelling units for the profile of demand response baselines over four time horizons.
Model6 h Ahead12 h Ahead18 h Ahead24 h Ahead
MAE (kW)RMSE (kW)MAPE (%)MAE (kW)RMSE (kW)MAPE (%)MAE (kW)RMSE (kW)MAPE (%)MAE (kW)RMSE (kW)MAPE (%)
ClassicMLR5.275.7926.746.717.9724.096.778.0925.286.788.1623.79
Ridge5.265.8026.506.687.9523.926.718.0424.986.748.1223.63
Lasso5.135.6626.496.467.7423.606.367.7124.026.397.7822.66
PolyR5.275.8026.726.717.9624.096.768.0825.246.778.1523.76
Bayesian5.295.8326.176.707.9723.736.688.0224.676.718.0923.27
KernelR7.447.9137.058.269.4430.567.879.1629.837.929.2728.39
SGDReg7.718.2635.568.7510.2131.108.5410.1532.408.5110.1930.19
ARIMA5.356.8325.927.129.2224.538.0310.3129.579.1711.5630.94
SVR4.365.1220.105.917.3019.736.027.4421.356.257.7320.66
EnsembleXGBoost3.904.6712.014.525.9214.834.726.0916.745.046.4616.79
LightGBM4.495.5420.325.537.0019.195.446.9520.175.617.1119.38
GB3.855.0716.195.427.1217.415.427.0919.025.527.1218.17
RF5.476.3626.456.097.5322.325.937.4022.916.097.6321.79
Bagging4.895.7423.815.747.1520.885.697.1122.035.937.3921.15
AdaBoost6.497.1937.156.637.8527.796.377.6427.236.517.9125.51
CatBoost6.277.3530.797.279.0526.697.158.9227.247.419.2526.23
Deep LearningANN2.813.4911.953.594.8213.933.985.5515.194.915.9115.56
DNN2.763.7711.713.944.4413.583.265.7314.014.225.3214.66
CNN2.533.4510.233.844.9713.053.795.1314.844.255.3415.49
RNN2.653.5610.593.114.3511.593.695.2313.314.425.6915.41
LSTM2.373.289.353.944.7013.523.785.2113.274.195.2714.65
GRU2.583.1211.323.874.3211.383.505.9813.404.475.8715.65
BiLSTM2.133.0810.423.184.3913.033.405.7913.534.745.0114.18
Table A4. Average performance of forecasting models at the aggregation level of 50 dwelling units for the profile of demand response baselines over four time horizons.
Table A4. Average performance of forecasting models at the aggregation level of 50 dwelling units for the profile of demand response baselines over four time horizons.
Model6 h Ahead12 h Ahead18 h Ahead24 h Ahead
MAE (kW)RMSE (kW)MAPE (%)MAE (kW)RMSE (kW)MAPE (%)MAE (kW)RMSE (kW)MAPE (%)MAE (kW)RMSE (kW)MAPE (%)
ClassicMLR3.193.5230.913.954.7326.043.984.9128.194.064.9427.04
Ridge3.163.5030.513.924.7125.783.944.7727.834.034.9126.84
Lasso3.033.3930.393.794.6225.773.744.6226.593.864.7725.61
PolyR3.193.5230.873.944.7226.013.974.8128.154.054.9327.01
Bayesian3.153.4930.013.914.7125.513.914.7527.464.014.8926.40
KernelR4.074.3838.254.845.6032.064.665.4832.924.795.6632.10
SGDReg4.164.5037.495.466.5635.415.576.7242.625.476.6438.87
ARIMA3.063.8628.714.616.0428.114.956.3733.915.516.9334.96
SVR2.562.9922.363.394.2320.563.434.3122.993.654.5722.87
EnsembleXGBoost1.682.0913.072.793.6117.332.933.7619.663.154.0619.93
LightGBM2.593.1322.183.224.0920.383.324.1923.083.444.3822.28
GB2.102.6817.623.104.0518.623.334.3322.453.424.4521.66
RF3.393.8432.903.634.4025.853.594.3926.993.714.5825.71
Bagging2.873.3526.933.394.2322.913.414.2524.973.584.4924.18
AdaBoost3.874.2343.033.964.6631.473.834.5931.503.914.7129.63
CatBoost3.694.3732.284.095.0926.704.165.1929.054.325.4228.45
Deep LearningANN1.481.9110.883.123.9913.992.813.8718.102.993.6518.44
DNN1.411.9210.372.753.6712.282.383.4717.452.483.5417.34
CNN1.651.9912.363.334.3214.492.793.2318.182.953.8217.58
RNN1.441.9312.032.833.7413.902.103.9916.712.693.5916.07
LSTM1.832.3613.932.583.4611.352.033.8216.052.803.6517.58
GRU1.511.9312.742.693.5012.962.163.9116.452.983.8716.03
BiLSTM1.341.8810.852.543.3611.692.203.1917.632.433.1815.36
Table A5. Average performance of forecasting models at the aggregation level of 20 dwelling units for the profile of demand response baselines over four time horizons.
Table A5. Average performance of forecasting models at the aggregation level of 20 dwelling units for the profile of demand response baselines over four time horizons.
Model6 h Ahead12 h Ahead18 h Ahead24 h Ahead
MAE (kW)RMSE (kW)MAPE (%)MAE (kW)RMSE (kW)MAPE (%)MAE (kW)RMSE (kW)MAPE (%)MAE (kW)RMSE (kW)MAPE (%)
ClassicMLR0.981.1123.641.471.8831.481.491.9134.771.562.0532.87
Ridge0.971.1023.341.461.8631.181.471.8934.481.552.0432.71
Lasso1.041.1726.161.501.8932.781.511.9236.281.582.0734.06
PolyR0.971.1123.611.471.8731.451.481.9134.731.562.0532.84
Bayesian0.961.0923.121.451.8530.971.461.8834.321.532.0132.49
KernelR1.391.5233.601.782.1639.521.762.1543.261.832.2840.52
SGDReg1.461.6034.892.443.2656.332.753.5877.942.563.4266.43
ARIMA1.381.7232.741.812.3137.422.062.6249.162.262.8847.77
SVR0.800.9618.201.291.7126.111.451.7930.561.441.9529.07
EnsembleXGBoost0.760.9212.821.451.8624.341.431.9228.621.401.8228.55
LightGBM0.841.0119.401.301.9227.331.541.9733.051.421.8930.78
GB0.851.0919.621.471.9730.511.552.0737.381.622.1734.50
RF1.311.4536.071.571.9337.111.601.9743.201.642.0938.98
Bagging0.941.1124.511.871.9731.221.551.8538.451.491.9535.11
AdaBoost1.421.5441.111.651.9840.821.692.0447.211.722.1242.89
CatBoost1.281.4830.721.672.0835.631.672.1140.751.732.2337.62
Deep LearningANN0.620.7511.591.351.7923.121.341.6826.131.181.6024.45
DNN0.670.8211.960.961.2717.471.081.3720.041.171.5424.36
CNN0.560.6411.231.211.7320.241.331.6926.991.291.6226.06
RNN0.420.578.031.041.4317.531.321.5825.311.341.7326.38
LSTM0.360.456.571.141.4818.431.371.5925.651.161.5922.74
GRU0.460.578.821.261.7520.451.371.7626.051.321.8426.66
BiLSTM0.520.649.381.291.8020.361.261.4924.221.021.0923.33
Table A6. Average performance of forecasting models at the aggregation level of 10 dwelling units for the profile of demand response baselines over four time horizons.
Table A6. Average performance of forecasting models at the aggregation level of 10 dwelling units for the profile of demand response baselines over four time horizons.
Model6 h Ahead12 h Ahead18 h Ahead24 h Ahead
MAE (kW)RMSE (kW)MAPE (%)MAE (kW)RMSE (kW)MAPE (%)MAE (kW)RMSE (kW)MAPE (%)MAEkWRMSE (kW)MAPE (%)
ClassicMLR0.580.6845.960.991.2949.761.011.3454.521.071.4051.49
Ridge0.570.6745.490.971.2849.121.021.3554.061.061.3951.32
Lasso0.610.7151.540.991.3052.261.031.3156.681.071.4153.54
PolyR0.580.6845.890.981.2949.681.011.3154.261.061.3951.46
Bayesian0.570.6745.270.971.2748.911.031.3353.911.051.3751.17
KernelR0.750.8457.381.071.3456.461.081.3662.601.151.4658.37
SGDReg0.730.8449.941.622.2775.101.912.57111.651.792.4496.08
ARIMA0.871.0970.381.251.6463.031.341.7370.621.521.9567.06
SVR0.530.6537.910.941.2443.681.021.3450.211.041.3947.35
EnsembleXGBoost0.540.6230.040.931.1939.991.011.3245.680.971.2946.09
LightGBM0.570.6744.040.981.2247.411.061.3954.221.021.3349.81
GB0.540.6937.811.021.4147.251.091.4556.341.111.5151.93
RF0.951.0394.241.121.3572.881.101.3574.161.161.4665.96
Bagging0.540.6542.840.991.2148.061.091.3356.751.011.3251.85
AdaBoost0.971.0497.031.171.4077.481.171.4080.411.181.4571.25
CatBoost0.750.8955.991.111.4256.761.131.4563.471.191.5358.31
Deep LearningANN0.430.5325.770.881.1335.810.981.2238.330.911.1939.43
DNN0.390.4822.810.871.1235.350.991.2338.640.811.0535.09
CNN0.480.5623.630.871.1636.510.911.1937.750.831.0735.55
RNN0.320.4318.310.721.0128.020.861.1435.280.881.1235.66
LSTM0.440.5927.710.710.9828.280.911.2437.680.841.1135.89
GRU0.400.5824.510.751.0428.870.921.2136.720.891.1839.41
BiLSTM0.400.5723.030.740.9928.330.861.1834.580.631.0734.47

References

  1. Pardalis, G.; Mahapatra, K.; Mainali, B. Comparing public-and private-driven one-stop-shops for energy renovations of residential buildings in Europe. J. Clean. Prod. 2022, 365, 132683. [Google Scholar] [CrossRef]
  2. Habib, M.; Timoudas, T.O.; Ding, Y.; Nord, N.; Chen, S.; Wang, Q. A hybrid machine learning approach for the load prediction in the sustainable transition of district heating networks. Sustain. Cities Soc. 2023, 99, 104892. [Google Scholar] [CrossRef]
  3. Chreim, B.; Esseghir, M.; Merghem-Boulahia, L. Energy management in residential communities with shared storage based on multi-agent systems: Application to smart grids. Eng. Appl. Artif. Intell. 2023, 126, 106886. [Google Scholar] [CrossRef]
  4. Junker, R.G.; Azar, A.G.; Lopes, R.A.; Lindberg, K.B.; Reynders, G.; Relan, R.; Madsen, H. Characterizing the energy flexibility of buildings and districts. Appl. Energy 2018, 225, 175–182. [Google Scholar] [CrossRef]
  5. Hu, M.; Xiao, F.; Wang, S. Neighborhood-level coordination and negotiation techniques for managing demand-side flexiblity in residential microgrids. Renew. Sustain. Energy Rev. 2021, 135, 110248. [Google Scholar] [CrossRef]
  6. Jadhav, P.; More, D.; Salkuti, S.R. Smart residential distribution energy management system with integration of demand response and aggregator. Clean. Responsible Consum. 2023, 9, 100115. [Google Scholar] [CrossRef]
  7. Xu, F.Y.; Zhang, T.; Lai, L.L.; Zhou, H. Shifting boundary for price-based residential demand response and applications. Appl. Energy 2015, 146, 353–370. [Google Scholar] [CrossRef]
  8. Balakumar, P.; Vinopraba, T.; Chandrasekarn, K. Real time implementation of demand side management scheme for IoT enabled PV integrated smart residential building. J. Build. Eng. 2022, 52, 1044485. [Google Scholar] [CrossRef]
  9. Chen, Y.; Xu, P.; Gu, J.; Schmidt, F.; Li, W. Measures to improve energy demand flexibility in buildings for demand response (DR): A review. Energy Build. 2018, 177, 125–139. [Google Scholar] [CrossRef]
  10. Ziras, C.; Heinrich, C.; Pertl, M.; Bindner, H.W. Experimental flexiblity identifcation of aggregated residential thermal loads using behind-the-meter data. Appl. Energy 2019, 242, 1407–1421. [Google Scholar] [CrossRef]
  11. Valles, M.; Bello, A.; Reneses, J.; Frias, P. Probabilistic characterization of electricity consumer responsiveness to economic incentives. Appl. Energy 2018, 216, 296–310. [Google Scholar] [CrossRef]
  12. Sekhar, C.; Dahiya, R. Robust framework based on hybrid learning approach for short-term load forecasting of building electricity demand. Energy 2023, 265, 126660. [Google Scholar] [CrossRef]
  13. Li, R.; Satchwell, A.J.; Finn, D.; Christensen, T.H.; Kummert, M.; Le Dreau, J.; Lopes, R.A.; Madsen, H.; Salom, J.; Henze, G.; et al. Ten questions concerning energy flexibility in buildings. Build. Environ. 2022, 223, 109461. [Google Scholar] [CrossRef]
  14. Chen, Y.; Xu, P.; Chu, Y.; Li, W.; Wu, Y.; Ni, L.; Bao, Y.; Wang, K. Short-term electrical load forecasting using the support vector regression (SVR) model to calculate the demand response baseline for office buildings. Appl. Energy 2017, 195, 659–670. [Google Scholar] [CrossRef]
  15. Zhang, L.; Chen, Y.; Yan, Z. Predicting the short-term electricity demand based on the weather variables using a hybrid CatBoost-PPSO model. J. Build. Eng. 2023, 7, 106432. [Google Scholar] [CrossRef]
  16. Lin, J.; Fernandez, J.A.; Rayhana, R.; Zaji, A.; Zhang, R.; Herrera, O.E.; Liu, Z.; Merida, W. Predictive analytics for building power demand: Day-ahead forecasting and anomaly prediction. Energy Build. 2022, 255, 111670. [Google Scholar] [CrossRef]
  17. Kazmi, H.; Fu, C.; Miller, C. Ten questions concerning data-driven modelling and forecasting operational energy demand at building and urban scale. Build. Environ. 2023, 239, 110407. [Google Scholar] [CrossRef]
  18. Zhou, M.; Yu, J.; Sun, F.; Wang, M. Forecasting of short term electric power consumption for different types buildings using improved transfer learning: A case study of primary school in China. J. Build. Eng. 2023, 78, 107618. [Google Scholar] [CrossRef]
  19. Ghenai, C.; Al-Mufti, O.A.A.; Al-Isawi, O.A.M.; Amirah, L.H.L.; Merabet, A. Short-term building electrical load forecasting using adaptive neuro-fuzzy inference system (ANFIS). J. Build. Eng. 2022, 52, 104323. [Google Scholar] [CrossRef]
  20. Correa-Florez, C.A.; Michiorri, A.; Kariniotakis, G. Robust optimization for day-ahead market participation for smart-home aggregators. Appl. Energy 2018, 229, 433–445. [Google Scholar] [CrossRef]
  21. Ebrahimi, J.; Abedini, M. A two-stage framework for demand-side management and energy savings of various buildings in multi smart grid using robust optimization algorithms. J. Build. Eng. 2022, 53, 104486. [Google Scholar] [CrossRef]
  22. Ghasemi, A.; Hojjat, M.; Saebi, J.; Neisaz, H.R.; Hosseinzade, M.R. An investigation of the customer baseline load (CBL) calculation for industrial demand response participants—A regional case study from Iran. Sustain. Oper. Comput. 2023, 4, 88–95. [Google Scholar] [CrossRef]
  23. Wijaya, T.K.; Vasirani, M.; Aberer, K. When bias matters: An economic assessment of demand response baselines for residential customers. IEEE Trans. Smart Grid 2014, 5, 1755–1763. [Google Scholar] [CrossRef]
  24. Zhang, Y.; Chen, W.; Xu, R.; Black, J. A cluster-based on method for calculating baselines for residential loads. IEEE Trans. Smart Grid 2016, 7, 2368–2377. [Google Scholar] [CrossRef]
  25. Wang, F.; Li, K.; Liu, C.; Mi, Z.; Shafie-khah, M.; Catalao, J.P.S. Synchronous pattern matching principle based residential demand response baseline estimation: Mechanism analysis and approach description. IEEE Trans. Smart Grid 2018, 9, 6972–6985. [Google Scholar] [CrossRef]
  26. Coughlin, K.; Piette, M.A.; Goldman, C.; Kiliccote, S. Statistical analysis of baseline load models for non-residential buildings. Energy Build. 2009, 41, 374–381. [Google Scholar] [CrossRef]
  27. Granderson, J.; Price, P.N. Development and application of a statistical methodology to evaluate the predictive accuracy of building energy baseline models. Energy 2014, 66, 981–990. [Google Scholar] [CrossRef]
  28. Walter, T.; Price, P.N.; Sohn, M.D. Uncertainty estimation improves energy measurement and verification procedures. Appl. Energy 2014, 130, 230–236. [Google Scholar] [CrossRef]
  29. Hatton, L.; Charpentier, P.; Matzner-Lober, E. Statistical estimation of the residential baseline. IEEE Trans. Power Syst. 2016, 31, 1752–1759. [Google Scholar] [CrossRef]
  30. Sharifi, R.; Fathi, S.H.; Vahidinasab, V. Customer baseline load models for residential sector in a smart-grid environment. Energy Rep. 2016, 2, 74–81. [Google Scholar] [CrossRef]
  31. Gassar, A.A.A.; Yun, G.Y.; Kim, S. Data-driven approach to prediction of residential energy consumption at urban scales in London. Energy 2019, 187, 115973. [Google Scholar] [CrossRef]
  32. Li, K.; Wang, F.; Mi, Z.; Fotuhi-Firuzabad, M.; Duic, N.; Wang, T. Capacity and output power estimation approach of individual behind-the-meter distributed photovoltaic system for demand response baseline estimation. Appl. Energy 2019, 253, 113595. [Google Scholar] [CrossRef]
  33. Strivastav, A.; Tewari, A.; Dong, B. Baseline building energy modeling and localized uncertainty quantification using Gaussian mixture models. Energy Build. 2013, 65, 438–447. [Google Scholar] [CrossRef]
  34. Zhang, Y.; Ai, Q.; Li, Z. Improving aggregated baseline load estimation by Gaussian mixture model. Energy Rep. 2020, 6, 1221–1225. [Google Scholar] [CrossRef]
  35. Bampoulas, A.; Pallonetto, F.; Mangina, E.; Finn, D.P. An ensemble learning-based framework for assessing the energy flexibility of residential buildings with multicomponent energy systems. Appl. Energy 2022, 315, 118947. [Google Scholar] [CrossRef]
  36. Sha, H.; Xu, P.; Lin, M.; Peng, C.; Dou, Q. Development of a multi-granularity energy forecasting toolkit for demand response baseline calculation. Appl. Energy 2021, 289, 116652. [Google Scholar] [CrossRef]
  37. Tao, P.; Xu, F.; Dong, Z.; Zhang, C.; Peng, X.; Zhao, J.; Li, K.; Wang, F. Graph convolutional network-based aggregated demand response baseline load estimation. Energy 2022, 251, 123847. [Google Scholar] [CrossRef]
  38. Park, S.; Ryu, S.; Choi, Y.; Kim, J.; Kim, H. Data-driven baseline estimation of residential buildings for demand response. Energies 2015, 8, 10239–10259. [Google Scholar] [CrossRef]
  39. Jazaeri, J.; Alpean, T.; Gordon, R.; Brandao, M.; Hoban, T.; Seeling, C. Baseline methodologies for small scale residential demand response. In In Proceedings of the IEEE Innovative Smart Grid Technologies—Asia (ISGT-Asia), Melbourne, Australia, 28 November–1 December 2016; pp. 747–752. [Google Scholar] [CrossRef]
  40. Schwarz, P.; Mohajeryami, S.; Cecchi, V. Building a better baseline for residential demand response programs: Mitigating the effects of customer heterogeneity and random variations. Electronics 2020, 9, 570. [Google Scholar] [CrossRef]
  41. Rahim, S.; Ahmad, H. Data-driven multi-layered intelligent energy management system for domestic decentralized power distributed systems. J. Build. Eng. 2023, 68, 106113. [Google Scholar] [CrossRef]
  42. Ramos, P.V.B.; Villela, S.M.; Silva, W.N.; Dias, B.H. Residential energy consumption forecasting using deep learning models. Appl. Energy 2023, 350, 121705. [Google Scholar] [CrossRef]
  43. Santos, M.L.; Garcia, S.D.; Garcia-Santiago, X.; Ogando-Martinez, A.; Camarero, F.E.; Gil, G.B.; Ortega, P.C. Deep learning and transfer learning techniques applied to short-term load forecasting of data-poor buildings in local energy communities. Energy Build. 2023, 292, 113164. [Google Scholar] [CrossRef]
  44. Fouladfar, M.H.; Soppelsa, A.; Nagpal, H.; Fedrizzi, R.; Franchini, G. Adaptive thermal load prediction in residential buildings using artificial neural networks. J. Build. Eng. 2023, 77, 107464. [Google Scholar] [CrossRef]
  45. Xu, Y.; Li, F.; Asgari, A. Prediction and optimization of heating and cooling loads in a residential building based on multi-layer perceptron neural network and different optimization algorithms. Energy 2022, 240, 122692. [Google Scholar] [CrossRef]
  46. Bourdeau, M.; Zhai, X.Q.; Nefzaoui, E.; Guo, X.; Chatellier, P. Modeling and forecasting building energy consumption: A review of data-driven techniques. Sustain. Cities Soc. 2019, 48, 101533. [Google Scholar] [CrossRef]
  47. Afzal, S.; Ziapour, B.M.; Shokri, A.; Shakibi, H.; Sobhani, B. Building energy consumption prediction using multilayer perceptron neural network-assisted models: Comparison of different optimization algorithms. Energy 2023, 22, 128446. [Google Scholar] [CrossRef]
  48. Shaqour, A.; Ono, T.; Hagishima, A.; Farzaneh, H. Electrical demand aggregation effects on the performance deep learning-based short-term load forecasting of a residential building. Energy AI 2022, 8, 100141. [Google Scholar] [CrossRef]
  49. Olu-Ajayi, R.; Alaka, H.; Suaimon, I.; Sunmola, F.; Ajayi, S. Building energy consumption prediction for residential buildings using deep learning and other machine learning techniques. J. Build. Eng. 2022, 45, 103406. [Google Scholar] [CrossRef]
  50. Wang, L.; Xie, D.; Zhou, L.; Zhang, Z. Application of the hybrid neural network model for energy consumption prediction of office buildings. J. Build. Eng. 2023, 72, 106503. [Google Scholar] [CrossRef]
  51. Shaikh, A.K.; Nazir, A.; Khalique, N.; Shah, A.-S.; Adhikari, N. A new approach to seasonal energy consumption forecasting using temporal convolutional networks. Results Eng. 2023, 19, 101296. [Google Scholar] [CrossRef]
  52. Khalil, M.; McGough, A.S.; Pourmirza, Z.; Pazhoohesh, M.; Walker, S. Machine learning, deep learning and statistical analysis for forecasting building energy consumption—A systematic review. Eng. Appl. Artif. Intell. 2022, 115, 105287. [Google Scholar] [CrossRef]
  53. Walser, T.; Sauer, A. Typical load profile-supported convolutional neural network for short-term load forecasting in the industrial sector. Energy AI 2021, 5, 100104. [Google Scholar] [CrossRef]
  54. Hewamalage, H.; Bergmeir, C.; Bandara, K. Recurrent neural networks for time series forecasting: Current status and future directions. Int. J. Forecast. 2021, 37, 388–427. [Google Scholar] [CrossRef]
  55. Amalou, I.; Mouhni, N.; Abdali, A. Multivariate time series prediction by RNN architectures for energy consumption forecasting, TRMEES22-Fr, EURACA Conference 2022. Energy Rep. 2022, 8, 1084–1091. [Google Scholar] [CrossRef]
  56. Wang, F.; Xuan, Z.; Zhen, Z.; Li, K.; Wang, T.; Shi, M. A day-ahead PV power forecasting method based on LSTM-RNN model and time correlation under partial daily pattern prediction. Energy Convers. Manag. 2020, 212, 112766. [Google Scholar] [CrossRef]
  57. AL-Alimi, D.; AlRassas, A.M.; Al-qaness, M.A.A.; Cai, Z.; Aseeri, A.O.; Elaziz, M.A.; Ewees, A.A. TLIA: Time-series forecasting model using short-term memory integrated with artificial neural networks for volatile energy markets. Appl. Energy 2023, 343, 121230. [Google Scholar] [CrossRef]
  58. Peng, C.; Tao, Y.; Chen, Z.; Zhang, Y.; Sun, X. Multi-source transfer learning guided ensemble LSTM for building multi-load forecasting. Expert Syst. Appl. 2022, 202, 117194. [Google Scholar] [CrossRef]
  59. Li, K.; Mu, Y.; Yang, F.; Wang, H.; Yan, Y.; Zhang, C. A novel short-term multi-energy load forecasting method for integrated energy system based on feature separation-fusion technology and improved CNN. Appl. Energy 2023, 351, 121823. [Google Scholar] [CrossRef]
  60. Song, Y.; Xie, H.; Zhu, Z.; Ji, R. Predicting energy consumption of chiller plant using WOA-BiLSTM hybrid prediction model: A case study for a hospital building. Energy Build. 2023, 300, 113642. [Google Scholar] [CrossRef]
  61. Xu, L.; Hu, M.; Fan, C. Probabilistic electrical load forecasting for buildings using Bayesian deep neural networks. J. Build. Eng. 2022, 46, 103853. [Google Scholar] [CrossRef]
  62. Zhang, X.; Zhong, C.; Zhang, J.; Wang, T.; Ng, W.W.Y. Robust recurrent neural networks for time series forecasting. Neurocomputing 2023, 526, 143–157. [Google Scholar] [CrossRef]
  63. Li, D.; Sun, G.; Miao, S.; Gu, Y.; Zhang, Y.; He, S. A short-term electric load forecast method based on improved sequence-to sequence GRU with adaptive temporal dependence. Int. J. Electr. Power Energy Syst. 2022, 137, 107627. [Google Scholar] [CrossRef]
  64. Meng, Y.; Yun, S.; Zhao, Z.; Guo, J.; Li, X.; Ye, D.; Jia, L.; Yang, L. Short-term electricity load forecasting based on a novel data preprocessing system and data reconstruction strategy. J. Build. Eng. 2023, 77, 107432. [Google Scholar] [CrossRef]
  65. Maltais, L.-G.; Gosselin, L. Forecasting of short-term lighting and plus load electricity consumption in single residential units: Development and assessment of data-driven models for different horizons. Appl. Energy 2022, 307, 118229. [Google Scholar] [CrossRef]
  66. DIMOSIM (DIstrict Modeller and SIMulator), CSTB (Centre Scientifique et Technique du Bâtiment). Available online: https://dimosim.cstb.fr/index.html (accessed on 19 November 2023).
  67. Garreau, E.; Abdelouadoud, Y.; Herrera, E.; Keilholz, W.; Kyriakodis, G.-E.; Partenay, V.; Riederer, P. District MOdeller and SIMulator (DIMOSIM)—A dynamic simulation platform based on a bottom-up approach for district and territory energetic assessment. Energy Build. 2021, 251, 111354. [Google Scholar] [CrossRef]
  68. Martinez, S.; Vellei, M.; Le Dreau, J. Demand-side flexibility in residential district: What are the main sources of uncertainty? Energy Build. 2022, 255, 111595. [Google Scholar] [CrossRef]
  69. Fathi, S.; Srinivasan, R.; Fenner, A.; Fathi, S. Machine learning applications in urban building energy performance forecasting: A systematic review. Renew. Sustain. Energy Rev. 2020, 133, 110287. [Google Scholar] [CrossRef]
  70. Sun, Y.; Haghighat, F.; Fung, B.C.M. A review of the-state-of-art in data-driven approaches for building energy prediction. Energy Build. 2020, 221, 110022. [Google Scholar] [CrossRef]
  71. Gassar, A.A.A.; Cha, S.H. Energy prediction techniques for large-scale buildings towards a sustainable built environment: A review. Energy Build. 2020, 224, 110238. [Google Scholar] [CrossRef]
  72. Wei, Y.; Zhang, X.; Shi, Y.; Xia, L.; Pan, S.; Wu, J.; Han, M.; Zhao, X. A review of data-driven approaches for prediction and classification of building energy consumption. Renew. Sustain. Energy Rev. 2018, 82, 1027–1047. [Google Scholar] [CrossRef]
  73. Yu, H.F.; Zhong, F.; Du, Y.; Xie, X.; Wang, Y.; Zhang, X.; Huang, S. Short-term cooling and heating loads forecasting of building district energy system based on data-driven models. Energy Build. 2023, 298, 113513. [Google Scholar] [CrossRef]
  74. Wang, Z.; Hong, T.; Piette, M.A. Building thermal load prediction through shallow machine learning and deep learning. Appl. Energy 2020, 263, 114683. [Google Scholar] [CrossRef]
  75. Sehovac, L.; Grolinger, K. Deep learning for load forecasting: Sequance to sequence recurrent neural networks with attention. IEEE Access 2020, 8, 36411–364226. [Google Scholar] [CrossRef]
  76. Fan, T.J.; Gramfort, A.; Grisel, O.; Halchenko, Y.; Hug, N.; Jalali, A.; Jerphanion, J.; Leraitre, G.; Mueller, A.; Metzen, J.-H.; et al. Scikit-Learn (Free and Open-Source Machine Learning Library), Version 1.1.3. 2007. Available online: https://scikit-learn.org/stable/ (accessed on 30 November 2023).
  77. Abadi, M.; Agarwal, A.; Barham, P.; Brevdo, E.; Chen, Z.; Citro, C.; Corrado, G.S.; Davis, A.; Dean, J.; Devin, M.; et al. Large-Scale Machine Learning on Heterogeneous Systems, Version 2.7.0; TensorFlow Organization (Free and Open-Source Software Library for Machine Learning and Artificial Intelligence): 2015. Available online: https://www.tensorflow.org/ (accessed on 30 November 2023).
  78. Dab, K.; Henao, N.; Nagarcheth, S.; Dube, Y.; Sansret, S.; Agbossou, K. Consensus-based time-series clustering approach to short-term load forecasting for residential electricity demand. Energy Build. 2023, 299, 113550. [Google Scholar] [CrossRef]
  79. Tofallis, C. A better measure of relative prediction accuracy for model selection and model estimation. J. Oper. Res. Soc. 2015, 66, 1352–1362. [Google Scholar] [CrossRef]
  80. Kumari, P.; Toshniwal, D. Extreme gradient boosting and neural network based ensemble learning approach to forecast hourly solar irradiance. J. Clean. Prod. 2021, 279, 123285. [Google Scholar] [CrossRef]
  81. Wang, Z.; Wang, Y.; Srinivasan, R.S. A novel ensemble learning approach to support building energy use prediction. Energy Build. 2018, 159, 109–122. [Google Scholar] [CrossRef]
Figure 1. Architecture of (a) deep neural network (DNN) and (b) convolutional neural network (CNN) methods.
Figure 1. Architecture of (a) deep neural network (DNN) and (b) convolutional neural network (CNN) methods.
Buildings 14 02242 g001
Figure 2. Architecture of (a) recurrent neural network (RNN) and (b) long short-term memory (LSTM) methods.
Figure 2. Architecture of (a) recurrent neural network (RNN) and (b) long short-term memory (LSTM) methods.
Buildings 14 02242 g002
Figure 3. Architecture of (a) bidirectional long short-term memory (BiLSTM) and (b) gated recurrent unit (GRU) methods.
Figure 3. Architecture of (a) bidirectional long short-term memory (BiLSTM) and (b) gated recurrent unit (GRU) methods.
Buildings 14 02242 g003
Figure 4. Research methodology used in this work, including simulations and model development.
Figure 4. Research methodology used in this work, including simulations and model development.
Buildings 14 02242 g004
Figure 5. Hourly baseline energy demand behaviors for dwelling units over the days of the five heating months, as the energy demand baselines are the average for space heating, appliances, and lighting.
Figure 5. Hourly baseline energy demand behaviors for dwelling units over the days of the five heating months, as the energy demand baselines are the average for space heating, appliances, and lighting.
Buildings 14 02242 g005
Figure 6. Energy demand distributions (demand response baseline profile) of dwelling units during the five heating months: (a) hourly energy demand and (b) total daily energy demand.
Figure 6. Energy demand distributions (demand response baseline profile) of dwelling units during the five heating months: (a) hourly energy demand and (b) total daily energy demand.
Buildings 14 02242 g006
Figure 7. Example of the input feature importance estimated by the PCC technique and determined by SHAP tests using values of demand response baselines for all dwelling units.
Figure 7. Example of the input feature importance estimated by the PCC technique and determined by SHAP tests using values of demand response baselines for all dwelling units.
Buildings 14 02242 g007
Figure 8. Process of training and testing deep learning and traditional machine learning models using the demand response baseline values of dwelling units.
Figure 8. Process of training and testing deep learning and traditional machine learning models using the demand response baseline values of dwelling units.
Buildings 14 02242 g008
Figure 9. Comparison of forecasted and actual demand response baselines for dwelling units based on the performance of deep learning models over various time horizons, considering a single test period.
Figure 9. Comparison of forecasted and actual demand response baselines for dwelling units based on the performance of deep learning models over various time horizons, considering a single test period.
Buildings 14 02242 g009
Figure 10. Comparison of forecast errors for deep learning models at each hour and over various time horizons, considering a single test period.
Figure 10. Comparison of forecast errors for deep learning models at each hour and over various time horizons, considering a single test period.
Buildings 14 02242 g010
Figure 11. Comparison of forecasted and actual demand response baselines for dwelling units based on the performance of tree-based ensemble models over various time horizons, considering a single test period.
Figure 11. Comparison of forecasted and actual demand response baselines for dwelling units based on the performance of tree-based ensemble models over various time horizons, considering a single test period.
Buildings 14 02242 g011
Figure 12. Comparison of forecast errors for ensemble models at each hour and over various time horizons, considering a single test period.
Figure 12. Comparison of forecast errors for ensemble models at each hour and over various time horizons, considering a single test period.
Buildings 14 02242 g012aBuildings 14 02242 g012b
Figure 13. Comparison of forecasted and actual demand response baselines for dwelling units based on the performance of classic models over various time horizons, considering a single test period.
Figure 13. Comparison of forecasted and actual demand response baselines for dwelling units based on the performance of classic models over various time horizons, considering a single test period.
Buildings 14 02242 g013aBuildings 14 02242 g013b
Figure 14. Comparison of forecast errors for classic models at each hour and over various time horizons, considering a single test period.
Figure 14. Comparison of forecast errors for classic models at each hour and over various time horizons, considering a single test period.
Buildings 14 02242 g014
Figure 15. Comparison of MAPE values for deep learning models considering different aggregation levels of the demand response baseline profile of the dwelling units over four time horizons: (a) 6 h, (b) 12 h, (c) 18 h, and (d) 24 h ahead forecasts.
Figure 15. Comparison of MAPE values for deep learning models considering different aggregation levels of the demand response baseline profile of the dwelling units over four time horizons: (a) 6 h, (b) 12 h, (c) 18 h, and (d) 24 h ahead forecasts.
Buildings 14 02242 g015
Figure 16. Comparison of MAPE values for ensemble models considering different aggregation levels of the demand response baseline profile of the dwelling units over four time horizons: (a) 6 h, (b) 12 h, (c) 18 h, and (d) 24 h ahead forecasts.
Figure 16. Comparison of MAPE values for ensemble models considering different aggregation levels of the demand response baseline profile of the dwelling units over four time horizons: (a) 6 h, (b) 12 h, (c) 18 h, and (d) 24 h ahead forecasts.
Buildings 14 02242 g016
Figure 17. Comparison of MAPE values for classic models considering different aggregation levels of the demand response baseline profile of the dwelling units over four time horizons: (a) 6 h, (b) 12 h, (c) 18 h, and (d) 24 h ahead forecasts.
Figure 17. Comparison of MAPE values for classic models considering different aggregation levels of the demand response baseline profile of the dwelling units over four time horizons: (a) 6 h, (b) 12 h, (c) 18 h, and (d) 24 h ahead forecasts.
Buildings 14 02242 g017
Table 1. Distribution of the reviewed studies according to method, study scale, data type, and time horizon for estimation.
Table 1. Distribution of the reviewed studies according to method, study scale, data type, and time horizon for estimation.
MethodScaleTemporal GranularlyData SizeTime-Horizon for EstimationReference
Statistical regression138 residential customersSub-hourly6 months2-dayZiras et al. [10]
SVRFour office buildingsHourly12 days9 am–5 pm
(9 h)
Chen et al. [14]
Averaging-based32 industrial customersHourly50 days--Ghasemi et al. [22]
Averaging-based782 residential customersHourly12 months--Wijaya et al. [23]
Cohort-based6427 residential customersHourly122 days0 am–23 pm
(24 h)
Zhang et al. [24]
Cohort-based736 residential customersHourly--3-day, 3-week, 3-monthWang et al. [25]
Statistical regression-based33 commercial buildingsHourly16 months3-day, 5-day, 10-dayCoughlin et al. [26]
Statistical regression-based29 commercial officesSub-hourly16 months4, 7, 10 monthsGranderson and Price [27]
Statistical regression-based17 office buildingsHourly12–27 months~monthWalter et al. [28]
Statistical regression-basedIndividual consumersSub-hourlyWinter months~5-dayHatton et al. [29]
Statistical regression-basedCityHourly10 days~5 daySharifi et al. [30]
SVR300 householdsHourly6 months10-hLi et al. [32]
GMRIndividual retail storeSub-hourly12 months--Srivastav et al. [33]
GMR441 customers--12 months--Zhang et al. [34]
MNN, SVR, RF, XGBoostIndividual houseHourly12 months1 h, 1-dayBampoulas et al. [35]
MLR, SVR, RF, CatBoost, LightGBM, ANN20 commercial buildingsHourly, Daily24 months--Sha et al. [36]
GCN3561 household customersSub-hourly12 months24 hTao et al. [37]
Self-organizing map, K-means clustering3 multi-story buildingsHourly27 months~5–10 daysPark et al. [38]
Averaging-based, ANN, MLR, PolyR66 household customersHourly4 months24 hJazaeri et al. [39]
Averaging-based, exponential smoothing, regression200 customersHourly, Daily12 months12-daySchwarz et al. [40]
Table 2. Occupancy characteristics of the targeted dwelling units in Atlantech in the city of La Rochelle.
Table 2. Occupancy characteristics of the targeted dwelling units in Atlantech in the city of La Rochelle.
MinMaxMeanStandard Deviation
Household composition (person/m2)Couple with children0.0040.0240.0140.004
Couple without children0.0080.0260.0140.003
Single0.0080.0260.0160.004
Occupant status (person/m2)Employed 0.0040.0740.0250.014
Unemployed0.0110.0730.0310.018
Stay-at-home0.0040.0180.0120.003
Student0.0090.0630.0290.011
Retired0.0080.0590.0210.009
Average lighting and electric equipment energy density (kW/m2)0.0191.2410.2900.203
Heating system operationComfort setpoint (°C)17.322.4019.831.26
Use of setback (%)--80-
COP of heat pumps (-)1.714.953.330.81
Table 3. Input features selected as predictors for forecasting models.
Table 3. Input features selected as predictors for forecasting models.
Input Feature (Factor)AbbreviationUnit
Previous energy demand patterns24 hH-1×24kW
48 hH-2×24kW
72 hH-3×24kW
WeatherOutdoor temperatureT_outdoor°C
Director solar radiationR_directW/m2
Diffuse solar radiationR_diffuseW/m2
Dwelling working schedule (occupancy)Hour of the dayhourhour
Day of the weekdayofweekDay
Day of the monthdayofmonthday
Table 4. Hyperparameter tuning information for forecasting models based on deep learning and classic and ensemble methods.
Table 4. Hyperparameter tuning information for forecasting models based on deep learning and classic and ensemble methods.
ModelHyperparametersSearch ScopesOptimal Values
Case ICase II
ClassicRidgeRegularization_coefficient for a norm of the weight_vector ( l 2 ){3, 6, 10, 13, 20, 50, 100, 200}10050
LassoRegularization_coefficient for a norm of the weight_vector ( l 1 ){0.001, 0.01, 0.1, 3, 5, 10, 20, 30}103
PolyRRegularization_coefficient (alphas){1 × 10−6, 1 × 10+6, 1 × 10−3, 1 × 10+3, 1 × 10−2, 1 × 10+2}{1 × 10−6, 1 × 10+6}{1 × 10−6, 1 × 10+6}
Maximum number of iterations{300, 1000, 3000}30003000
BayesienRegularization_coefficients (parameters: alphas, lambdas){1 × 10−6, 1 × 10+6, 1 × 10−3, 1 × 10+3, 1 × 10−2, 1 × 10+2, 1 × 10−1, 1 × 10+1}{1 × 10−1, 1 × 10+1}{1 × 10−1, 1 × 10+1}
Maximum number of iterations{100, 300, 1000, 3000}30001000
SGDRegRegularization_coefficient (alpha){0.001, 0.01, 1, 10}0.10.1
Maximum number of passes (epochs){300, 600, 1000, 3000, 6000}30002000
KernelRTypes of Kernel used in the algorithm {str, callable, linear}LinearLinear
Degree of Kernel{2, 3, 6}33
Regularization_coefficient (alphas){0.1, 1, 10, 30, 100}10030
SVRTypes of Kernel used in the algorithm {linear, sigmoid, rbf, poly}rbfrbf
Regularization_parameter “regressor_gamma”{0.001, 0.01, 0.1, 1, 10}{0.001, 0.01, 0.1}{0.001, 0.01, 0.1}
Regularization_parameter “regressor_C”{50, 60, 70, 100, 300, 1000}{50, 60, 70}{50, 60, 70}
EnsembleXGBoostN_estimators{100, 250, 500, 900}600500
Maximum tree depth{3, 6, 7, 8, 9, 10}73
Learning rate{0.001, 0.01, 0.02, 0.03}0.010.01
Minimum loss reduction “gamma”{0.02, 0.05, 0.1, 3}33
Subsmaple{0.3, 0.6, 0.8, 0.9, 1}0.80.6
LightGBMN_estimators{200, 300, 800, 1000}800600
Maximum tree depth{3, 4, 6, 8, 10, 15}105
Learning rate{0.05, 0.01, 0.03, 1.0}0.10.1
Number of leaves{10, 60, 100, 300}300300
Subsmaple{0.2, 0.5, 0.6, 0.75, 1}0.750.75
GBN_estimators (trees){20, 60, 200, 300}200200
Maximum tree depth{2, 4, 6, 8, 10}22
Learning rate{0.01, 0.1, 0.2, 0.3}0.30.3
RFN_estimators (trees){30, 100, 150, 200}200200
Maximum tree depth{4, 5, 6, 10, 15}1515
Maximum features{sqrt, auto}sqrtsqrt
Minimum number of samples {2, 3, 4, 5, 6}200200
AdaBoostLearning rate{0.001, 0.1, 0.2, 0.3}0.10.1
N_estimators (trees) {100, 200, 250, 300}300300
BaggingMinimum number of features{0.1, 0.3, 0.6, 1, 4}11
Maximum number of samples{1, 3, 7, 10, 15, 30}11
N_estimators (trees){200, 500, 700, 1000}10001000
CatBoostLearning rate{0.01, 0.03, 0.1, 1}11
Maximum tree depth{3, 10, 20, 30}33
Deep LearningANNNumber of hidden layers{1}11
Number of neurons per hidden layer{32, 64, 80, 100, 156}156100
Number of epochs{30, 100, 300, 600, 800}600600
Learning rate{0.001, 0.01, 0.02, 0.03}0.0010.001
Dropout rate{0, 0.1, 0.2, 0.3, 0.4}0.20.2
DNNNumber of hidden layers{2, 3, 4, 5, 6}23
Number of neurons per hidden layer{20, 30, 64, 80}6464
Number of epochs{30, 200, 300, 400, 600}300300
Learning rate{0.001, 0.01, 0.03, 0.04}0.0010.001
Dropout rate{0, 0.1, 0.2, 0.3, 0.4}0.20.2
CNNNumber of hidden layers{2, 3, 4, 5, 6}32, 3
Number of neurons per hidden layer{20, 32, 50, 64}6432, 64
Number of epochs{30, 100, 200, 300, 500}500300
Learning rate{0.001, 0.01, 0.03, 0.05}0.0010.001
Dropout rate{0, 0.1, 0.2, 0.3, 0.4}0.20.2
RNNNumber of hidden layers{2, 3, 4, 5, 6}2, 32
Number of neurons per hidden layer{16, 32, 64, 76}6432, 64
Number of epochs{30, 100, 300, 400, 500}500400
Learning rate{0.001, 0.01, 0.02, 0.03}0.0010.01
Dropout rate {0, 0.1, 0.2, 0.3, 0.4}0.20.1
LSTMNumber of hidden layers{2, 3, 4, 5, 6}2, 32
Number of neurons per hidden layer{16, 32, 64, 70}32, 6450
Number of epochs{30, 200, 300, 600}600300
Learning rate{0.001, 0.01, 0.02, 0.03}0.010.01
Dropout rate {0, 0.1, 0.2, 0.3, 0.4}0.20.1
GRUNumber of hidden layers{2, 3, 4, 5, 6}2, 32
Number of neurons per hidden layer{20, 40, 50, 64}6464, 40
Number of epochs{30, 100, 300, 400, 500}500300
Learning rate{0.001, 0.01, 0.02, 0.03}0.010.01
Dropout rate {0, 0.1, 0.2, 0.3, 0.4}0.20.1
BiLSTMNumber of hidden layers{2, 3, 4, 5}2, 32
Number of neurons per hidden layer {20, 32, 50, 64}32, 6432, 64
Number of epochs{30, 100, 200, 300, 400}300200
Learning rate{0.001, 0.01, 0.02, 0.04}0.010.01
Dropout rate {0, 0.1, 0.2, 0.3, 0.4}0.20.1
Notes: Case I and Case II refer to forecasting models with demand response baseline profiles with all dwelling units and different aggregation levels of dwelling units. Adam and Relu are also optimization techniques and activation functions for deep learning models.
Table 5. Average performance of deep learning and classic and ensemble models in forecasting aggregated demand response baselines for dwelling units over various time horizons.
Table 5. Average performance of deep learning and classic and ensemble models in forecasting aggregated demand response baselines for dwelling units over various time horizons.
Model6 h Ahead12 h Ahead18 h Ahead24 h Ahead
MAE (kW)RMSE (kW)MAPE (%)MAE (kW)RMSE (kW)MAPE (%)MAE (kW)RMSE (kW)MAPE (%)MAE (kW)RMSE (kW)MAPE (%)
ClassicMLR18.6319.9834.5821.6424.4925.6521.7625.0724.5621.8925.6122.73
Ridge18.5219.8834.0621.5724.4425.3721.5824.9124.2421.7425.4622.45
Lasso17.3918.7229.7521.0624.0923.4220.8424.3722.4721.0224.8520.92
PolyR18.6219.9734.5221.6424.4925.6221.7325.0524.5221.8825.5922.70
Bayesian18.5319.8933.6521.6424.5225.1921.6124.9424.0621.7625.4922.31
KernelR23.2424.8936.8828.0932.1829.6427.3731.9127.4828.2733.2126.59
SGDReg23.4424.7741.2925.8328.7230.3224.4827.8428.1924.5128.3525.44
ARIMA17.3122.5827.6021.3226.9723.2424.1531.0124.7227.4235.4725.97
SVR14.1916.1622.5418.1321.6318.7718.3922.0918.7119.4623.4418.18
EnsembleXGBoost8.1410.2811.5613.9517.6114.0114.3918.0714.4315.7819.8414.65
LightGBM12.2115.2718.3215.8619.8816.0416.3120.4716.5116.8821.1515.81
GB10.6113.7515.6814.9119.0414.6915.5819.8115.4716.0920.3414.86
RF15.1618.0124.0118.9022.9719.9318.6322.7119.5319.0723.4718.41
Bagging14.9117.3324.5418.1521.6819.8317.9521.6619.4418.6922.8318.45
AdaBoost20.3422.4539.7021.4224.7627.7420.5824.2125.3120.8524.8823.27
CatBoost20.8424.7633.7924.6229.8827.1724.7530.2926.3625.7331.6125.09
Deep LearningANN6.497.3010.1811.9615.1211.8114.0917.4913.9814.4518.2714.49
DNN5.897.178.8613.0717.3512.0814.9717.2412.8713.9316.7013.67
CNN6.927.359.1513.1717.1113.2714.1517.3312.9614.0917.6613.97
RNN7.858.039.9211.6315.1011.1812.6216.3612.7413.2416.6213.58
LSTM5.478.939.2311.9215.9411.7311.6315.3512.2712.8516.2912.03
GRU5.817.788.9211.5115.5511.3110.4813.9211.2213.0416.3212.52
BiLSTM5.417.079.0811.1514.8611.1410.3113.1311.1112.4115.1211.59
Table 6. Performance of different forecasting models to evaluate energy reductions due to demand response events for all dwelling units during peak heating days over 3 h ahead.
Table 6. Performance of different forecasting models to evaluate energy reductions due to demand response events for all dwelling units during peak heating days over 3 h ahead.
ModelMAE (kW)RMSE (kW)MAPE (%)
Classic MLR13.3215.1922.82
Ridge12.9814.7820.68
Lasso11.9013.6319.99
PolyR13.9815.6321.01
Bayesian14.0615.7221.09
KernelR23.7225.9634.24
SGDReg26.6829.3346.32
ARIMA22.1826.1236.94
SVR8.4410.2615.75
EnsembleCatBoost11.7312.4118.50
AdaBoost10.1612.5217.03
Bagging10.2011.7516.02
RF9.0910.6914.01
GB10.9511.9814.93
LightGBM15.6516.3523.15
XGBoost8.019.5813.12
Deep learningANN5.226.039.56
DNN5.426.2310.35
CNN7.127.5512.64
RNN4.925.4010.16
LSTM4.194.628.02
GRU4.865.279.59
BiLSTM4.975.879.89
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Gassar, A.A.A. Short-Term Energy Forecasting to Improve the Estimation of Demand Response Baselines in Residential Neighborhoods: Deep Learning vs. Machine Learning. Buildings 2024, 14, 2242. https://doi.org/10.3390/buildings14072242

AMA Style

Gassar AAA. Short-Term Energy Forecasting to Improve the Estimation of Demand Response Baselines in Residential Neighborhoods: Deep Learning vs. Machine Learning. Buildings. 2024; 14(7):2242. https://doi.org/10.3390/buildings14072242

Chicago/Turabian Style

Gassar, Abdo Abdullah Ahmed. 2024. "Short-Term Energy Forecasting to Improve the Estimation of Demand Response Baselines in Residential Neighborhoods: Deep Learning vs. Machine Learning" Buildings 14, no. 7: 2242. https://doi.org/10.3390/buildings14072242

APA Style

Gassar, A. A. A. (2024). Short-Term Energy Forecasting to Improve the Estimation of Demand Response Baselines in Residential Neighborhoods: Deep Learning vs. Machine Learning. Buildings, 14(7), 2242. https://doi.org/10.3390/buildings14072242

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop