Next Article in Journal
Spatiotemporal Differentiation and Influencing Factors of Green Technology Innovation Efficiency in the Construction Industry: A Case Study of Chengdu–Chongqing Urban Agglomeration
Next Article in Special Issue
Influences of Heat Rejection from Split A/C Conditioners on Mixed-Mode Buildings: Energy Use and Indoor Air Pollution Exposure Analysis
Previous Article in Journal
Assessment of Web Panel Zone in Built-up Box Columns Subjected to Bidirectional Cyclic Loads
Previous Article in Special Issue
Deep Forest-Based DQN for Cooling Water System Energy Saving Control in HVAC
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Hybrid Residential Short-Term Load Forecasting Method Using Attention Mechanism and Deep Learning

1
School of Electronic and Information Engineering, Suzhou University of Science and Technology, Suzhou 215009, China
2
Suzhou Smart City Research Institute, Suzhou University of Science and Technology, Suzhou 215009, China
3
Kashi Institute of Electronics and Information Industry, Kashi 844199, China
4
State Grid Suzhou Power Supply Company, Suzhou 215004, China
*
Author to whom correspondence should be addressed.
Buildings 2023, 13(1), 72; https://doi.org/10.3390/buildings13010072
Submission received: 10 November 2022 / Revised: 14 December 2022 / Accepted: 22 December 2022 / Published: 28 December 2022

Abstract

:
Development in economics and social society has led to rapid growth in electricity demand. Accurate residential electricity load forecasting is helpful for the transformation of residential energy consumption structure and can also curb global climate warming. This paper proposes a hybrid residential short-term load forecasting framework (DCNN-LSTM-AE-AM) based on deep learning, which combines dilated convolutional neural network (DCNN), long short-term memory network (LSTM), autoencoder (AE), and attention mechanism (AM) to improve the prediction results. First, we design a T-nearest neighbors (TNN) algorithm to preprocess the original data. Further, a DCNN is introduced to extract the long-term feature. Secondly, we combine the LSTM with the AE (LSTM-AE) to learn the sequence features hidden in the extracted features and decode them into output features. Finally, the AM is further introduced to extract and fuse the high-level stage features to achieve the prediction results. Experiments on two real-world datasets show that the proposed method is good at capturing the oscillation characteristics of low-load data and outperforms other methods.

1. Introduction

With the development of smart cities and smart homes, the daily electricity units of the residents are becoming increasingly numerous, resulting in a more complicated power system. As shown in Figure 1, a typical residential power supply includes biomass power generation, photovoltaic power generation, hydropower generation, and wind power generation. The residential electricity demand poses a huge potential threat to maintaining the stability of the power system. However, the existing energy supply structure is still dominated by thermal power generation, and the carbon emissions of thermal power generation will lead to climate warming, which, on the contrary, leads to an increase in energy demand. Hence, accurate residential power load forecasting is important. Generally, power load forecasting is divided into three categories: short-term load forecasting, medium-term load forecasting, and long-term load forecasting [1]. Short-term load forecasting can predict the sum of energy consumption within a few minutes or hours, which can optimize the energy dispatch and reduce the micro-grid primary energy loss [2]. Further, residents can reduce electricity costs by formulating electricity utilization strategies with the current pricing scheme. In addition, short-term load forecasting provides a judgment basis for some abnormal power information and guarantees the safety of people’s lives and property. Therefore, in this paper, we mainly consider how to design a suitable short-term load forecasting method to improve forecasting accuracy.
Short-term load forecasting is a time series forecasting problem. The factors that influence residential short-term load forecasting are complex and significant, including social events, electricity price adjustments, human behavior, and other uncertainties. Initially, researchers relied on artificial feature analysis to obtain empirical models, which is time-consuming and inaccurate [3]. Due to the widely different distribution of residential electricity consumption, load forecasting tends to utilize machine learning and deep learning to optimize the results. Support vector machine (SVM) [4], autoregressive integrated moving average (ARIMA) [5], extreme gradient boosting regressor (XGBoost) [6], artificial neural network (ANN) [7], and long short-term memory network (LSTM) [8] are the most commonly used methods [9]. Machine learning methods exhibit some benefits: (1) they have high computational efficiency; (2) they are highly interpretable in their basic form [10]. Although machine learning methods perform well in load forecasting, deep learning for load forecasting can obtain better results. First, deep learning does not require the creation of feature engineering in machine learning. Second, deep learning has a good generalization [11].
Currently, many studies have focused on the accuracy of short-term load forecasting. However, they have ignored trend tracking in oscillating data, especially in valley data. The oscillating data are likely to correspond to the operation of some power consumption units. If this situation cannot be predicted accurately, then for future fault monitoring and other studies, the system’s stability will be challenging to achieve because of the increased likelihood of misjudgment. Therefore, this paper proposes a hybrid model, i.e., DCNN-LSTM-AE-AM, for residential short-term load forecasting. Figure 2 shows the architecture of the proposed DCNN-LSTM-AE-AM. By broadening the temporal horizon, we utilize the dilated convolutional neural network to extract temporal features from the time series. Then, the LSTM-AE is applied to mine the electricity consumption characteristics thoroughly. Finally, an attention mechanism (AM) is used to reflect the importance of behaviors in load prediction. The main contributions of this paper are listed as follows:
Considering that individual data loss may still occur due to various conditions, we propose the T-nearest neighbors (TNN) algorithm to solve the problem of missing values, which can estimate the missing load according to the load data of adjacent similar days.
We propose a hybrid short-term residential electricity load forecasting model (DCNN-LSTM-AE-AM). This proposed model focuses on the trend tracking of oscillating data, which can be captured with almost no delay, and provides a technical basis for predicting power failures in advance. Compared with other methods, DCNN-LSTM-AE-AM can capture the valley load data, which improves the prediction accuracy.
The proposed DCNN-LSTM-AE-AM model is validated on two real-world datasets and compared with the existing methods. Experimental results show that this model improves the prediction results and has a good generalization.
The rest of this paper is organized as follows. Section 2 reviews the related work. Section 3 proposes the main results on how to design a hybrid short-term residential electricity load forecasting model. In Section 4, the experimental results are demonstrated, and some comparisons with the other three models are made. Finally, Section 5 concludes the paper and future work.

2. Related Work

In this section, we discuss the previous research on short-term load forecasting. The existing approaches for feature extraction can be divided into three: manual feature screening, traditional machine learning, and deep earning.
In the early days, load data were less affected by uncertain factors such as human behavior and climate change. Researchers first processed the data by relying on expert experience. Then, with the application of some statistical methods, they tried to build a prediction model. For example, Taylor analyzed the seasonal cycle within the day and week as the key features and processed them with five statistical methods to obtain an optimal forecasting result [12]. Statistical methods are rich in theory and highly interpretable for some specific characteristics. However, due to the growth in the living standard, the approaches are difficult for experts to interpret the complex data.
With the advance of big data and artificial intelligence, data-driven load forecasting technologies have received attention extensively [13,14,15,16,17]. The computing capacity of hardware devices has greatly influenced short-term load forecasting. He et al. [18] utilized ARIMA to design a high-frequency short-term load forecasting model. The model divided the inputs by season and then used hourly load data to predict energy consumption for the next month. Cao et al. [19] divided the dataset by the characteristics of similar daily meteorological conditions and applied ARIMA for prediction. Although the above models predicted well in the same season, such models based on seasons are extremely dependent on manual screening and do not have the ability to generalize. Cai et al. [20] proposed an energy prediction model that combined the K-means and data mining methods to analyze the energy consumption of 16,000 residential buildings. Mohammadi et al. [21] proposed a hybrid model based on sliding window empirical mode decomposition (SWEMD) to predict the power consumption of small buildings. They also proposed an algorithm to optimize the model parameters. Chauhan et al. [22] designed a hybrid model based on SVM and ensemble learning for prediction, where load data were processed separately through hourly and daily resolutions. Experiments in Aguilar Madrid and Antonio [6] showed that the XGBoost could obtain the best performance from the machine learning algorithms set. Massaoudi et al. [23] proposed an ensemble-based LGBM-XGB-MLP hybrid model to improve prediction performance. Although traditional machine learning models can achieve good results in load forecasting, they all require more effort in the feature selection and parameter optimization process [11].
Deep learning has received explosive growth in various fields. Some typical networks for deep learning, such as ANN [24,25,26], convolutional neural network (CNN) [27], recurrent neural network (RNN) [28], and LSTM [29], have been widely used in load forecasting. Chen et al. [30] added periodic features and utilized a deep residual network (ResNet) to predict the hourly residential load, whose evaluation performance had been improved. From the aspect of demand-side management (DSM), Kong et al. [31] proposed an improved deep belief network (DBN) method to forecast 1-h-level loads. Compared with others, this method could significantly improve both the day-ahead and week-ahead load forecasting results. Dong et al. [32] proposed a distributed deep belief network (DDBN) with Markov switching topology, which improved distributed communication stability and prediction accuracy.
It should be noted that energy consumption is generated by a variety of complex factors, so how to utilize the nonlinear models to improve prediction accuracy is the main consideration. Recently, CNN has injected vitality into short-term load forecasting [33,34,35]. Amarasinghe et al. [36] tried to apply CNN for prediction on the residential load dataset [37]. Experimental results show that this method is effective, but the prediction accuracy needs to be improved. Sadaei et al. [38] proposed a hybrid algorithm combining CNN and fuzzy time series (FTS) for forecasting. This method converted multivariate time series into multi-channel images. The proposed model overcame some advanced time series models for short-term load forecasting with better results and solved the problem of over-fitting.
RNN, LSTM, and gated recurrent neural network (GRU) are designed for time-series data, which is sensitive to temporal features. They can fully mine time-related features between adjacent data. Rahman et al. [39] utilized RNN to forecast the 1-h commercial and residential load data in the medium and long-term forecasting and thus realized the load trend tracking. However, it is challenging for RNNs to achieve convergence on data within a long time interval. Kong et al. [8] relied on the multi-layer LSTM to predict individual energy consumption. Compared with others, Kong et al. [8] not only solved the problem of RNN but also improved the prediction accuracy. Li et al. [40] proposed an improved GRU to dynamically capture temporal correlations within the forecast period, enabling the model to adapt to different datasets.
In the field of load forecasting, it is more inclined to utilize the feature extraction capabilities of the hybrid model to improve forecasting accuracy. Shi et al. [41] proposed a pooling method based on the LSTM network, which further improved the prediction accuracy of LSTM through backpropagation and pooling. Jiang et al. [42] proposed a hybrid model based on CNN and LSTM to predict household energy consumption. Unlike other combined methods, this method divided the input into two parts: long and short data. Then, they obtained a result with sufficient feature interaction through data fusion. Yue et al. [43] combined ensemble empirical mode decomposition (EEMD), permutation entropy (PE), feature selection (FS), LSTM, and bayesian optimization algorithm (BOA) to optimize the prediction accuracy. Further, some reasonable explanations were made for the reconstructed subsequences. Lin et al. [44] proposed an AM-based auto-encoder structure for LSTM, which provided superior prediction accuracy and had a good generalization. Wei et al. [45] proposed detrend singular spectrum fluctuation analysis (DSSFA) to extract trend and periodic components and then input these components into LSTM to improve short-term forecasting accuracy. Laouafi et al. [46] proposed an adaptive hybrid ensemble method named CMKP-EG-SVR and optimized the result of the mixture model through a gaussian-based error correction strategy.

3. Model Architecture of Residential Short-Term Load Forecasting

The goal of residential short-term load forecasting is to improve residents’ electricity experience. It is an essential part of energy supply management. In this section, we mainly introduce a complete forecasting architecture that can improve residential load forecasting accuracy.
Figure 3 shows the process of residential short-term load forecasting, DCNN-LSTM-AE-AM. In Figure 3, the missing data are first processed based on the TNN algorithm, and the DCNN layer extracts the initial features. Then, the LSTM-AE is used to extract the spatiotemporal feature information. Finally, the AM is introduced to analyze the importance of the extracted features and outputs the final prediction result.

3.1. Data Processing

Residential energy consumption data are collected by smart meters. Smart meters can accurately collect and transmit data to the data management center through various communication networks. Considering that the communication network is susceptible to interference by multiple factors, data loss is inevitable. Therefore, the missing data needs to be processed by some methods. This paper proposes the T-nearest neighbors (TNN) algorithm to fill in the missing data. At time t, TNN is defined as follows:
I t 1 K I t K 2 T + I t ( K 2 1 ) T + + I t T + I t + T + + I t + ( K 2 1 ) T + I t + K 2 T ,
where K represents the number of selected adjacent values; T represents the interval period; I t is the output of TNN at time t. The algorithm can solve the problem that the duration of missing data is relatively long. When the duration of the missing data is too long, the missing data will be ignored.
Moreover, the collected load data usually has a small amount of singular data, which will affect the overall model. Therefore, it is necessary to scale these data to some fixed range so that they will conform to a certain distribution. In this paper, we use a linear normalization, i.e., the max-min normalization, to process the load data, which is defined as follows:
θ n o r m = θ θ min θ max θ min ,
where θ n o r m represents the normalized output; θ represents the current input; θ max and θ min represent the upper and lower bounds of the current sequence input, respectively.

3.2. Dilated Convolutional Neural Network

CNN has been widely used in image processing ever since it was proposed [36]. Recently, time series data has also tried to use CNN to deal with short-term load forecasting. The core of CNN is weight sharing. Each CNN has a convolution kernel, which shares different weights according to the convolution operation. However, CNN will lose this feature information with long-term regularity, such as valley oscillation load [47]. To broaden the horizon of CNN, in this paper, we transform the convolution computation of continuous data into the convolution computation of skipping data, which is called the dilated convolutional neural network (DCNN) [48,49]. For a τ -dimension input vector ν R τ and a kernel w : { 0 , , k 1 } R , the dilated convolution operation on element s of input vector ν is defined as:
y s = i = 0 k 1 ν s + r · i · w i ,
where the dilated convolution adjustment rate r is expressed as the interval step size for selecting input data; k represents the kernel size. In addition, τ represents the time step.
Figure 4a,b illustrate the internal structures of CNN and DCNN, respectively. In Figure 4, CNN has one dimension, a kernel of one, and a dilation rate of one, while DCNN has one dimension, a kernel of one, and a dilation rate of one. The first layer is the original input layer, the second layer is the hidden layer, and the third layer is the output layer. In addition, after adding the convolution layer, the activation function and pooling layer are added appropriately to help the backpropagation of the gradient.

3.3. LSTM-Based Autoencoder

3.3.1. Long Short-Term Memory Network

RNN is a forward-propagating sequential neural network. When it deals with long-duration data, it usually faces some challenges, such as gradient disappearance and gradient explosion. Hochreiter and Schmidhuber [29] proposed an improved network based on the RNN structure and named it LSTM. The internal structure of the LSTM is shown in Figure 5, where the memory cell can retain information from a long time ago, and the forget gate can choose to discard some feature information. Backpropagation in the LSTM strengthens the interaction ability of context information and reserves more useful spatiotemporal feature information.
The update principles of the LSTM are defined as:
f t = σ W f t x t + W f h h t 1 + b f ,
u t = σ W t x x t + W u h h t 1 + b u ,
g t = tanh W g x x t + W g h h t 1 + b g ,
o t = σ W o x x t + W o h h t 1 + b o ,
c t = g t u t + c t 1 f t ,
h t = tanh c t o t ,
where f t , u t , g t , o t in Equations (4)–(7) represent the information at the forget gate, input gate, input node, and output gate at time t; σ , tanh are the multiplication calculations and activation functions; b f , b u , b g , b o are the bias parameters of the corresponding processing units; c t , h t in Equations (8) and (9) represent memory cells; W f t , W f h , W t x , W u h , W g x , W g h , W o x , and W o h are the weight matrices of the corresponding processing units; ⊙ represents element multiplication; FC represents the fully connected layer. These units use functions σ and t a n h to continuously compress the input x t to a smaller range.

3.3.2. Bidirectional Long Short-Term Memory Network

The bidirectional long short-term memory network (BiLSTM) is a variant of LSTM. As shown in Figure 6, BiLSTM is a special network structure formed by superimposing two LSTM layers. The LSTMs synchronously train the input data at the same time step. These two LSTM layers differ in that one input the data in a positive temporal order, and the other processes it in a reverse temporal order. This structure not only utilizes the information of the previous moment but also relies on the information of the latter moment [50,51].
Let h t and h t be the hidden states of forward and backward propagation, respectively. Then, h t , h t , and the output H t of BiLSTM are calculated as follows:
h t = L S T M ξ t , S t 1 , t 1 , T ,
h t = L S T M ξ t , S t + 1 , t T , 1 ,
H t = F C h t , h t .
where ξ t represents the current input at current time t; S t represents the internal state in the LSTM, that is, the memory cells and the hidden state; T represents the time steps.

3.3.3. Autoencoder

Autoencoder (AE), as an artificial neural network (ANN), is a conceptual network structure with an encoder and decoder. The AE aims to find an optimal set of connection weights by minimizing the reconstruction error between the original input and the output [52]. For any AE, there is a n-dimensional input vector φ t and an output vector ϵ t with random dimensions. The input φ t can be mapped to output ϵ t according to the following mapping functions:
Θ = f W ˜ · φ t + α ,
ϵ t = ρ W ^ · Θ + β ,
where Θ is the mapping output of the encoding layer; W ˜ and W ^ are two weight matrices; f and ρ are two activation functions; α and β are the bias parameters of the encoder and decoder, respectively.
Obviously, simple nonlinear AE is difficult for time series data feature extraction. As shown in Figure 3, in this paper, the LSTM and BiLSTM are combined to construct the AE. In order to learn more feature information and increase the sensitivity of the contextual information connection, the encoder part adopts two BiLSTM layers. In addition, the increase in the depth of the neural network structure is conducive to the extraction and fusion of load features. In the decoder part, we only need to add two LSTM layers as the feature analysis layer to reduce the unnecessary network computing burden.

3.3.4. Attention Mechanism

The AM pays more attention to the important parts of reconstructing the cognitive world instead of making an average judgment on the whole. Figure 7 shows the structure of the AM. In Figure 7, FC is the fully connected layer, which fuses with the output of the AM. The output of the AM can be calculated as follows:
μ t = Ω t η t .
where η t = { h 1 , , h χ } is the output decoded by LSTM-AE and is a χ -dimensional hidden state vector at time t; Ω t = λ 1 , , λ χ is a weight matrix.
The matrix Ω t in (15) can be implemented according to the following procedure. First, the output η t represents the input of the AM. Then, the alignment model a ( · ) aligns the input with the output vector ϕ t = ε 1 , , ε χ . The alignment score ϕ t is calculated as follows:
ϕ t = a δ t 1 , η t .
In this study, the alignment model a δ t 1 , η t represents tanh δ t 1 η t + γ , where the cell state δ t 1 decoded by LSTM-AE represents the χ -dimensional hidden state vectors at time t 1 ; γ is a vector of bias parameters. Finally, each element λ j is computed by applying a softmax operation:
λ j = exp ε j i = 1 χ exp ε i ,
where i and j represent the i-th and j-th elements in ϕ t . After injecting the AM, we also apply multiple fully connected layers as the output layer to complete the final load forecasting.

4. Experiment Results and Analysis

In this section, we will show the effectiveness of our proposed DCNN-LSTM-AE-AM load forecasting method through some experiments.

4.1. Experiment Settings

4.1.1. Dataset Selection

The experiments use two real-world power load datasets to test our proposed method’s robustness and generalization. The household electricity consumption dataset from the UCI machine learning library (IHEPC) [37] records the energy consumption information of a house from 2006 to 2010. It contains multiple attributes: date, time, global active power, global reactive power, voltage, current, and active power of three-room types: kitchen, bathroom, and bedroom. There are 2,075,269 1-min-level data, including 25,979 missing values. Table 1 shows detailed information on IHEPC. The global active power represents the actual power consumption, so this paper only takes the global active power as the input. Another dataset is from the smart grid and smart city (SGSC) [53] projects carried out by the Australian government and industry consortium Ausgrid. Data in the SGSC dataset are collected from 10,000 households and some retail stores in New South Wales (NSW) from 2010 to 2014. In this paper, the IHPEC is organized as the sum of energy consumption within 15 min and 1 h. Since the SGSC only provides electricity consumption per half hour, the SGSC is organized as the sum of energy consumption within 30 min and 1 h. Figure 8 shows the randomly selected data of IHPEC at a 15-min resolution and the randomly selected data of one household from SGSC at a 30-min resolution.

4.1.2. Experiment Setup

All networks are built based on Python3.6, Keras2.2, and TensorFlow2.0. The device is configured with a 2.6 GHz intel i9 CPU and a 16GB NVIDIA TESLA T4 GPU.
According to the findings in [8], some rules of thumb for hyperparameter selection are adopted. Since hyperparameter selection is a time-consuming task, in this paper, we try to use different combinations of parameters to obtain the optimal performance in MSE. Table 2 lists the hyperparameter settings of the proposed DCNN-LSTM-AE-AM. The first 1D-Conv layer uses the DCNN with a kernel of 3, 12 filters, and a dilation rate of 2 to extract features and ReLU as output results. The number of the second 1D-Conv layer’s filters is upgraded to 24. These two SpatialDropout layers randomly zero the parameters with probabilities 0.1 and 0.2, respectively. In the LSTM-AE, 32 units are used in all four temporal models. In the final output layer, the AM uses 32 units, and the three Dense layers use 96, 32, and 1 unit, respectively. We set the maximum training epoch to 50. All other methods are also tested on the same equipment, hyperparameter configuration, and environment to allow horizontal comparisons. In addition, XGBoost obtains optimal performance with the number of estimators set to 30 after constantly changing the number of estimators.
In this paper, we split the dataset into a training set and a test set, whose ratios are designed as 0.67 and 0.33 . We fill the missing values in the training set according to the TNN algorithm. The missing values in the test set are ignored to prevent prior knowledge leakage.

4.1.3. Evaluation Metric

It is well known that the classification task can be evaluated by accuracy in percentage. However, this kind of accuracy is not appropriate for evaluating any regression task. In this paper, in order to objectively evaluate the fairness and integrity of the methods, we use four evaluation metrics: MAE, RMSE, MSE, and MAPE. The MAE is a method of averaging quantization errors. Compared with MAE, the other methods have made some improvements. MSE pays more attention to the influence of outliers on the overall prediction effect. RMSE performs arithmetic square root on the overall basis of the MSE, which amplifies the difference of the MSE. MAPE can focus on the gap between the error and the actual value. The formulations for the MAE, RMSE, MSE, and MAPE are listed as follows:
M A E = 1 N i = 1 N ψ p r e d ψ ,
M S E = 1 N i = 1 N ( ψ p r e d ψ ) 2 ,
R M S E = 1 N i = 1 N ( ψ p r e d ψ ) 2 ,
M A P E = 100 % N i = 1 N ψ p r e d ψ ψ .
where N represents the total amount of the load data; ψ and ψ p r e d represent the actual load and the predicted load, respectively. For all those evaluation metrics, we have that the smaller the value, the more accurate the model.

4.2. Influence of Hyperparameters

This subsection sets various optimized hyperparameters to achieve the best performance. We evaluate the performance of the proposed model under these hyperparameters by the MSE. To ensure the fairness of the experiments, we compare the influence of batch size, optimizer, and learning rate on the MSE with other parameters that remain unchanged. Figure 9 shows the MSE of the proposed method under different hyperparameter settings. From Figure 9, we can have that our method has a certain degree of sensitivity to hyperparameters. According to the experimental results in Figure 8, we choose the batch size of 64, the optimizer as Adam, and the learning rate as 0.001.

4.3. Influence of Time Step

Time step τ , i.e., the length of the input data, is one of the factors affecting the robustness and accuracy of the proposed short-term load forecasting model. In deep networks, long time-series data often results in overfitting. Hence, the data length also affects the performance of our proposed method. The MSE, RMSE, MAE, and MAPE values under different data lengths are shown in Table 3. The evaluation metrics with lengths between 8 and 14 have very little difference. Therefore, in these following experiments, we let the data length be 12.

4.4. Performance Evaluation on IHEPC Dataset

The performance of the proposed DCNN-LSTM-AE-AM is compared with some sole models and some hybrid models. The sole models include XGBoost, DCNN, LSTM [8], and AM; while the hybrid models include LSTM-AE [54], CNN-LSTM [42], DCNN-AM, DCNN-LSTM-AE, and LSTM-AE-AM. The evaluation metrics comparison among different methods at 15-min and 1-h resolutions are shown in Table 4. From Table 4, we have that: (1) our proposed method obtains the best metrics of 0.0041 in MSE, 0.0640 in RMSE, 0.0333 in MAE, and 0.6757 in MAPE at the 15-min resolution; (2) while for the dataset at the 1-h resolution, our proposed method outperforms others with metrics of 0.0086 in MSE, 0.0926 in RMSE, 0.0667 in MAE, and 0.7257 in MAPE at the 1-h resolution. (3) Compared with the existing methods, the overall prediction accuracy of the DCNN-LSTM-AE-AM at different resolutions has obtained a little improvement. (4) Performances of those sole models are worse than those of hybrid models. It is worth noting that there is not much difference between the performance of DCNN-AM, LSTM-AE-AM, and DCNN-LSTM-AE-AM.
To show the effect of individual data in the model, we also use four box-figures to show the performance of different models in MSE, RMSE, MAE, and MAPE. As shown in Figure 10, four subfigures are used to show the performance of the IHEPC dataset at a 15-min resolution. Since we have given the average results in Table 4, the average results are no longer marked in the subfigures, and only the median results are marked. From Figure 10, one can have that: (1) the proposed method outperforms others in all the evaluation metrics; (2) the median of the metrics is lower than that of other methods; (3) the MSE, RMSE, and MAE obtain significant improvement, while the increase in MAPE is relatively small.
Figure 10 also shows that all the mentioned forecasting methods are still far from accurate prediction, which is the direction we need to focus on. Further, CNN-LSTM, LSTM-AE, and our proposed method all inherit the feature extraction capability of LSTM.
To demonstrate our proposed model’s trend-tracking capability, we show the prediction results of the mentioned four sole models and the six hybrid models in Figure 11 and Figure 12. From Figure 11, it can be seen that the tracking trend of deep learning methods, except for the AM, can outperform traditional machine learning. DCNN is always trying to match the low-load data. The LSTM has outstanding temporal feature extraction and good trend tracking, but the error with respect to the actual load is large. Although the overall performance of using only AM is not good, the AM is suitable for forecasting low-load data. The performance of AM benefits from local feature fusion capability.
From Figure 12, one can have that: (1) Different hybrid models have different prediction results. (2) The LSTM-AE is similar to the LSTM, which further narrows the numerical gap. (3) Although the CNN-LSTM is sensitive to data with large fluctuations in value and predicts the result with a small error, there is a large error when capturing the valley values. This is one of the reasons for the large value of MAPE. (4) DCNN-AM and LSTM-AE-AM are increasingly perceptive to the low-load data. The addition of AM in the last layer is definitely the main contributor to improving the prediction performance of the hybrid model on the low-load data. (5) Compared with the other five mentioned hybrid models, DCNN-LSTM-AE-AM can quickly capture the data in this situation and predict the valley data very well. This benefits from the broadened time horizon of the DCNN. The introduction of AM makes the weights closer to the actual load, which significantly reduces the prediction error. In addition, the results of DCNN-LSTM-AE suggest that no combination of hybrid models can improve the prediction results.

4.5. Performance Evaluation on SGSC Dataset

Our proposed method is also tested on the SGSC dataset, where it shows excellent accuracy and has a good generalization. Table 5 shows the evaluation metrics comparisons of different forecasting methods on the SGSC dataset. The performance of the proposed method on the SGSC dataset is similar to that on the IHEPC dataset, and both have obtained the best results for each evaluation metric. The MSE of the proposed method obviously gets improved compared with other methods. The other evaluation metrics also have a small increase.
As shown in Figure 13, we compare the actual load with the prediction results. In this study, a residential user is arbitrarily selected, and the proposed DCNN-LSTM-AE-AM can easily predict the load at the next moment. The proposed method can capture the overall trend of the actual load with small time offsets and numerical errors. In addition, the method also has good predictability for the valley data. This confirms that DCNN and AM have a good ability to capture long-term regular data.

5. Conclusions

This paper proposes a short-term load forecasting model to predict residential energy consumption. A hybrid electric load forecasting model, i.e., DCNN-LSTM-AE-AM, is constructed with the introduction of existing deep learning methods. We use multiple similar-days data in the vicinity of missing values as the basis for inferring the original data and use the TNN algorithm to fill in the missing data. In the initial feature-extraction stage, DCNN broadens the time horizon to retain the load features. LSTM-AE is used to improve the analysis capability of features. In the feature fusion stage, the importance of features in each period is summarized based on AM, which enhances the final prediction accuracy. The validity of the proposed method is verified on two real-world datasets. Experimental results show that the proposed method improves the accuracy of residential load forecasting and can capture low-load data features.
In future work, we need to improve the accuracy of residential load forecasting by exploiting residential lifestyle features. Moreover, how to achieve real-time forecasting through methods such as online learning is another work.

Author Contributions

Conceptualization, X.J.; methodology, K.Y.; software, H.H.; validation, Y.Z. and H.H.; formal analysis, X.J. and R.B.; investigation, X.J.; resources, R.B.; data curation, D.C.; writing—original draft preparation, X.J.; writing—review and editing, Z.C.; visualization, Y.Z.; supervision, Z.C. and R.B.; project administration, D.C.; funding acquisition, Z.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (NSFC) under grant nos. 51874205; Jiangsu Qinglan Project.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The figures and tables used to support the findings of this study are included in the article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Jacob, M.; Neves, C.; Vukadinović Greetham, D. Forecasting and Assessing Risk of Individual Electricity Peaks; Springer Nature: Berlin/Heidelberg, Germany, 2020. [Google Scholar]
  2. Hernandez, L.; Baladron, C.; Aguiar, J.M.; Carro, B.; Sanchez-Esguevillas, A.J.; Lloret, J.; Massana, J. A survey on electric power demand forecasting: Future trends in smart grids, microgrids and smart buildings. IEEE Commun. Surv. Tutor. 2014, 16, 1460–1495. [Google Scholar] [CrossRef]
  3. Ghofrani, M.; Hassanzadeh, M.; Etezadi-Amoli, M.; Fadali, M.S. Smart meter based short-term load forecasting for residential customers. In Proceedings of the 2011 North American Power Symposium, Boston, MA, USA, 4–6 August 2011; pp. 1–5. [Google Scholar]
  4. Ullah, I.; Ahmad, R.; Kim, D. A prediction mechanism of energy consumption in residential buildings using hidden markov model. Energies 2018, 11, 358. [Google Scholar] [CrossRef] [Green Version]
  5. Lee, Y.S.; Tong, L.I. Forecasting time series using a methodology based on autoregressive integrated moving average and genetic programming. Knowl.-Based Syst. 2011, 24, 66–72. [Google Scholar] [CrossRef]
  6. Aguilar Madrid, E.; Antonio, N. Short-term electricity load forecasting with machine learning. Information 2021, 12, 50. [Google Scholar] [CrossRef]
  7. Fumo, N.; Biswas, M.R. Regression analysis for prediction of residential energy consumption. Renew. Sustain. Energy Rev. 2015, 47, 332–343. [Google Scholar] [CrossRef]
  8. Kong, W.; Dong, Z.Y.; Jia, Y.; Hill, D.J.; Xu, Y.; Zhang, Y. Short-term residential load forecasting based on LSTM recurrent neural network. IEEE Trans. Smart Grid 2017, 10, 841–851. [Google Scholar] [CrossRef]
  9. Ahmad, A.S.; Hassan, M.Y.; Abdullah, M.P.; Rahman, H.A.; Hussin, F.; Abdullah, H.; Saidur, R. A review on applications of ANN and SVM for building electrical energy consumption forecasting. Renew. Sustain. Energy Rev. 2014, 33, 102–109. [Google Scholar] [CrossRef]
  10. Arik, S.Ö.; Pfister, T. Tabnet: Attentive interpretable tabular learning. In Proceedings of the AAAI Conference on Artificial Intelligence, Online, 2–9 February 2021; Volume 35, pp. 6679–6687. [Google Scholar]
  11. Janiesch, C.; Zschech, P.; Heinrich, K. Machine learning and deep learning. Electron. Mark. 2021, 31, 685–695. [Google Scholar] [CrossRef]
  12. Taylor, J.W. Short-term load forecasting with exponentially weighted methods. IEEE Trans. Power Syst. 2011, 27, 458–464. [Google Scholar] [CrossRef]
  13. Hammad, M.A.; Jereb, B.; Rosi, B.; Dragan, D. Methods and models for electric load forecasting: A comprehensive review. Logist. Supply Chain Sustain. Glob. Chall. 2020, 11, 51–76. [Google Scholar] [CrossRef]
  14. Liu, X.; Zhang, Z.; Song, Z. A comparative study of the data-driven day-ahead hourly provincial load forecasting methods: From classical data mining to deep learning. Renew. Sustain. Energy Rev. 2020, 119, 109632. [Google Scholar] [CrossRef]
  15. Panda, S.K.; Ray, P.; Salkuti, S.R. A Review on Short-Term Load Forecasting Using Different Techniques. In Recent Advances in Power Systems; Springer: Singapore, 2022; pp. 433–454. [Google Scholar]
  16. Vanting, N.B.; Ma, Z.; Jørgensen, B.N. A scoping review of deep neural networks for electric load forecasting. Energy Inform. 2021, 4, 49. [Google Scholar] [CrossRef]
  17. Haben, S.; Arora, S.; Giasemidis, G.; Voss, M.; Greetham, D.V. Review of low voltage load forecasting: Methods, applications, and recommendations. Appl. Energy 2021, 304, 117798. [Google Scholar] [CrossRef]
  18. He, H.; Liu, T.; Chen, R.; Xiao, Y.; Yang, J. High frequency short-term demand forecasting model for distribution power grid based on ARIMA. In Proceedings of the 2012 IEEE International Conference on Computer Science and Automation Engineering (CSAE), Zhangjiajie, China, 25–27 May 2012; Volume 3, pp. 293–297. [Google Scholar]
  19. Cao, X.; Dong, S.; Wu, Z.; Jing, Y. A data-driven hybrid optimization model for short-term residential load forecasting. In Proceedings of the 2015 IEEE International Conference on Computer and Information Technology; Ubiquitous Computing and Communications; Dependable, Autonomic and Secure Computing; Pervasive Intelligence and Computing, Liverpool, UK, 26–28 October 2015; pp. 283–287. [Google Scholar]
  20. Cai, H.; Shen, S.; Lin, Q.; Li, X.; Xiao, H. Predicting the energy consumption of residential buildings for regional electricity supply-side and demand-side management. IEEE Access 2019, 7, 30386–30397. [Google Scholar] [CrossRef]
  21. Mohammadi, M.; Talebpour, F.; Safaee, E.; Ghadimi, N.; Abedinia, O. Small-scale building load forecast based on hybrid forecast engine. Neural Process. Lett. 2018, 48, 329–351. [Google Scholar] [CrossRef]
  22. Chauhan, M.; Gupta, S.; Sandhu, M. Short-Term Electric Load Forecasting Using Support Vector Machines. ECS Trans. 2022, 107, 9731. [Google Scholar] [CrossRef]
  23. Massaoudi, M.; Refaat, S.S.; Chihi, I.; Trabelsi, M.; Oueslati, F.S.; Abu-Rub, H. A novel stacked generalization ensemble-based hybrid LGBM-XGB-MLP model for short-term load forecasting. Energy 2021, 214, 118874. [Google Scholar] [CrossRef]
  24. Lee, K.; Cha, Y.; Park, J. Short-term load forecasting using an artificial neural network. IEEE Trans. Power Syst. 1992, 7, 124–132. [Google Scholar] [CrossRef] [Green Version]
  25. Bisht, B.S.; Holmukhe, R.M. Electricity load forecasting by artificial neural network model using weather data. IJEET Trans. Power Syst. 2013, 4, 91–99. [Google Scholar]
  26. Kuo, P.H.; Huang, C.J. A high precision artificial neural networks model for short-term energy load forecasting. Energies 2018, 11, 213. [Google Scholar] [CrossRef] [Green Version]
  27. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 60, 25–31. [Google Scholar] [CrossRef] [Green Version]
  28. Sherstinsky, A. Fundamentals of recurrent neural network (RNN) and long short-term memory (LSTM) network. Phys. D Nonlinear Phenom. 2020, 404, 132306. [Google Scholar] [CrossRef] [Green Version]
  29. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  30. Chen, K.; Chen, K.; Wang, Q.; He, Z.; Hu, J.; He, J. Short-term load forecasting with deep residual networks. IEEE Trans. Smart Grid 2018, 10, 3943–3952. [Google Scholar] [CrossRef] [Green Version]
  31. Kong, X.; Li, C.; Zheng, F.; Wang, C. Improved deep belief network for short-term load forecasting considering demand-side management. IEEE Trans. Power Syst. 2019, 35, 1531–1538. [Google Scholar] [CrossRef]
  32. Dong, Y.; Dong, Z.; Zhao, T.; Li, Z.; Ding, Z. Short term load forecasting with markovian switching distributed deep belief networks. Int. J. Electr. Power Energy Syst. 2021, 130, 106942. [Google Scholar] [CrossRef]
  33. Deng, Z.; Wang, B.; Xu, Y.; Xu, T.; Liu, C.; Zhu, Z. Multi-scale convolutional neural network with time-cognition for multi-step short-term load forecasting. IEEE Access 2019, 7, 88058–88071. [Google Scholar] [CrossRef]
  34. Aouad, M.; Hajj, H.; Shaban, K.; Jabr, R.A.; El-Hajj, W. A CNN-Sequence-to-Sequence network with attention for residential short-term load forecasting. Electr. Power Syst. Res. 2022, 211, 108152. [Google Scholar] [CrossRef]
  35. Zhang, G.; Bai, X.; Wang, Y. Short-time multi-energy load forecasting method based on CNN-Seq2Seq model with attention mechanism. Mach. Learn. Appl. 2021, 5, 100064. [Google Scholar] [CrossRef]
  36. Amarasinghe, K.; Marino, D.L.; Manic, M. Deep neural networks for energy load forecasting. In Proceedings of the 2017 IEEE 26th International Symposium on Industrial Electronics (ISIE), Edinburgh, UK, 19–21 June 2017; pp. 1483–1488. [Google Scholar]
  37. UCI. Individual Household Electric Power Consumption Data Set. Available online: https://archive.ics.uci.edu/ml/datasets/Individual+household+electric+power+consumption (accessed on 4 March 2020).
  38. Sadaei, H.J.; e Silva, P.C.d.L.; Guimarães, F.G.; Lee, M.H. Short-term load forecasting by using a combined method of convolutional neural networks and fuzzy time series. Energy 2019, 175, 365–377. [Google Scholar] [CrossRef]
  39. Rahman, A.; Srikumar, V.; Smith, A.D. Predicting electricity consumption for commercial and residential buildings using deep recurrent neural networks. Appl. Energy 2018, 212, 372–385. [Google Scholar] [CrossRef]
  40. Li, D.; Sun, G.; Miao, S.; Gu, Y.; Zhang, Y.; He, S. A short-term electric load forecast method based on improved sequence-to-sequence GRU with adaptive temporal dependence. Int. J. Electr. Power Energy Syst. 2022, 137, 107627. [Google Scholar] [CrossRef]
  41. Shi, H.; Xu, M.; Li, R. Deep learning for household load forecasting—A novel pooling deep RNN. IEEE Trans. Smart Grid 2017, 9, 5271–5280. [Google Scholar] [CrossRef]
  42. Jiang, L.; Wang, X.; Li, W.; Wang, L.; Yin, X.; Jia, L. Hybrid multitask multi-information fusion deep learning for household short-term load forecasting. IEEE Trans. Smart Grid 2021, 12, 5362–5372. [Google Scholar] [CrossRef]
  43. Yue, W.; Liu, Q.; Ruan, Y.; Qian, F.; Meng, H. A prediction approach with mode decomposition-recombination technique for short-term load forecasting. Sustain. Cities Soc. 2022, 85, 104034. [Google Scholar] [CrossRef]
  44. Lin, J.; Ma, J.; Zhu, J.; Cui, Y. Short-term load forecasting based on LSTM networks considering attention mechanism. Int. J. Electr. Power Energy Syst. 2022, 137, 107818. [Google Scholar] [CrossRef]
  45. Wei, N.; Yin, L.; Li, C.; Wang, W.; Qiao, W.; Li, C.; Zeng, F.; Fu, L. Short-term load forecasting using detrend singular spectrum fluctuation analysis. Energy 2022, 256, 124722. [Google Scholar] [CrossRef]
  46. Laouafi, A.; Laouafi, F.; Boukelia, T.E. An adaptive hybrid ensemble with pattern similarity analysis and error correction for short-term load forecasting. Appl. Energy 2022, 322, 119525. [Google Scholar] [CrossRef]
  47. Khan, N.; Haq, I.U.; Khan, S.U.; Rho, S.; Lee, M.Y.; Baik, S.W. DB-Net: A novel dilated CNN based multi-step forecasting model for power consumption in integrated local energy systems. Int. J. Electr. Power Energy Syst. 2021, 133, 107023. [Google Scholar] [CrossRef]
  48. Chen, L.C.; Papandreou, G.; Schroff, F.; Adam, H. Rethinking atrous convolution for semantic image segmentation. arXiv 2017, arXiv:1706.05587. [Google Scholar]
  49. Bai, S.; Kolter, J.Z.; Koltun, V. Convolutional sequence modeling revisited. In Proceedings of the 2018 International Conference of Learning Representation Workshop, Vancouver, BC, Canada, 30 April–3 May 2018; Available online: https://openreview.net/forum?id=BJEX-H1Pf (accessed on 26 June 2022).
  50. Graves, A.; Fernández, S.; Schmidhuber, J. Bidirectional LSTM networks for improved phoneme classification and recognition. In Proceedings of the International Conference on Artificial Neural Networks, Warsaw, Poland, 11–15 September 2005; Springer: Berlin/Heidelberg, Germany, 2005; pp. 799–804. [Google Scholar]
  51. Jia, M.; Huang, J.; Pang, L.; Zhao, Q. Analysis and research on stock price of LSTM and bidirectional LSTM neural network. In Proceedings of the 3rd International Conference on Computer Engineering, Information Science & Application Technology (ICCIA 2019), Chongqing, China, 30–31 May 2019; Atlantis Press: Paris, France, 2019; pp. 467–473. [Google Scholar]
  52. Wang, T.; Lai, C.S.; Ng, W.W.; Pan, K.; Zhang, M.; Vaccaro, A.; Lai, L.L. Deep autoencoder with localized stochastic sensitivity for short-term load forecasting. Int. J. Electr. Power Energy Syst. 2021, 130, 106954. [Google Scholar] [CrossRef]
  53. Smart-Grid Smart-City Customer Trial Data. Available online: https://data.gov.au/data/dataset/smart-grid-smart-city-customer-trial-data (accessed on 26 June 2022).
  54. Khan, Z.A.; Hussain, T.; Ullah, A.; Rho, S.; Lee, M.; Baik, S.W. Towards efficient electricity forecasting in residential and commercial buildings: A novel hybrid CNN with a LSTM-AE based framework. Sensors 2020, 20, 1399. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Structure of residential energy supply.
Figure 1. Structure of residential energy supply.
Buildings 13 00072 g001
Figure 2. Architecture of the proposed DCNN-LSTM-AE-AM.
Figure 2. Architecture of the proposed DCNN-LSTM-AE-AM.
Buildings 13 00072 g002
Figure 3. Process of residential short-term load forecasting.
Figure 3. Process of residential short-term load forecasting.
Buildings 13 00072 g003
Figure 4. Structure of CNN and DCNN: (a) CNN; (b) DCNN.
Figure 4. Structure of CNN and DCNN: (a) CNN; (b) DCNN.
Buildings 13 00072 g004
Figure 5. Internal structure of LSTM.
Figure 5. Internal structure of LSTM.
Buildings 13 00072 g005
Figure 6. Standard structure of BiLSTM.
Figure 6. Standard structure of BiLSTM.
Buildings 13 00072 g006
Figure 7. Structure of attention mechanism.
Figure 7. Structure of attention mechanism.
Buildings 13 00072 g007
Figure 8. Visualization of selected datasets: (a) IHEPC at a 15-min resolution; (b) SGSC at a 30-min resolution.
Figure 8. Visualization of selected datasets: (a) IHEPC at a 15-min resolution; (b) SGSC at a 30-min resolution.
Buildings 13 00072 g008
Figure 9. MSE of the proposed model under different hyperparameters.
Figure 9. MSE of the proposed model under different hyperparameters.
Buildings 13 00072 g009
Figure 10. Prediction results of different methods on the IHEPC dataset at a 15-min resolution.
Figure 10. Prediction results of different methods on the IHEPC dataset at a 15-min resolution.
Buildings 13 00072 g010
Figure 11. Prediction results of some sole models on the IHEPC dataset at a 15-min resolution.
Figure 11. Prediction results of some sole models on the IHEPC dataset at a 15-min resolution.
Buildings 13 00072 g011
Figure 12. Prediction results of some hybrid models on the IHEPC dataset at a 15-min resolution.
Figure 12. Prediction results of some hybrid models on the IHEPC dataset at a 15-min resolution.
Buildings 13 00072 g012
Figure 13. Prediction results of DCNN-LSTM-AE-AM on the SGSC dataset at a 30-min resolution.
Figure 13. Prediction results of DCNN-LSTM-AE-AM on the SGSC dataset at a 30-min resolution.
Buildings 13 00072 g013
Table 1. Information from the IHEPC dataset.
Table 1. Information from the IHEPC dataset.
VariableDescription
DataRecorded date
TimeCurrent moment
Global active powerSum of active power per minute
Global reactive powerSum of reactive power per minute
VoltageVoltage per minute
Global intensitySum of current per minute
Sub metering1Power used by kitchen per minute
Sub metering2Power used by room per minute
Sub metering3Power used by bathroom per minute
Table 2. Hyperparameter settings.
Table 2. Hyperparameter settings.
NetworkHyperparameters
1D-ConvThe convolution kernel size is 3, the number of filters is 12, the dilation rate is 2, and the activation function is ReLU
1D-ConvThe convolution kernel size is 3, the number of filters is 24, the dilation rate is 2, and the activation function is ReLU
1D-SpatialDropout0.1
BiLSTM32 units
BiLSTM32 units
LSTM32 units
LSTM32 units
1D-SpatialDropout0.2
Attention32 units
Dense96 units
Dense32 units
Dense1 unit
Table 3. Performance of our proposed method under different data lengths.
Table 3. Performance of our proposed method under different data lengths.
LengthMSERMSEMAEMAPE
60.004340.06590.03510.7599
80.004200.06480.03320.8521
100.004180.06470.03440.7662
120.004110.06410.03310.6175
140.004150.06440.03340.6757
160.004260.06530.03380.7421
180.004440.06660.03480.7952
200.004500.06710.03570.7853
220.004580.06770.03660.8322
240.004460.06680.03700.8964
Table 4. Prediction performance comparisons on the SGSC dataset.
Table 4. Prediction performance comparisons on the SGSC dataset.
MethodResolutionMSERMSEMAEMAPE
XGBoost15 min0.04630.21520.15410.8683
1 h0.04100.20250.11200.7268
DCNN15 min0.04030.20070.11920.6869
1 h0.04120.20300.11320.7284
LSTM15 min0.04910.22160.12631.0384
1 h0.05480.23410.13561.4831
AM15 min0.06630.25750.12281.4286
1 h0.06870.26210.13461.3788
LSTM-AE15 min0.01790.13380.08550.7863
1 h0.02320.15230.08780.8265
CNN-LSTM15 min0.01580.12570.07120.6930
1 h0.01970.14030.09970.7556
DCNN-AM15 min0.00810.09000.04520.6938
1 h0.00910.09540.04660.7028
DCNN-LSTM-AE15 min0.02220.14900.07090.7380
1 h0.02950.17180.08230.7428
LSTM-AE-AM15 min0.00430.06560.03780.6854
1 h0.00950.09750.06820.7530
DCNN-LSTM-AE-AM15 min0.00410.06400.03330.6757
1 h0.00860.09270.06670.7257
Table 5. Prediction performance comparisons on the SGSC dataset.
Table 5. Prediction performance comparisons on the SGSC dataset.
MethodResolutionMSERMSEMAEMAPE
XGBoost30 min0.04560.21350.12030.9298
1 h0.04030.20070.09800.8296
DCNN30 min0.04330.20810.13290.6796
1 h0.04650.21560.13660.7862
LSTM30 min0.04830.21980.12981.1001
1 h0.05220.22850.14011.5131
AM30 min0.06520.25530.18931.6235
1 h0.06890.26250.20061.7692
LSTM-AE30 min0.01660.12880.07990.7567
1 h0.02180.14760.07620.7992
CNN-LSTM30 min0.01420.11920.08040.6728
1 h0.01820.13490.09910.7256
DCNN-AM30 min0.00830.09110.04890.6118
1 h0.00890.09430.04960.6412
DCNN-LSTM-AE30 min0.02420.15560.07560.7128
1 h0.02890.17000.08620.7196
LSTM-AE-AM30 min0.00480.06930.03840.6203
1 h0.00560.07480.06960.6495
DCNN-LSTM-AE-AM30 min0.00410.06400.03290.5901
1 h0.00810.09000.06570.6336
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ji, X.; Huang, H.; Chen, D.; Yin, K.; Zuo, Y.; Chen, Z.; Bai, R. A Hybrid Residential Short-Term Load Forecasting Method Using Attention Mechanism and Deep Learning. Buildings 2023, 13, 72. https://doi.org/10.3390/buildings13010072

AMA Style

Ji X, Huang H, Chen D, Yin K, Zuo Y, Chen Z, Bai R. A Hybrid Residential Short-Term Load Forecasting Method Using Attention Mechanism and Deep Learning. Buildings. 2023; 13(1):72. https://doi.org/10.3390/buildings13010072

Chicago/Turabian Style

Ji, Xinhui, Huijie Huang, Dongsheng Chen, Kangning Yin, Yi Zuo, Zhenping Chen, and Rui Bai. 2023. "A Hybrid Residential Short-Term Load Forecasting Method Using Attention Mechanism and Deep Learning" Buildings 13, no. 1: 72. https://doi.org/10.3390/buildings13010072

APA Style

Ji, X., Huang, H., Chen, D., Yin, K., Zuo, Y., Chen, Z., & Bai, R. (2023). A Hybrid Residential Short-Term Load Forecasting Method Using Attention Mechanism and Deep Learning. Buildings, 13(1), 72. https://doi.org/10.3390/buildings13010072

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop