1. Introduction
MGs, characterized by cleanliness, low-carbon emissions, and openness, have garnered significant attention due to the rapid development of renewable energy in recent years [
1]. As a critical solution to improve the consumption of distributed power sources and the reliability of power supply, MGs have become essential to reducing fossil energy pollution and promoting sustainable development in China [
2]. However, the intermittent, volatile, and uncertain nature of renewable energy sources poses significant challenges to the stable operation of MGs [
3]. Additionally, the expansion of the MG system and the increased number of its components also impose more demanding requirements on the optimal scheduling method. The traditional scheduling method based on numerical model optimization, scenario matching, and manual manning makes it difficult to meet the demand. Hence, studying fast, accurate, and intelligent scheduling decision methods holds immense practical value and significance [
4].
Currently, the predominant method for solving MG optimal scheduling problems is a numerical calculation method grounded in optimization theory. Common optimal scheduling models encompass MILP [
5], dynamic programming [
6], distributed optimization [
7], etc. Likewise, common model-solving algorithms involve intelligent algorithms [
8], second-order cone relaxation methods [
9], Lagrange relaxation methods, etc. However, as the uncertainties of the source–load dual-side within MGs escalate, solving the optimal scheduling problem under such uncertainties becomes a more realistic and challenging research problem [
10]. Some researchers construct uncertainty planning models. The main modeling and solution methods are stochastic planning [
11], chance-constrained planning [
12], etc. Among these, robust optimization [
13,
14] methods have been proven to be an effective method of solving MG uncertain optimization problems. They aim at optimal operation under the worst-case scenario. However, their overly pessimistic view of uncertain variables may lead to solution results that are too conservative to be economical. The mathematical models of these methods are relatively complex and computationally expensive. The other researchers used a multi-timescale optimal scheduling strategy [
15], which can be classified into day-ahead scheduling and intra-day stage according to the timescale. Among these, the model predictive control (MPC) technique is a widely employed modeling approach [
16]. How to enhance intra-day real-time scheduling computational efficiency is still a challenge.
In summary, the traditional optimization theory-based scheduling methods rely on strict mathematical derivation, requiring researchers to participate. With the MG evolving into a new system characterized by increased uncertainty and complexity, the traditional optimization scheduling methods are gradually becoming inadequate to meet the demands of MG operation [
17]. Several critical problems of this method are as follows:
- (1)
The traditional mathematical optimization methods cannot model the components of the MG in a fast and refined manner, but it is difficult to describe the physical characteristics of the actual operation of the components using a simplified model [
18].
- (2)
The traditional mathematical MG scheduling models are often nonlinear and nonconvex, which is a typical nondeterministic polynomial problem (NP-hard). The problem is demanding on the solution algorithm, and it is not easy to find the optimal solution.
- (3)
The computational process of traditional mathematical optimization methods is complex and inefficient, and it is difficult to adapt to the real-time solution of optimization scheduling problems with uncertainty under complex and variable system operating conditions [
19].
- (4)
The traditional mathematical optimization methods ignore the significance of historical data and historical decision-making plans and fail to make use of the valuable historical decision-making data information accumulated during the system’s operation.
Recently, the rapid development of computer technology has made neural networks (NN) an important driver of the new technological revolution and industrial change [
20]. A new intelligent decision-making method using NNs based on big data technology may be a more effective way of thinking, which may help to break through the limitations of mathematical optimization solution methods. Unlike traditional optimization methods, the decision-making method based on NNs no longer depends on specific mathematical models or algorithms; instead, it is trained using extensive real data [
21]. This method can greatly simplify the process and complexity of modeling and solving the optimal scheduling problem, and cope with various theoretical problems and challenges that keep emerging through its self-learning and self-evolution process. It can potentially facilitate the transition from manual supervision to machine intelligence-based monitoring in the domain of MG scheduling. Moreover, when the data make centralized training of models bitter due to factors such as privacy and size, the idea of distributed frameworks [
22,
23] can also be referred to for decentralized training of small models and then aggregated to a big model. This allows great flexibility in the implementation of the method.
Figure 1 shows the transition from the traditional optimization method to the NN-based method.
Several scholars have attempted to utilize artificial intelligence (AI) techniques in the field of scheduling decisions. The literature [
24] utilized long and short-term memory (LSTM) to establish the mapping from system load to unit output. However, the constructed network structure is relatively simple, and the results are unconstrained. The literature [
25] uses a multi-layer perceptron (MLP) to learn and mimic the scheduling decision of a smart grid, and an iterative algorithm is used to correct the output of the NNs so that it satisfies the actual constraints. The literature [
26] applies a feedforward neural network (FNN) for the optimal scheduling of combined heat and power (CHP) systems, which enhances computational efficiency by about 7000 times while permitting suboptimal cost. Although previous studies have demonstrated that NNs are feasible and effective in optimal energy scheduling, the current research still faces some issues:
- (1)
Only load data are used as training inputs without considering the influence of other system state data on the scheduling decision results. This approach cannot fully extract the feature information embedded in the valuable historical operation data.
- (2)
Only using a shallow or single network model to build the scheduling mapping relationship, the accuracy of the output results is low.
- (3)
The decision results from the NNs-based scheduling method will inevitably violate some actual constraints, and there is no reasonable and efficient solution to this issue.
To address the above issues, this paper proposes a two-stage optimal scheduling method for MGs. The proposed method aims to enhance the effectiveness of the NNs-driven scheduling method and the MG’s ability to handle uncertain fluctuations and address the limitations of the traditional mathematical model-driven and manual participation scheduling methods. In the day-ahead part (1 h timescale), which does not require high timeliness, a MILP model is used to obtain the MG’s operating plan. In the day-ahead part (15 min timescale), a DNN scheduling decision network is used for fast-rolling optimization. The main contributions of this paper are as follows:
- (1)
An intra-day rolling optimization model based on DNNs and big data is proposed, which is trained using the dataset clustered by the K-means algorithm to improve generalizability and accuracy.
- (2)
A novel CNN-Bi LSTM scheduling decision network is proposed, digging deep feature information in the system operation data by CNN and establishing the accurate mapping between input and output by Bi LSTM.
- (3)
A power balance correction algorithm is proposed to fine-tune the DNN outputs to quickly satisfy all practical constraints.
The proposed method can effectively reduce the complexity of solving the optimal scheduling problem and significantly improve computational efficiency (reducing the solution time for intra-day rolling optimization to milliseconds), which also improves the intelligence level of MGs. The rest of this paper is organized as follows:
Section 2 presents the basic mathematical optimization model for day-ahead MG scheduling.
Section 3 presents the DNN-based intra-day rolling optimal scheduling method.
Section 4 presents simulation experiments and analyses.
Section 5 presents the conclusions of this paper.
3. Deep Neural Network-Based Intra-Day Rolling Optimization Method
A data-driven DNN-based scheduling method is proposed in this paper to address the shortcomings and difficulties of traditional methods in intra-day rolling optimization. Instead of relying on specific mathematical models, it trains with large amounts of real data and makes scheduling decisions by high-dimensional matrix multiplication [
28]. This method can reduce the complexity of solving the optimal scheduling problem and significantly improve computational efficiency.
3.1. Intra-Day MPC Rolling Optimization Forms
In this paper, the DNN scheduling decision network is used as an optimizer for MPC to perform intra-day rolling optimization. MPC is an alternating process of continuously rolling local optimization and continuously rolling control role implementation. By obtaining ultra-short-term power forecast information in real-time during intra-day scheduling and using the actual scheduling results and new forecast information as feedback, MPC rolling optimization forms can greatly reduce the impact of MG uncertainties on optimal operating scheduling. The general steps of MPC rolling optimization can be expressed as follows:
Step 1. Based on the current moment and the current system state, the system state in the future period is obtained by a certain prediction model.
Step 2. Based on the system state in a future period, the optimization problem in that period is solved to obtain the control sequence in that period.
Step 3. Only the action of the first moment of the control sequence is applied to the system, and the above steps are repeated for the next moment.
3.2. Total Framework of the DNN-Based Intra-Day Scheduling Decision Method
The overall framework of the DNN-based intra-day scheduling method is shown in
Figure 3, which mainly includes: the training dataset construction stage, offline training stage, and online decision stage.
- (1)
Training dataset construction stage. To improve the accuracy and reduce the pressure on the network’s generalizability, the numerous real operating data of MG collected are clustered by the K-means algorithm [
29], dividing into different training sets. The net system load demand
, which is a 96-dimensional time series represented as
, is used as the clustering index.
- (2)
Offline training stage. A two-dimensional time series feature map containing the system operation state is constructed as the input for the CNN-Bi LSTM network. The optimal scheduling plan is the network’s output, training multiple scheduling decision networks with different training datasets.
- (3)
Online decision stage. The system’s ultra-short-term prediction state is combined with the day-ahead operation plan and fed into the well-trained CNN-Bi LSTM network. The outputs of the network are fine-tuned by a power correction algorithm to get the final scheduling decision.
3.3. Introduction to Deep Neural Networks
3.3.1. Convolutional Neural Networks
The efficient feature extraction ability of the CNN makes it the most widely used model in the field of deep learning. The CNN primarily comprises a convolutional layer and a pooling layer. The convolutional layer performs effective nonlinear local feature extraction using convolutional kernels, while the pooling layer compresses the extracted features and generates more significant feature information to enhance generalization capability [
30]. The basic structure of the CNN is shown in
Figure 4.
3.3.2. Bidirectional Long and Short-Term Memory Networks
We start by introducing the LSTM network, which contains forgetting gates, input gates, and output gates, and the basic structure is shown in
Figure 5.
In
Figure 5,
and
represent Sigmoid and Tanh activation functions, respectively. The calculation of the data within LSTM is as follows:
where
and
denote the weight matrix and bias vector, respectively.
represents dot product.
and
denote the output of the last and current moments, respectively.
and
denote the memory state of the last and current moments, respectively.
is the Intermediate state of the network.
and
denote that the current states add degree and output degrees, respectively.
is the input of the current moment.
and
represent the sigmoid and tanh activation functions, respectively.
The LSTM structure gathers feature information only from the current input and past time series at each time while disregarding feature information from future time series. In this paper, bidirectional LSTM is used as the back-end mapping network of the scheduling decision network to improve the accuracy of the decision results and the performance of temporal feature extraction. The Bi LSTM is a variant structure of LSTM that includes both forward LSTM and backward LSTM layers [
31]. The Bi LSTM structure enables it to gather information from both forward and backward directions, enabling the network to consider past and future data. This enhances the model’s feature extraction ability without requiring additional data. The structure of Bi LSTM is illustrated in
Figure 6.
3.4. The CNN-Bi LSTM Intra-Day Scheduling Decision Network
Trained by a large amount of real operation data, the CNN-Bi LSTM intra-day scheduling decision network can learn the regularity between the system state and the scheduling decision result. Once the parameters are fixed in the network, it can provide the optimal scheduling plan extremely fast under any operating scenario.
3.4.1. Input and Output of the CNN-Bi LSTM
The CNN-Bi LSTM scheduling decision network imitates the idea of MPC for intra-day rolling optimization, with the prediction domain set to 2 h and the control domain set to 15 min. To deeply mine the implicit value information in the system operating data, we set the input
of this network in the form of a 2-D time series grayscale graph. The output
of the network is the optimal scheduling plan. The specific expression is as follows:
is a matrix consisting of the intra-day state vector of the system (power load, WT, and PV) in the period to and the day-ahead operating plan vector of controllable devices (SB, UG, and MTs) in the corresponding time. The number 9 indicates the number of input features. is a vector consisting of the controllable devices’ intra-day optimal operating plan in the period to . The number 6 indicates the number of controllable devices in output features. is the day-ahead operating plan, and is the intra-day optimal operating plan. Since the MPC prediction domain is set to 2 h, the is set as 7 in this paper and all the above variables are real.
3.4.2. Structure of the CNN-Bi LSTM
Since the mapping relationship between the system operating state and scheduling decision is complex, this paper uses a multilayer CNN-Bi LSTM network for deep mining of the data. This network is mainly constituted by a three-layer CNN and a three-layer Bi LSTM, and linked by a Flatten layer. The CNN primarily extracts the power correlation feature, while the Bi LSTM focuses on extracting the power time series feature. The batch normalization (BN) layer can solve the problem of numerical instability in DNNs, making the distribution of individual features in the same batch similar. In this paper, the BN layer is inserted between each convolutional layer and pooling layer to normalize the features in the network and accelerate training. The dropout layer is the layer used after each Bi LSTM to enhance the generalization performance of the network. Finally, the data are adjusted to a vector output in the specified size through a fully connected (Dense) layer. The specific structure of the proposed CNN-Bi LSTM in this paper is shown in
Figure 7.
3.4.3. Settings of the CNN-Bi LSTM
To better extract and abstract the input feature, the number of convolutional kernels is set to 64, 128, and 256, and the size of convolutional kernels is set to 7 × 7, 5 × 5, and 3 × 3. The number of neurons of Bi LSTM is set to 256, 128, and 64, respectively, and the drop rate of the dropout layer is set uniformly to 0.25.
Normalize the training data of the network to between 0 and 1 using the maximum–minimum normalization method. The network is trained using the Adam optimization algorithm [
32] and the root mean square error (RMSE) is set as the loss function of the network, which is defined as follows:
where
and
are the true and predicted scheduling plans for
th device at time
, respectively.
is the number of controllable devices in MG.
3.5. The Power Balance Correction Algorithm
Like load forecasting, the DNN-based scheduling method is fundamentally a process of nonlinear regression. Consequently, the output inevitably does not meet certain practical constraints. To address this issue, we use a power balance correction algorithm (PBC) to adjust the output, making it practical for use.
Inspired by the average consistency algorithm, we utilize the difference between total power demand and total generation at time
as the consistency indicator. The outputs from DNN are updated by iteration (Equations (30) and (31)). Any updated results that violate the operating constraints of the device require additional correction (Equation (32)). This algorithm is denoted as follows:
where
is the number of iterations.
and
are the power of
th power generator at time
and the total number of generators in MG, respectively.
So far, MG’s intra-day optimal scheduling model based on CNN-Bi LSTM-PBC is completely constructed.