Figure 1.
Generalized time series architecture.
Figure 1.
Generalized time series architecture.
Figure 2.
Relation between input and output dimensionality for frame-based time series modelling: (a) sequence-to-point, (b) sequence-to-subsequence, and (c) sequence-to-sequence.
Figure 2.
Relation between input and output dimensionality for frame-based time series modelling: (a) sequence-to-point, (b) sequence-to-subsequence, and (c) sequence-to-sequence.
Figure 3.
Overview of implemented modules and functionalities in the PyDTS toolkit. Inputs and preprocessing are indicated in red, features and data sequencing in yellow, modelling in green, postprocessing in blue, and visual elements and outputs in purple.
Figure 3.
Overview of implemented modules and functionalities in the PyDTS toolkit. Inputs and preprocessing are indicated in red, features and data sequencing in yellow, modelling in green, postprocessing in blue, and visual elements and outputs in purple.
Figure 4.
Internal data pipeline of PyDTS including training and testing modules and external data, model, and setup databases.
Figure 4.
Internal data pipeline of PyDTS including training and testing modules and external data, model, and setup databases.
Figure 5.
Grid search for the optimal number of input samples depending on the time series problem.
Figure 5.
Grid search for the optimal number of input samples depending on the time series problem.
Figure 6.
DL layer architectures for DNNs, LSTM, and CNN models. For CNNs, the notation of the convolutional layer is Conv1D(x,y) with x being the number of filters and y being the kernel size. For pooling layers MaxPool(x,y), x is the size and y the stride, while for LSTM and DNN layers, x denotes the number of neurons.
Figure 6.
DL layer architectures for DNNs, LSTM, and CNN models. For CNNs, the notation of the convolutional layer is Conv1D(x,y) with x being the number of filters and y being the kernel size. For pooling layers MaxPool(x,y), x is the size and y the stride, while for LSTM and DNN layers, x denotes the number of neurons.
Figure 7.
Predicted appliance current draw for 12 h for three different (FRE, HPE, and CDE) appliances from the AMPds2 dataset on 9 January 2013 at 12:00 p.m.
Figure 7.
Predicted appliance current draw for 12 h for three different (FRE, HPE, and CDE) appliances from the AMPds2 dataset on 9 January 2013 at 12:00 p.m.
Figure 8.
Forecasted power consumption and error for phase L1 for 1 week using RF as regression model.
Figure 8.
Forecasted power consumption and error for phase L1 for 1 week using RF as regression model.
Figure 9.
Predicted temperature for stator winding and rotor magnet for IDs 60, 62, and 72.
Figure 9.
Predicted temperature for stator winding and rotor magnet for IDs 60, 62, and 72.
Figure 10.
Confusion matrices for (a) raw, (b) statistical, and (c) frequency domain features for the CNN model.
Figure 10.
Confusion matrices for (a) raw, (b) statistical, and (c) frequency domain features for the CNN model.
Figure 11.
Ground-truth and predicted remaining cell charge and prediction error using the best-performing DNN model (for visibility, the predicted output has been filtered with a median filter of a length of 100 samples).
Figure 11.
Ground-truth and predicted remaining cell charge and prediction error using the best-performing DNN model (for visibility, the predicted output has been filtered with a median filter of a length of 100 samples).
Figure 12.
Feature ranking for the nonlinear modelling task for 13 features: coolant/ambient temperature (, ), stator voltages (, , ), stator currents (, , ), torque (), rotational speed (), apparent power (), and products or current/power and rotational speed (, ).
Figure 12.
Feature ranking for the nonlinear modelling task for 13 features: coolant/ambient temperature (, ), stator voltages (, , ), stator currents (, , ), torque (), rotational speed (), apparent power (), and products or current/power and rotational speed (, ).
Table 1.
Comparison of relevant properties between different modelling approaches: (+): comparatively better, (o): neutral, and (-): comparatively worse.
Table 1.
Comparison of relevant properties between different modelling approaches: (+): comparatively better, (o): neutral, and (-): comparatively worse.
Properties | Ref. and Eq. | Linear Algebra | Statistical Modelling | Machine Learning |
---|
Runtime | [51] | o | + | - |
Memory | [51] | o | + | - |
Interpretability | (12)–(14) | + | o | - |
Dimensionality | (10), (11), (14) | o | - | + |
Transferability | [52] | o | - | + |
Nonlinear | (10), (11), (14) | o | - | + |
Hyperparameters | (10), (11), (14) | o | + | - |
Training data | [53] | + | o | - |
Table 2.
Short description of the datasets. The feature column includes the following abbreviations: active power (P), reactive power (Q), apparent power (S), current (I), voltage (V), temperature (T), relative humidity (), solar irradiance (), wind speed (), rotational speed (n), torque (M), and acceleration (A). Similarly, the outputs include the appliance current (), the per-phase power (), the stator winding and rotor magnet temperatures (), the motor state, and the remaining battery charge ().
Table 2.
Short description of the datasets. The feature column includes the following abbreviations: active power (P), reactive power (Q), apparent power (S), current (I), voltage (V), temperature (T), relative humidity (), solar irradiance (), wind speed (), rotational speed (n), torque (M), and acceleration (A). Similarly, the outputs include the appliance current (), the per-phase power (), the stator winding and rotor magnet temperatures (), the motor state, and the remaining battery charge ().
Name | Ref. | Scenario | Length | Sampling | Features | Output | Max | Mean | Std |
---|
AMPds2 | [54] | Denoise | 2 y | 60 s | P, Q, S, I | | 105 | 0.8 | 10.9 |
Energy | [55] | Forecast | 1 y | 10 min | P, T, , | | 52.2 | 23.7 | 12.2 |
Motor Temp. | [47] | Nonlinear | 185 h | 0.5 s | V, I, T, M, n | | 141.4 | 57.5 | 22.7 |
Ford Motor | [56] | Anomaly | 1.4 h | 2 ms | | s | 1.0 | 0.49 | 0.50 |
Battery Health | [57] | Degradation | 57 days | 2.5 s | V, I, T | | 1.92 | 1.54 | 0.17 |
Table 3.
Optimized model parameters for ML approaches including KNN, RF, and SVM.
Table 3.
Optimized model parameters for ML approaches including KNN, RF, and SVM.
Model | Parameter | Optimal | Range | Step |
---|
KNN | Neighbors | 140 | 10–200 | 5 |
RF | Max. Depth | 10 | 5-25 | 5 |
Split | 4 | 2–10 | 2 |
#-Trees | 128 | 2–256 | |
SVM | Kernel | rbf | linear, rbf, poly | - |
C | 100 | 1–200 | 20 |
Gamma | 0.1 | 0.001–1 | |
Table 4.
Hyper- and solver parameters for deep learning models including DNN, CNN, and LSTM.
Table 4.
Hyper- and solver parameters for deep learning models including DNN, CNN, and LSTM.
Hyperparameters | Solver Parameters |
---|
Batch | 1000 | optimizer | adam |
Epochs | 50–200 | loss | mae |
Patience | 15 | Learning rate | |
Validation steps | 50 | Beta1 | 0.9 |
Shuffle | False | Beta2 | 0.999 |
Table 5.
Average results (A) for the energy disaggregation task for fivefold cross-validation using different models and accuracy metrics. The best performances are indicated with bold notation.
Table 5.
Average results (A) for the energy disaggregation task for fivefold cross-validation using different models and accuracy metrics. The best performances are indicated with bold notation.
Model | NMSE | RMSE | MSE | MAE | MAX |
---|
CNN | 92.48 | 0.64 | 0.41 | 0.08 | 29.01 |
LSTM | 94.51 | 0.60 | 0.36 | 0.08 | 30.54 |
DNN | 94.39 | 0.66 | 0.44 | 0.08 | 31.85 |
RF | 81.39 | 0.63 | 0.40 | 0.10 | 28.60 |
KNN | 74.11 | 1.15 | 1.32 | 0.21 | 31.09 |
Table 6.
Per-device results (A) for the energy disaggregation task for fivefold cross-validation using LSTM as regression model and different accuracy metrics.
Table 6.
Per-device results (A) for the energy disaggregation task for fivefold cross-validation using LSTM as regression model and different accuracy metrics.
Device | NMSE | RMSE | MSE | MAE | MAX |
---|
DWE | 49.79 | 0.87 | 0.76 | 0.12 | 6.76 |
FRE | 95.15 | 0.24 | 0.06 | 0.13 | 3.41 |
HPE | 97.55 | 0.63 | 0.40 | 0.07 | 7.21 |
WOE | 91.50 | 0.63 | 0.40 | 0.03 | 30.61 |
CDE | 97.66 | 0.62 | 0.38 | 0.02 | 40.73 |
Avg | 94.51 | 0.60 | 0.36 | 0.08 | 30.54 |
Table 7.
Comparison with the literature for the energy disaggregation task.
Table 7.
Comparison with the literature for the energy disaggregation task.
Ref. | Year | Model | NMSE | RMSE | MAE |
---|
[60] | 2016 | HMM | 94.1% | - | - |
[61] | 2019 | CNN | 93.9% | - | - |
[62] | 2020 | CNN | 94.7% | - | - |
[58] | 2021 | CNN | 95.8% | - | - |
[43] | 2022 | CNN | 94.7% | 0.48 | 0.06 |
This Work | 2023 | LSTM | 94.5% | 0.60 | 0.08 |
Table 8.
Forecasting errors (kW) using Seq2Point for a 24 h ahead prediction window with different models and accuracy metrics using fivefold cross-validation. The best performances are indicated with bold notation.
Table 8.
Forecasting errors (kW) using Seq2Point for a 24 h ahead prediction window with different models and accuracy metrics using fivefold cross-validation. The best performances are indicated with bold notation.
Model | NMSE | RMSE | MSE | MAE | MAX |
---|
CNN | 95.72 | 3.62 | 13.10 | 2.77 | 18.49 |
LSTM | 95.55 | 3.85 | 14.82 | 2.88 | 18.19 |
DNN | 95.61 | 3.74 | 13.99 | 2.85 | 17.90 |
RF | 97.50 | 2.42 | 5.87 | 1.60 | 17.88 |
KNN | 93.98 | 4.96 | 24.60 | 3.88 | 18.63 |
Table 9.
Forecasting errors (kW) using Seq2Seq for a 24 h ahead prediction window with different models and accuracy metrics using fivefold cross-validation. The best performances are indicated with bold notation.
Table 9.
Forecasting errors (kW) using Seq2Seq for a 24 h ahead prediction window with different models and accuracy metrics using fivefold cross-validation. The best performances are indicated with bold notation.
Model | NMSE | RMSE | MSE | MAE | MAX |
---|
CNN | 95.88 | 3.54 | 12.53 | 2.67 | 18.61 |
LSTM | 95.99 | 3.01 | 9.06 | 2.36 | 12.12 |
DNN | 95.66 | 3.71 | 13.76 | 2.81 | 17.26 |
Table 10.
Temperature prediction results for 5-fold cross validation using different regression models and performance metrics. Due to memory restrictions the LSTM input was reduced to 500 samples. The best performances are indicated with bold notation.
Table 10.
Temperature prediction results for 5-fold cross validation using different regression models and performance metrics. Due to memory restrictions the LSTM input was reduced to 500 samples. The best performances are indicated with bold notation.
Model | NMSE | RMSE | MSE | MAE | MAX |
---|
| | | | | | | | | |
---|
CNN | 97.67 | 95.19 | 4.54 | 7.59 | 20.61 | 57.61 | 3.06 | 5.53 | 76.43 | 54.18 |
LSTM | 96.71 | 93.23 | 6.39 | 10.6 | 40.83 | 112.4 | 4.28 | 7.85 | 77.15 | 60.05 |
DNN | 97.37 | 95.21 | 5.32 | 7.81 | 28.30 | 61.00 | 3.43 | 5.59 | 76.52 | 59.20 |
RF | 96.04 | 94.66 | 7.63 | 8.30 | 58.22 | 68.89 | 5.26 | 4.43 | 73.73 | 47.87 |
KNN | 86.40 | 89.85 | 22.79 | 14.98 | 519.4 | 224.4 | 17.39 | 11.45 | 82.24 | 57.96 |
Table 11.
Results for MSE (K²) and MAX (K) errors for different testing IDs, their respective time (hr), and temperature hot spots using a CNN regression model per hot spot.
Table 11.
Results for MSE (K²) and MAX (K) errors for different testing IDs, their respective time (hr), and temperature hot spots using a CNN regression model per hot spot.
ID | Time | Stator Winding | Stator Tooth | Stator Yoke | Magnet |
---|
MSE | MAX | MSE | MAX | MSE | MAX | MSE | MAX |
---|
60 | 1.7 | 2.41 | 5.03 | 1.68 | 4.28 | 1.16 | 3.14 | 22.62 | 9.90 |
62 | 3.3 | 2.75 | 6.23 | 1.25 | 3.78 | 1.22 | 3.96 | 17.49 | 9.74 |
74 | 3.0 | 3.33 | 6.18 | 2.42 | 5.43 | 1.80 | 5.00 | 14.47 | 10.81 |
Avg | 8.0 | 2.90 | 6.23 | 1.78 | 5.43 | 1.42 | 5.00 | 17.45 | 10.81 |
Table 12.
Comparison for temperature prediction using different models and number of input features.
Table 12.
Comparison for temperature prediction using different models and number of input features.
Ref. | Year | Model | MSE | MAX | Features |
---|
[63] | 2021 | MLP | 5.58 | 14.29 | 81 |
[63] | 2021 | OLS | 4.47 | 9.85 | 81 |
[47] | 2020 | CNN | 4.43 | 15.54 | 81 |
[59] | 2023 | TNN | 2.87 | 6.02 | 5 |
This Work | 2023 | CNN | 5.89 | 10.81 | 13 |
Table 13.
Classification results in terms of ACC and F1 for anomaly detection using different classification models. The best performances are indicated with bold notation.
Table 13.
Classification results in terms of ACC and F1 for anomaly detection using different classification models. The best performances are indicated with bold notation.
Model | Raw | Statistical | Frequency |
---|
ACC | F1 | ACC | F1 | ACC | F1 |
---|
CNN | 92.35 | 92.34 | 56.52 | 55.87 | 94.85 | 94.85 |
LSTM | 51.06 | 50.52 | 55.30 | 54.90 | 51.59 | 35.12 |
DNN | 80.15 | 80.15 | 56.52 | 56.13 | 94.77 | 94.77 |
RF | 72.80 | 72.77 | 59.09 | 59.10 | 92.42 | 92.42 |
KNN | 72.80 | 72.76 | 58.11 | 58.12 | 88.94 | 88.90 |
SVM | 51.59 | 35.12 | 58.41 | 58.01 | 94.47 | 94.47 |
Table 14.
Degradation errors for different regression models and performance metrics using Seq2Point learning. The best performances are indicated with bold notation.
Table 14.
Degradation errors for different regression models and performance metrics using Seq2Point learning. The best performances are indicated with bold notation.
Model | NMSE | RMSE | MSE | MAE | MAX |
---|
CNN | 98.00 | 0.08 | 0.01 | 0.06 | 0.36 |
LSTM | 97.85 | 0.08 | 0.01 | 0.07 | 0.39 |
DNN | 98.64 | 0.06 | 0.01 | 0.04 | 0.49 |
RF | 95.15 | 0.16 | 0.03 | 0.15 | 0.38 |
KNN | 97.43 | 0.10 | 0.01 | 0.08 | 0.35 |
Table 15.
Degradation errors for different regression models and performance metrics using Seq2Seq learning. The best performances are indicated with bold notation.
Table 15.
Degradation errors for different regression models and performance metrics using Seq2Seq learning. The best performances are indicated with bold notation.
Model | NMSE | RMSE | MSE | MAE | MAX |
---|
CNN | 98.26 | 0.07 | 0.01 | 0.05 | 0.34 |
LSTM | 97.74 | 0.09 | 0.01 | 0.07 | 0.41 |
DNN | 97.85 | 0.09 | 0.01 | 0.07 | 0.41 |
Table 16.
Intratransferability scenario based on load forecasting between phases 1 (L1) and 2 (L2). The best performances are indicated with bold notation.
Table 16.
Intratransferability scenario based on load forecasting between phases 1 (L1) and 2 (L2). The best performances are indicated with bold notation.
Model | L2 (Train L2) | L2 (Train L1) | Loss (%) |
---|
NMSE | RMSE | MAE | NMSE | RMSE | MAE | NMSE | RMSE | MAE |
---|
CNN | 92.02 | 4.22 | 3.36 | 87.61 | 6.34 | 5.19 | 4.79 | 50.24 | 54.46 |
LSTM | 93.21 | 3.58 | 2.81 | 92.88 | 3.70 | 2.94 | 0.35 | 3.35 | 4.63 |
DNN | 92.81 | 3.86 | 3.03 | 87.44 | 6.40 | 5.25 | 5.79 | 65.80 | 73.27 |
RF | 96.02 | 2.35 | 1.71 | 93.28 | 3.44 | 2.78 | 2.85 | 46.38 | 62.57 |
KNN | 91.66 | 4.37 | 3.49 | 89.07 | 5.58 | 4.56 | 2.83 | 27.69 | 30.66 |
Table 17.
Intertransferability scenario based on energy disaggregation between different consumer households (REDD-1,2). The best performances are indicated with bold notation.
Table 17.
Intertransferability scenario based on energy disaggregation between different consumer households (REDD-1,2). The best performances are indicated with bold notation.
Model | REDD2 (Train REDD2) | REDD2 (Train REDD1) | Loss (%) |
---|
NMSE | RMSE | MAE | NMSE | RMSE | MAE | NMSE | RMSE | MAE |
---|
CNN | 92.60 | 39.44 | 5.45 | 76.12 | 70.83 | 16.57 | 16.48 | 79.59 | 204.0 |
LSTM | 86.65 | 84.36 | 9.83 | 71.26 | 94.88 | 19.95 | 15.39 | 12.47 | 102.9 |
DNN | 85.02 | 76.83 | 11.03 | 55.19 | 106.4 | 31.10 | 29.83 | 38.49 | 181.9 |
RF | 89.19 | 41.38 | 7.96 | 75.88 | 67.77 | 16.74 | 13.31 | 63.77 | 110.3 |
KNN | 92.48 | 31.32 | 5.54 | 70.09 | 79.57 | 20.76 | 22.39 | 154.1 | 274.7 |
Table 18.
Model size of the trained model including all parameters for different scenarios.
Table 18.
Model size of the trained model including all parameters for different scenarios.
Model | Denoise | Forecast | Nonlinear | Anomaly | Degradation |
---|
30 × 4 | 144 × 8 | 1000 × 13 | 500 × 1 | 140 × 3 |
---|
CNN | 2.91 MB | 6.37 MB | 32.0 MB | 9.49 MB | 6.20 MB |
LSTM | 4.29 MB | 4.30 MB | 4.34 MB | 4.26 MB | 4.27 MB |
DNN | 1.92 MB | 5.00 MB | 40.6 MB | 2.30 MB | 2.81 MB |
RF | 37.7 MB | 12.1 MB | 58.4 MB | 2.80 MB | 9.16 MB |
KNN | 3.94 GB | 0.33 GB | 26.9 GB | 7.05 MB | 162.4 MB |
Table 19.
Training (T) and inference time (I) per sample (s) for different models and scenarios.
Table 19.
Training (T) and inference time (I) per sample (s) for different models and scenarios.
Model | Denoise | Forecast | Non-Linear | Anomaly | Degradation |
---|
T | I | T | I | T | I | T | I | T | I |
---|
CNN | 530 | 59 | 2570 | 120 | 2610 | 190 | 8650 | 478 | 1540 | 109 |
LSTM | 540 | 87 | 6790 | 255 | 10,300 | 556 | 6540 | 893 | 2410 | 232 |
DNN | 310 | 22 | 1500 | 33 | 3070 | 95 | 3760 | 76 | 1510 | 31 |
RF | 9 × | 15 | 5710 | 5.5 | 20 × | 24 | 90 | 20 | 2170 | 3.1 |
KNN | 0 | 6 × | 0 | 967 | 0 | 42 × | 0 | 97 | 0 | 854 |
Table 20.
Temperature prediction results for stator winding and magnet temperature in terms of MSE (K²) for different testing IDs and models. Baseline scenarios are denoted with ‘Base’, while reduced-order configurations are denoted with ’MOR’.
Table 20.
Temperature prediction results for stator winding and magnet temperature in terms of MSE (K²) for different testing IDs and models. Baseline scenarios are denoted with ‘Base’, while reduced-order configurations are denoted with ’MOR’.
ID | Time (h) | Stator Winding | Rotor Magnet |
---|
Base | MOR | Base | MOR |
---|
60 | 1.7 | 2.41 | 1.34 | 22.62 | 16.68 |
62 | 3.3 | 2.75 | 1.79 | 17.49 | 31.11 |
74 | 3.0 | 3.33 | 2.37 | 14.47 | 15.39 |
Avg | 8.0 | 2.90 | 1.91 | 17.45 | 22.15 |