Detection of Electricity Theft Behavior Based on Improved Synthetic Minority Oversampling Technique and Random Forest Classifier
Abstract
:1. Introduction
- (1)
- Considering the high unbalance of power user-side dataset and the shortcomings of existing methods, a new K-SMOTE method was proposed to deal with the unbalanced initial datasets. The proposed method can reduce the impact of detection accuracy caused by unbalanced data.
- (2)
- Combined with the unbalanced data, considering the limitation of setting decision tree in RF algorithm, the improved random forest classifier was applied to detect electricity theft behaviors. The efficiency of power theft detection could be greatly improved because multiple decision trees were running in parallel. Then, the improved RF algorithm and K-SMOTE oversampling algorithm were combined to establish the electricity theft detection system, which considered the unbalance of the users’ electricity dataset.
- (3)
- The detection method of this paper had higher detection accuracy and better stability compared with the existing methods.
2. Proposed Algorithm
2.1. SMOTE
- (1)
- For a sample xi in minority class samples set X, calculate the Euclidean distance from this sample to all other samples in the set, and get its k nearest neighbor, denoted as yj (j = 1, 2, …, k).
- (2)
- Sampling rate is set according to the data unbalance ratio to determine the sampling magnification. For data xi, n numbers are randomly selected from their K-nearest neighbors, and new data can be constructed as follows:
2.2. K-Means Clustering Algorithm
- (1)
- Dataset D, denoted as {x1, x2, x3, …, xn}, randomly select k initial cluster center as μ1, μ2, …, μk ∈ Dn.
- (2)
- Calculate the Euclidean distance using Equation (2), that is, calculate the distance d(xi, μj) between xi to each cluster center, find the minimum d, and divide xi into the same cluster as μi:
- (3)
- After all data have been calculated, the new clustering center of each class can be recalculated by Equation (3):
- (4)
- If the cluster to which each data belongs does not change with the increase of iteration process, go to Step 2. Otherwise, go to Step 5.
- (5)
- Output clustering results.
2.3. K-SMOTE
- (1)
- Let M, O, P, and N represent the unbalanced electricity dataset on the user side, the PD, and the ND. T is the training set in P, S is the majority training set, and T and S constitute the total training set O. K is the number of the initial clusters, ui is the cluster center, and Xnew is the corresponding new interpolated data point set.
- (2)
- Determine the number of initial clusters K.
- (3)
- For T, K-means algorithm was used to perform clustering and record cluster centers. T was divided into K clusters, and the cluster center were {μ1, μ2, μ3, …, μk}.
- (4)
- SMOTE was used in T to achieve data interpolation based on cluster centers {μ1, μ2, μ3, …, μk}, then the interpolated dataset Xnew was obtained
- (5)
- T, S, and Xnew were combined to form new training set O’.
3. Random Forest Classification Based on K-SMOTE
- (1)
- Compared with existing classification algorithms, its average accuracy is at a high level [49].
- (2)
- It can process input data with high dimensional characteristics without dimensionality reduction [50].
- (3)
- An unbiased estimate of the internal generation error can be obtained during the generation process.
- (4)
- It is robust to default value problems.
- (5)
- Each decision tree in the random forest operates independently, realizing parallel operation of multiple decision trees and saving resources and computational time.
- (6)
- Randomness is reflected in the random selection of data and attributes, even if each tree is not pruned, there will be no overfitting.
3.1. Decision Tree
3.2. Discretization of Continuous Variable
- (1)
- Sort the values of continuous variables to find the maximum (MAX) and minimum (MIN).
- (2)
- If there are N values for continuous variables and each value is a breakpoint, the interval (MIN, MAX) is divided into N-1 intervals.
- (3)
- For each breakpoint Ai (i = 1, 2, …, N), Gini-index is calculated with A and B as intervals.
- (4)
- Select a breakpoint Ai with the largest Gini-index coefficient as the best split point of the continuous attribute.
3.3. Random Forest
3.3.1. Bootstrap Random Sampling
3.3.2. OOB Error Estimate
3.3.3. Random Forest
- (1)
- The unbalanced user-side dataset M is oversampled by K-SMOTE to obtain dataset M’.
- (2)
- Divide the training set Tr and test set Te of random forest.
- (3)
- Set the number of initial decision tree nTree.
- (4)
- Use the bootstrap method to select training data for each decision tree. The total features in M are K. Selecting n features randomly, n is calculated using Equation (15). Then, use the CART algorithm to generate the unpruned decision trees.
- (5)
- Input test set Te into each trained decision trees, and the classification result is determined according to the voting result of each decision tree. The voting classification formula is as follows:
- (6)
- The current OOB error is calculated according to Equation (9). If the OOB error converges, then go to Step 7. If the error does not converge, update the decision tree number nTree according to Equation (11) and return to Step 4.
- (7)
- Output classification result.
4. Simulation Results
- (1)
- The neurons’ number of input layer, hidden layer, and output layer in BPN were 20, 40 and 1, respectively. Learning rate was 0.3, momentum term was 0.3, batch volume was 100, and the maximum number of iterations was 50 [53].
- (2)
- The kernel function of SVM qA radial basis function (RBF), the parameter coefficient of kernel function penalty g WAS 0.07, and the penalty factor coefficient c of PD and ND were 1 and 0.01, respectively [30].
4.1. Evaluation Indexes
- (1)
- Accuracy (ACC): ACC is the ratio of the number of correct classifications to the total number of samples. The higher the value of ACC, the better is the performance of the detection algorithm. Mathematically ACC is defined as:
- (2)
- True Positive Rate (TPR): TPR describes the sensitivity of the detection model to PD. The higher the value of TPR, the better is the performance of the detection algorithm. TPR is defined as:
- (3)
- False Positive Rate (FPR): FPR refers to the proportion of data in ND, which actually belongs to ND, and is wrongly judged as PD by the detection algorithm. FPR is defined as:
- (4)
- True Negative Rate (TNR): TNR describes the sensitivity of the detection model to ND, which is defined as:
- (5)
- G-mean index: G-mean index is used for the evaluation of classifier performance [54]. Large G-mean index reveals better classification performance. The value of G-mean depends on the square root of the product of the accuracy of PD and ND. G-mean can reasonably evaluate the overall classification performance of unbalanced dataset, and it can be expressed as:
- (6)
- Receiver operating characteristic (ROC) and area under the ROC curve (AUC): Receiver operator characteristic chive (ROC) was originally created to test the performance of a radar [55]. ROC curve describes the relationship between the relative growth of FPR and TPR in the confusion matrix. For values output by the binary classification model, the closer the ROC curve is to the point (0, 1), the better the classification performance. Area under the ROC curve (AUC), is an index to evaluate the performance of the detection algorithm in the ROC curve. The AUC value of 1 corresponds to an ideal detection algorithm.
4.2. Unbalanced Processing of User-Side Data
4.3. Electricity Theft Detection Based on Improved RF
4.3.1. Determination of the Number of Decision Trees
4.3.2. Detection Results of RF
4.3.3. Comparison of Detection Performance of Different Algorithms
- (1)
- Without K-SMOTE, the ACC value and AUC of RF detection method were relatively low. However, with K-SMOTE, the ACC and AUC value of three detection methods were obviously improved, which was increased about 10%. This indicates that unbalanced datasets would affect the accuracy of detection algorithm, and K-SMOTE plays an effective role in improving machine learning accuracy.
- (2)
- The electricity user data processed by K-SMOTE were tested by BPN and SVM. The ACC mean values of SVM and BPN were 71.26% and 84.87%, respectively, and the mean values of AUC in SVM and BPN were 0.7236 and 0.8716, respectively. These indexes were lower than the ACC and AUC of RF, which were 94.53% and 0.9513, respectively. Thus, the performance of RF was superior to SVM and BPN.
5. Conclusions
- (1)
- K-SMOTE was proposed to avoid the influence of unbalanced data on the accuracy of the classifier.
- (2)
- The RF classifier, which was suitable for the nature of the user-side dataset, was used to detect electricity theft. The decision trees in RF classifier could work in parallel, which improved the detection efficiency and reduced the computational time.
- (3)
- Compared with the conventional detection methods, the proposed method featured higher accuracy and stronger stability.
Author Contributions
Funding
Conflicts of Interest
Appendix A
Acronyms | Full Name |
---|---|
SMOTE | synthetic minority oversampling technique |
RF | random forest |
PD | positive data |
ND | negative data |
OOB Error | out-of-bag data error |
TL | Technical loss |
NTL | non-technical losses |
PCA | principal component analysis |
FA | factor analysis |
ELM | extreme learning machine |
BC | balance cascade |
KNN | k-nearest neighbor |
TN | true negative |
TP | true positive |
FP | false positive |
FN | false negative |
ACC | Accuracy |
TPR | true positive rate |
FPR | false positive rate |
TNR | true negative rate |
IE | information entropy |
OOB | out of bag |
SVM | support vector machines |
BPN | back-propagation neural network |
References
- Lo, Y.L.; Huang, S.C.; Lu, C.N. Non-technical loss detection using smart distribution network measurement data. In Proceedings of the IEEE PES Innovative Smart Grid Technologies, Tianjin, China, 21–24 May 2012; pp. 1–5. [Google Scholar]
- Smith, T.B. Electricity theft: A comparative analysis. Energy Policy 2004, 32, 2067–2076. [Google Scholar] [CrossRef]
- Huang, S.C.; Lo, Y.L.C.N. Non-Technical Loss Detection Using State Estimation and Analysis of Variance. IEEE Trans. Power Syst. 2013, 28, 2959–2966. [Google Scholar] [CrossRef]
- Bhavna, B.; Mohinder, G. Reforming the Power Sector, Controlling Electricity Theft and Improving Revenue. Public Policy for the Private Sector. 2004. Available online: http://rru.worldbank.org/PublicPolicyJourna (accessed on 16 December 2019).
- Soma, S.S.R.D.; Wang, L.; Vijay, D.; Robert, C.G. High performance computing for detection of electricity theft. Int. J. Electr. Power Energy Syst. 2013, 47, 21–30. [Google Scholar]
- Smart Meters Help Reduce Electricity Theft, Increase Safety. BCHydro. Available online: https://www.bchydro.com/news/conservation/2011/smart_meters_energy_theft.html (accessed on 16 December 2019).
- Dzung, D.; Naedele, M.; Von Hoff, T.P.; Crevatin, M. Security for Industrial Communication Systems. Proc. IEEE Secur. Ind. Commun. Syst. 2005, 93, 1152–1177. [Google Scholar] [CrossRef]
- Krebs, B. FBI: Smart Meter Hacks Likely to Spread. Available online: http://krebsonsecurity.com/2012/04/fbi-smart-meter-hacks-likely-to-spread (accessed on 16 December 2019).
- Carlos, L.; Félix, B.; Iñigo, M.; Juan, I.; Guerrero, J.B.; Rocío, M. Integrated expert system applied to the analysis of non-technical losses in power utilities. Expert Syst. Appl. 2011, 38, 10274–10285. [Google Scholar]
- Han, W.; Xiao, Y. NFD: A practical scheme to detect non-technical loss fraud in smart grid. In Proceedings of 2014 IEEE International Conference on Communications (ICC), Sydney, NSW, Australia, 10–14 June 2014; pp. 605–609. [Google Scholar]
- Grochocki, D.; Huh, J.H.; Berthier, R. AMI threats, intrusion detection requirements and deployment recommendations. In Proceedings of the IEEE Third International Conference on Smart Grid Communications, Tainan, Taiwan, China, 5–8 November 2012; pp. 395–400. [Google Scholar]
- Hao, R.; Ai, Q.; Xiao, F. Architecture based on multivariate big data platform for analyzing electricity consumption behavior. Electr. Power Autom. Equip. 2017, 37, 20–27. [Google Scholar]
- Jiang, R.; Lu, R.; Wang, Y.; Luo, J.; Shen, C.; Shen, X. Energy-theft detection issues for advanced metering infrastructure in smart grid. Tsinghua Sci. Technol. 2014, 19, 105–120. [Google Scholar] [CrossRef]
- Lo, C.-H.; Ansari, N. CONSUMER: A Novel Hybrid Intrusion Detection System for Distribution Networks in Smart Grid. IEEE Trans. Emerg. Top. Comput. 2013, 1, 33–44. [Google Scholar] [CrossRef]
- McLaughlin, S.; Holbert, B.; Fawaz, A.; Berthier, R.; Zonouz, S. A multi-sensor energy theft detection framework for advanced metering infrastructures. IEEE J. Sel. Areas Commun. 2013, 31, 1319–1330. [Google Scholar] [CrossRef]
- Leite, J.B.; Sanches mantovani, J.R. Detecting and locating non–technical losses in modern distribution networks. IEEE Trans. Smart Grid 2018, 9, 1023–1032. [Google Scholar] [CrossRef] [Green Version]
- Tariq, M.; Poor, H. Electricity Theft Detection and Localization in Grid-Tied Microgrids. IEEE Trans. Smart Grid 2018, 9, 1920–1929. [Google Scholar] [CrossRef]
- Salinas, S.; Li, P. Privacy-Preserving Energy Theft Detection in Microgrids: A State Estimation Approach. IEEE Trans. Power Syst. 2016, 31, 883–894. [Google Scholar] [CrossRef]
- Cárdenas, A.A.; Amin, S.; Schwartz, G.; Dong, R.; Sastry, S. A game theory model for electricity theft detection and privacy-aware control in AMI systems. In Proceedings of the 50th Annual Allerton Conference on Communication, Control, and Computing (Allerton), Monticello, IL, USA, 1–5 October 2012; pp. 1830–1837. [Google Scholar]
- O’Leary, D.E. Summary of Previous Papers in Expert Systems Review. Intell. Syst. Account. Financ. Manag. 2016, 1, 3–7. [Google Scholar] [CrossRef]
- Punmiya, R.; Choe, S. Energy Theft Detection Using Gradient Boosting Theft Detector with Feature Engineering-Based Preprocessing. IEEE Trans. Smart Grid 2019, 10, 2326–2329. [Google Scholar] [CrossRef]
- Coma-Puig, B.; Carmona, J. Bridging the Gap between Energy Consumption and Distribution through Non-Technical Loss Detection. Energies 2019, 12, 1748. [Google Scholar] [CrossRef] [Green Version]
- Lu, X.; Zhou, Y.; Wang, Z.; Yi, Y.; Feng, L.; Wang, F. Knowledge Embedded Semi-Supervised Deep Learning for Detecting Non-Technical Losses in the Smart Grid. Energies 2019, 12, 3452. [Google Scholar] [CrossRef] [Green Version]
- Nagi, J.; Yap, K.S.; Tiong, S.K.; Ahmed, S.K.; Mohamad, M. Nontechnical Loss Detection for Metered Customers in Power Utility Using Support Vector Machines. IEEE Trans. Power Deliv. 2009, 25, 1162–1171. [Google Scholar] [CrossRef]
- Zhuang, C.J.; Zhang, B.; Hu, J.; Li, Q.; Zeng, R. Anomaly detection for power consumption patterns based on unsupervised learning. Proc. CSEE 2016, 36, 379–387. (In Chinese) [Google Scholar]
- Ghasemi, A.A.; Gitizadeh, M. Detection of illegal consumers using pattern classification approach combined with Levenberg-Marquardt method in smart grid. Int. J. Electr. Power Energy Syst. 2018, 99, 363–375. [Google Scholar] [CrossRef]
- Monedero, I.; Biscarri, F.; Leon, C.; Biscarri, J.; Millan, R. MIDAS: Detection of Non-technical Losses in Electrical Consumption Using Neural Networks and Statistical Techniques. In Proceedings of Computational Science and Its Applications-ICCSA 2006, International Conference, Glasgow, UK, 8–11 May 2006; Proceedings, Part V. Springer: Berlin/Heidelberg, Germany, 8 May 2006; pp. 725–734. [Google Scholar]
- Nizar, A.H.; Dong, Z.Y.; Wang, Y. Power Utility Nontechnical Loss Analysis with Extreme Learning Machine Method. IEEE Trans. Power Syst. 2008, 23, 946–955. [Google Scholar] [CrossRef]
- Muniz, C.; Figueiredo, K.; Vellasco, M.; Chavez, G.; Pacheco, M. Irregularity detection on low tension electric installations by neural network ensembles. In Proceedings of the International Joint Conference on Neural Networks, Atlanta, GA, USA, 14–19 June 2009; pp. 2176–2182. [Google Scholar]
- Nagi, G.; Yap, K.S.; Tiong, S.K.; Ahmed, S.K.; Nagi, F. Improving SVM-Based Nontechnical Loss Detection in Power Utility Using the Fuzzy Inference System. IEEE Trans. Power Deliv. 2011, 26, 1284–1285. [Google Scholar] [CrossRef]
- Viegas, J.L.; Esteves, P.R.; Vieira, S.M. Clustering-based novelty detection for identification of non-technical losses. Int. J. Electr. Power Energy Syst. 2018, 101, 301–310. [Google Scholar] [CrossRef]
- Buzau, M.M.; Tejedor-Aguilera, J.; Cruz-Romero, P.; Gómez-Expósito, A. Detection of Non-Technical Losses Using Smart Meter Data and Supervised Learning. IEEE Trans. Smart Grid 2019, 10, 2661–2670. [Google Scholar] [CrossRef]
- Messinis, G.M.; Hatziargyriou, N.D. Review of non-technical loss detection methods. Electr. Power Syst. Res. 2018, 158, 250–266. [Google Scholar] [CrossRef]
- Jokar, P.; Arianpoo, N.; Leung, V.C.M. Electricity Theft Detection in AMI Using Customers’ Consumption Patterns. IEEE Trans. Smart Grid 2015, 7, 216–226. [Google Scholar] [CrossRef]
- Figueroa, G.; Chen, Y.; Avila, N.; Chu, C. Improved practices in machine learning algorithms for NTL detection with imbalanced data. In Proceedings of the 2017 IEEE Power & Energy Society General Meeting, Chicago, IL, USA, 16–20 July 2017; pp. 1–5. [Google Scholar]
- Glauner, P.; Boechat, A.; Lautaro, D.; Radu, S.; Franck, B.; Yves, R.; Diogo, D. Large-scale detection of non-technical losses in unbalanced datasets. In Proceedings of the 2016 IEEE Power & Energy Society Innovative Smart Grid Technologies Conference (ISGT), Minneapolis, MN, USA, 6–9 September 2016; pp. 1–5. [Google Scholar]
- Liu, X.; Wu, J.; Zhou, Z. Exploratory Under sampling for Class-Unbalance Learning. IEEE Trans. Cybern. 2009, 39, 539–550. [Google Scholar]
- Victoria, L.; Alberto, F.; Salvador, G.; Palade, V.; Herrera, F. An insight into classification with unbalanced data: Empirical results and current trends on using data intrinsic characteristics. Inf. Sci. 2013, 250, 113–141. [Google Scholar]
- Del Río, S.; López, V.; Benítez, J.M.; Herrera, F. On the use of MapReduce for unbalanced big data using Random Forest. Inf. Sci. 2014, 285, 112–137. [Google Scholar] [CrossRef]
- Chawla, N.V.; Bowyer, K.W.; Hall, L.O.; Kegelmeyer, W.P. SMOTE: Synthetic minority over-sampling technique. J. Artif. Intell. Res. 2002, 16, 321–357. [Google Scholar] [CrossRef]
- Feilong, C.; Yuanpeng, T.; Miaomiaom, C. Sparse algorithms of Random Weight Networks and applications. Expert Syst. Appl. 2014, 41, 2457–2462. [Google Scholar]
- Song, X.; Guo, Z.; Guo, H.; Wu, S.; Wu, C. A new forecasting model based on forest for photovoltaic power generation. Power Syst. Prot. Control 2015, 43, 13–18. [Google Scholar]
- Monedero, I.; Félix, B.; Carlos, L.; Guerrero, J.I.; Jesús, B.; Rocío, M. Detection of frauds and other non-technical losses in a power utility using Pearson coefficient, Bayesian networks and decision trees. Int. J. Electr. Power Energy Syst. 2012, 34, 90–98. [Google Scholar] [CrossRef]
- Gao, D.; Zhang, Y.X.; Zhao, Y.H. Random forest algorithm for classification of multiwavelength data. Res. Astron. Astrophys. 2009, 9, 220–226. [Google Scholar] [CrossRef]
- Fernandez, A.; García, S.; Herrera, F.; Chawla, N.V. SMOTE for Learning from Imbalanced Data: Progress and Challenges, Marking the 15-year Anniversary. J. Artif. Intell. Res. 2018, 61, 863–905. [Google Scholar] [CrossRef]
- Bradley, R.; Fayyad, U. Refining Initial Points for K-Means Clustering. In Proceedings of the Fifteenth International Conference on Machine Learning; 1998; pp. 91–99. Available online: https://dl.acm.org/doi/10.5555/645527.657466 (accessed on 20 December 2019).
- Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
- Breiman, L. Bagging predictors. Mach. Learn. 1996, 24, 123–140. [Google Scholar] [CrossRef] [Green Version]
- Belgiu, M.; Drăguţ, L. Random forest in remote sensing: A review of applications and future directions. ISPRS J. Photogramm. Remote Sens. 2016, 114, 24–31. [Google Scholar] [CrossRef]
- Alessia, S.; Antonio, C.; Aldo, Q. Random Forest Algorithm for the Classification of Neuroimaging Data in Alzheimer’s Disease: A Systematic Review. Front. Aging Neurosci. 2017, 9, 329–336. [Google Scholar]
- Quinlan, J.R. Decision trees and decision-making. IEEE Trans. Syst. Man Cybern. 1990, 20, 339–346. [Google Scholar] [CrossRef]
- Wolpert, D.H.; Macready, W.G. An Efficient Method to Estimate Bagging’s Generalization Error. Mach. Learn. 1999, 35, 41–55. [Google Scholar] [CrossRef] [Green Version]
- Xu, G.; Tan, Y.P.; Dai, T.H. Sparse Random Forest Based Abnormal Behavior Pattern Detection of Electric Power User Side. Power Syst. Technol. 2017, 41, 1965–1973. [Google Scholar]
- Rao, R.B.; Krishnan, S.; Niculescu, R.S. Data mining for improved cardiac care. ACM SIGKDD Explor. Newsl. 2006, 1, 3–10. [Google Scholar] [CrossRef]
- Huang, Y.A.; You, Z.H.; Gao, X.; Wong, L.; Wang, L. Using Weighted Sparse Representation Model 561 Combined with Discrete Cosine Transformation to Predict Protein-Protein Interactions from Protein Sequence. BioMed Res. Int. 2015, 2015, 902198. [Google Scholar] [CrossRef] [PubMed] [Green Version]
Decision Tree | T1 | T2 | T3 | … | Tn | |
---|---|---|---|---|---|---|
Data | ||||||
(X1,Y1) | N | N | N | … | Y | |
(X2,Y2) | N | N | Y | … | Y | |
(X3,Y3) | N | N | Y | … | N | |
… | … | … | … | … | … | |
(X4,Y4) | Y | Y | N | … | N |
Predicted | Predicted Positive | Predicted Negative | |
---|---|---|---|
Actual | |||
True Positive | TP | FN | |
True Negative | FP | TN |
Algorithm | Random Oversampling | SMOTE | K-SMOTE |
---|---|---|---|
Data Repetition Rate/% | 95.02 | 30.50 | 15.84 |
Simulation Times | K-SMOTE | ACC/% |
---|---|---|
1 | with | 96.02 |
without | 83.24 | |
2 | with | 96.21 |
without | 82.01 | |
3 | with | 81.32 |
without | 91.35 | |
Mean | with | 94.53 |
without | 85.53 |
Simulation Times | Algorithm | ACC/% |
---|---|---|
1 | SVM | 70.21 |
BPN | 85.24 | |
2 | SVM | 72.54 |
BPN | 83.01 | |
3 | SVM | 71.02 |
BPN | 86.35 | |
Mean | SVM | 71.26 |
BPN | 84.87 |
© 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
Share and Cite
Qu, Z.; Li, H.; Wang, Y.; Zhang, J.; Abu-Siada, A.; Yao, Y. Detection of Electricity Theft Behavior Based on Improved Synthetic Minority Oversampling Technique and Random Forest Classifier. Energies 2020, 13, 2039. https://doi.org/10.3390/en13082039
Qu Z, Li H, Wang Y, Zhang J, Abu-Siada A, Yao Y. Detection of Electricity Theft Behavior Based on Improved Synthetic Minority Oversampling Technique and Random Forest Classifier. Energies. 2020; 13(8):2039. https://doi.org/10.3390/en13082039
Chicago/Turabian StyleQu, Zhengwei, Hongwen Li, Yunjing Wang, Jiaxi Zhang, Ahmed Abu-Siada, and Yunxiao Yao. 2020. "Detection of Electricity Theft Behavior Based on Improved Synthetic Minority Oversampling Technique and Random Forest Classifier" Energies 13, no. 8: 2039. https://doi.org/10.3390/en13082039
APA StyleQu, Z., Li, H., Wang, Y., Zhang, J., Abu-Siada, A., & Yao, Y. (2020). Detection of Electricity Theft Behavior Based on Improved Synthetic Minority Oversampling Technique and Random Forest Classifier. Energies, 13(8), 2039. https://doi.org/10.3390/en13082039