Next Article in Journal
Numerical Simulation Study on Seepage Theory of a Multi-Section Fractured Horizontal Well in Shale Gas Reservoirs Based on Multi-Scale Flow Mechanisms
Previous Article in Journal
Optimisation of Tray Drier Microalgae Dewatering Techniques Using Response Surface Methodology
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Distribution Grids Fault Location employing ST based Optimized Machine Learning Approach

1
Electrical Engineering Department, King Fahd University of Petroleum & Minerals, Dhahran 31261, Saudi Arabia
2
Electrical Engineering Department, Menoufia University, Shebin El-Kom 32511, Egypt
*
Author to whom correspondence should be addressed.
Energies 2018, 11(9), 2328; https://doi.org/10.3390/en11092328
Submission received: 11 July 2018 / Revised: 12 August 2018 / Accepted: 14 August 2018 / Published: 4 September 2018
(This article belongs to the Section F: Electrical Engineering)

Abstract

:
Precise information of fault location plays a vital role in expediting the restoration process, after being subjected to any kind of fault in power distribution grids. This paper proposed the Stockwell transform (ST) based optimized machine learning approach, to locate the faults and to identify the faulty sections in the distribution grids. This research employed the ST to extract useful features from the recorded three-phase current signals and fetches them as inputs to different machine learning tools (MLT), including the multilayer perceptron neural networks (MLP-NN), support vector machines (SVM), and extreme learning machines (ELM). The proposed approach employed the constriction-factor particle swarm optimization (CF-PSO) technique, to optimize the parameters of the SVM and ELM for their better generalization performance. Hence, it compared the obtained results of the test datasets in terms of the selected statistical performance indices, including the root mean squared error (RMSE), mean absolute percentage error (MAPE), percent bias (PBIAS), RMSE-observations to standard deviation ratio (RSR), coefficient of determination (R2), Willmott’s index of agreement (WIA), and Nash–Sutcliffe model efficiency coefficient (NSEC) to confirm the effectiveness of the developed fault location scheme. The satisfactory values of the statistical performance indices, indicated the superiority of the optimized machine learning tools over the non-optimized tools in locating faults. In addition, this research confirmed the efficacy of the faulty section identification scheme based on overall accuracy. Furthermore, the presented results validated the robustness of the developed approach against the measurement noise and uncertainties associated with pre-fault loading condition, fault resistance, and inception angle.

1. Introduction

The detailed information regarding the faulty area and fault location, plays a vital role in expediting the restoration process in electric utilities after being subjected to any kind of fault. Consequently, there is a growing research interest to locate faults and identify faulty sections in distribution grids efficiently, to reduce customer minute losses and revenue losses for the utilities. The available fault location techniques for distribution grids can be categorized into three major groups namely the impedance, the traveling wave, and the knowledge-based techniques [1,2,3]. The impedance-based technique, evaluates fault location using voltage and current measurements available at the substation end, as well as technical information of the distribution grids, including grid topology, load, and line data. This approach analyzes the fault location either in phase domain [4,5,6] or using symmetrical components [7,8,9]. However, the impedance-based fault location approach, frequently leads towards multiple estimations and they are generally based on a hectic iterative process. Additionally, the dynamic behavior of loads, the presence of laterals, inherent properties of distribution grids, measurement noise, and fault resistance and inception angle uncertainties, heavily degrade the performance of these techniques [1,2,3].
However, the development of advanced measurement infrastructure, communication systems, and the willingness of the electric system decision makers paved the way for a faster and more accurate fault location technique, called the traveling wave technique. The technique is based on the characteristic frequencies of the traveling waves associated with faults in the electrical grids [10], and can be ramified into A, B, and C-type [11]. The A-type and the B-type fault location techniques, are based on online measurements and use single/double ended recorded traveling waves originated by faults, whereas the C-type technique injects traveling waves manually to locate faults in electricity grids [12]. These techniques are very useful for long transmission/distribution lines, but distribution lines are short by nature. Consequently, they cannot effectively locate faults in short length distribution lines. The third category fault location technique namely, the knowledge-based approach, offers auspicious prospects in dealing with distribution grid faults considering their intrinsic complexities. In Reference [13], the artificial neural network (ANN) was employed to locate distribution grid faults based on voltage, current, and active power measurements collected from the substation end. The authors of Reference [14,15,16,17], combined the wavelet transform and different machine learning tools to locate faults in distribution grids. However, the wavelet transform (WT) based decompositions, do not uphold the phase information of the original signal and are sometimes sensitive to the presence of measurement noise [18,19,20].
However, most of the referenced fault location techniques did not consider the presence of measurement noise and uncertainties associated with pre-fault loading condition, fault resistance, and inception angle in locating faults and identifying faulty sections in power distribution grids. Additionally, an advanced signal processing technique (SPT) namely, the Stockwell transform that combines the advantages of wavelet transform and short time Fourier transform was employed recently to detect and classify faults in distribution grids in Reference [21], but not employed in locating faults and identifying faulty sections in distribution grids. Therefore, this research proposes the hybrid fault location and faulty section identification approach, for power distribution grids, combining the Stockwell transform and different machine learning tools. The proposed approach starts with the extraction of useful features from the ST decomposed faulty current signals, collected from different branches of the grids. Then it fetches the extracted characteristic features, as inputs of the different machine learning tools, including the MLP-NN, SVM, and ELM to get decisions on fault location and faulty section. Additionally, the proposed technique employs the CF-PSO to optimize the SVM and ELM key parameters, to achieve better accuracy in locating distribution grid faults. The presented results, demonstrated the independence of the proposed approach in the presence of measurement noise, pre-fault loading condition, fault resistance, and inception angle.

2. Background Theories

The combination of SPT and MLT, received significant attention in diagnosing power systems transients in recent years [22,23,24], where the researchers used SPT to collect characteristic features related to the electrical transients, and fetched them as inputs to MLT. The following parts of this section, briefly describe the employed SPT and MLT in this research.

2.1. Stockwell Transform

Among many SPT, the Fourier transform (FT), is one of the most widely used techniques to transform time domain signals into the frequency domain to analyze harmonics and design filters. However, the FT loses the temporal information and provides erroneous results for non-stationary signals. In response to this deficiency, Dennis Gabor (1946) employed a small sampling window of the regular interval to map a signal into a two-dimensional function of time and frequency, and the adaptation is known as short time Fourier transform (STFT) [25]. Unfortunately, the STFT cannot provide better time-frequency resolution simultaneously, where a better frequency resolution leads towards a worse time resolution, and vice versa. The wavelet transform (WT), resolves the resolution problem by employing larger windows at lower frequencies and smaller windows at higher frequencies [26]. However, the WT is sensitive to the presence of measurement noise and does not uphold the phase information of the non-stationary signals. Consequently, Stockwell et al. [18,19,20] developed the ST combining the benefits of the STFT and WT, that effectively upholds the referenced frequency and phase information. The ST of a signal x ( t ) can be defined as:
S ( τ , f ) = S { x ( t ) } = x ( t )   w ( τ t , f )   e j 2 π f t d t  
where, w ( t , f ) = | f | 2 π e t 2 f 2 2 is the Gaussian window function, f represents the frequency, and t and τ are the time variables.
The ST transforms the time domain signals, into both the frequency and time domains. The coefficients of this transform produce complete resolutions for each designated frequency and cover the whole temporal axis by picking up the possible values of τ where the τ indicates the center of the window function. Different values of f are employed to adjust the sizes of the Gaussian windows, with a view to realizing multiresolution over different frequencies over the temporal axis. However, the ST on a signal provides an M × N matrix, commonly known as the S-matrix, where all the elements are complex numbers. The rows of the matrix pertain to frequency characteristics and the columns pertain to time characteristics. However, the following equation gives the energy matrix of the same signal:
E = | S M × N | 2
The proposed fault location approach, generates three S-matrices and three E-matrices, from a three-phase current signal through ST. Hence, it creates new vectors namely Scmax, Srmax, and Emax from the S-matrices and the E-matrices. The Scmax-vectors contain the maximum absolute values of the columns of the S-matrices, whereas the Srmax-vectors contain the maximum absolute values of the rows of the S-matrices, and the Ecmax-vectors contain the absolute maximum values of the columns of the energy matrices. The approach also produces another set of vectors called Sc-phase-max, that contain the phase angles of the S-matrices associated with the elements of Scmax. Eventually, the proposed approach applies standard statistical techniques on the produced vectors and extracts characteristic features that contain fault signature. Among many features, the amplitude, the gradient of amplitude, mean value, standard deviation, entropy, skewness, kurtosis, time of occurrence, and energy of different harmonics are widely used features for the analysis of power quality transients [21,27,28,29].

2.2. Multilayer Perceptron Neural Network

Artificial neural networks have become very popular for many engineering applications, due to their parallel computing abilities and adaptiveness to external disturbances. The MLP-NN is one of the widely used ANN and consists of input, hidden, and output layers. The hidden layer uses squashing functions to process the inputs before sending them to the output layer. In order to minimize training error, the supervised learning algorithm optimizes randomly initiated connecting weights and biases [30,31,32]. In this paper, the total number of hidden neurons for the MLP-NN are chosen through a systematic trial and error approach.

2.3. Support Vector Machine

The SVM was first introduced by Boser et al. [33] with a view to efficient analysis of data, and was primarily restricted to classification problems only. With the passage of time, the SVM was extended to solve regression problems that construct an optimal geometric hyperplane to distinguish the available data, and to map them into the higher dimensional feature space [34]. It forms a separation surface employing various functions including sigmoidal, polynomial, and radial basis functions. This paper employed a MATLAB/SIMULINK based SVM-KM toolbox [35], to locate distribution grid faults. It considered the mean absolute percentage error (MAPE) of the test dataset as the objective function, while optimizing SVM parameters including regularization coefficients (λ and C), tolerance of termination criterion (ε), and kernel option (Ko) employing CF-PSO.

2.4. Extreme Learning Machine

Huang et al. [36] introduced the extreme learning machines for single hidden layer feedforward neural networks, that are a thousand-fold faster over the traditional feedforward neural networks, in attaining generalization performance efficiently. The ELM evaluates the output weights analytically, by picking the input weights randomly [37]. This paper employs an ELM toolbox [38], developed in the MATLAB/SIMULINK platform to locate distribution grid faults. Like the SVM, this research considers the MAPE of the test dataset as the objective function, while optimizing ELM parameters including the regularization coefficient (CR) and the kernel option (Kp) with the aid of CF-PSO.

2.5. Constriction-Factor Partcile Swarm Optimization

The population-based stochastic optimization approach widely known as the particle swarm optimization (PSO) mimics the social behavior of swarms [39]. Initially, it was considered that the inertia weight of the PSO technique gets decreased monotonously in each generation to update the velocity, hence to update the position of the swarms. This monotonicity sometimes leads towards fall of the quality of the solutions. Consequently, Clerc [40] introduced the idea of constriction factor to update the velocity by ensuring system convergence and effective diversified search. The following steps briefly discuss the CF-PSO technique:
Step 1: Initialization
The CF-PSO employs the following equations to initialize positions and velocities of the swarms randomly from the search space, for a pre-specified number of particles:
x i , j = U ( x m i n , j ,   x m a x j ) ,                   i ,   j
v i , j = U ( v m i n , j ,   v m a x j ) ,                   i ,   j
where, U represents the uniform distribution; i and j represent particle size and dimension of the problem; xmin and xmax refer to the lower and upper boundaries of the solution space; vmin and vmax represent the lower and upper boundaries of the velocities of the particles.
Step 2: Fitness assessment and updating the best solution
This step assesses the fitness of swarms, based on their randomly initiated positions. Hence, it stores the fitness values of all swarms as the individual best solutions for the first generation. From the second generation, this step compares the new fitness of the swarms with the fitness of the previous generation and updates the individual best solutions accordingly. Besides this, the step picks the global best solution from the individual best solutions in each generation. Finally, the CF-PSO chooses the optimal solution from the global best solutions.
Step 3: Checking the stopping criteria
The CF-PSO checks the stopping criteria after a certain number of generations, to avoid premature convergence. It stops the optimization algorithm if the objective function does not change for a pre-specified number of generations, or it reaches the targeted number of generations.
Step 4: Updating inertia weights, velocities, and positions
This step upgrades the inertia weights and the velocity of each swarm, with the aid of the following equations in each generation. Hence, this step upgrades the positions of the swarms based on the revised velocities.
w t = w m a x w m a x w m i n t m a x × t
v j , k t = c f ( w t × v j , k t 1 + c 1 r 1 t × [ x j , k * ( t 1 ) x j , k ( t 1 ) ] + c 2 r 2 t × [ x j , k * * ( t 1 ) x j , k ( t 1 ) ] )
c f = 2 | 2 φ ( φ 2 4 φ ) | ;   c 1 + c 2 = φ ;   | c 1 | | c 2 | ;   φ > 4
  x j , k t = v j , k t + x j , k ( t 1 )  
where, w t is the inertia weight of tth generation; v j , k t is the velocity of jth particle in tth generation; v j ,   k ( t 1 ) and x j , k ( t 1 ) are the velocity and the position of jth particle in (t − 1)th generation x j , k * * ( t 1 ) ; and x j , k * ( t 1 ) are the global and individual best positions of (t − 1)th generation. w m a x and w m i n are the maximum and minimum values of inertia weight; cf is the constriction factor; r1 and r2 are uniformly distributed random numbers in [0 1]; and c1 and c2 are known as the cognitive and social parameters, respectively.
Figure 1 shows the complete flowchart of the presented CF-PSO algorithm. However, this paper sets vmax and vmin to 4 and −4, respectively, and starts checking of the stopping criteria after 30 generations, through a systematic trial and error approach, and based on the experience as described in References [41,42,43]. Likewise, this paper sets other parameters as w m a x = 1.2, w m i n = 0.1 , 4.05 φ 4.15 ,   2.00 c 1 2.05 .

3. Proposed Fault Location Technique

The proposed fault location technique comprises of three steps namely the data generation, ST based feature extraction and selection, and training and testing of the machine learning tools to locate faults/identify faulty sections in the distribution grids. The following parts of this section describe the mentioned steps:

3.1. Data Generation

The proposed fault location/faulty section identification scheme, applies hundreds of faults on the selected test distribution feeders by varying the pre-fault loading condition, fault resistance, class, and inception angle. For each faulty case, the proposed technique records three-phase current signals from different locations/branches of the distribution feeders, with a sampling frequency of 20 kHz. The technique also keeps track of the fault locations and faulty sections, for future use to train different machine learning tools as the regression and classification problems, respectively. However, the ST extracted features are employed as inputs, and the fault locations and faulty sections are considered as the outputs.

3.2. ST Based Feature Extraction and Selection

To select the useful features, the proposed fault location technique decomposes one pre and one post cycle current signals recorded from an applied phase-A-to-ground (AG) fault, on the four-node test distribution feeder employing the ST. Hence, the developed approach calculates the following statistical features from the produced Scmax, Srmax, Emax, and Sc-phase-max vectors, as discussed in Section II (A):
  • F1 to F4: Standard deviations of the vectors.
  • F5 to F8: Maximum values of the vectors.
  • F9 to F12: Minimum values of the vectors.
  • F13 to F16: Mean values of the vectors.
  • F17 to F20: Entropies of the vectors.
  • F21 to F24: Skewness of the vectors.
  • F25 to F28: Kurtosis of the vectors.
Therefore, the proposed approach collects 28 features for the phase-A-current and repeats the same step for 500 similar faulty cases. Eventually, the technique ends up with a matrix of 500 × 28 for the phase-A-current. Then it selects twelve useful features, and removes twelve more redundant and insignificant features, based on their correlation factor. In addition, the technique also removes four more features, as they possess constant/zero values for all 500 faulty cases. Consequently, the developed technique selects a total of thirty-six (= 12 × 3) features, for each three-phase branch currents. Table 1 summarizes the feature selection and removal process.

3.3. Training and Testing of the Machine Learning Tools

As mentioned earlier, the proposed technique records three-phase faulty current signals from different locations/branches of the selected test distribution feeders and extracts thirty-six features employing the ST for each three-phase branch current signal, where the faulty cases are generated by varying the pre-fault loading condition, fault resistance, inception angle, class, and location. Then the technique employs the selected ST extracted features to the machine learning tools as inputs, and the fault locations/faulty sections as outputs. The proposed ST based MLT approach, considers the fault location problem as a regression problem, and the faulty section identification problem as a classification problem. It is worth mentioning, that the proposed approach employs 70% of the generated faulty cases for the training purpose and uses rest of the cases for testing purposes.
The proposed fault location technique (regression problem), evaluates different statistical performance indices, including the mean absolute percentage error (MAPE), root mean squared error (RMSE), percent bias (PBIAS), RMSE-observations to standard deviation ratio (RSR), the coefficient of determination (R2), Willmott’s index of agreement (WIA), and the Nash–Sutcliffe model efficiency coefficient (NSEC) on the test datasets, to investigate the efficacy of the employed machine learning tools. According to Lewis [44], any model is considered highly accurate for forecasting, if the developed model gives a MAPE less than 10%. Additionally, lower values of the RMSE and RSR indicate the strength of the model [45], whereas the ideal value of the PBIAS is zero, which says the developed model can predict desired outputs accurately. Negative and positive values of PBIAS, demonstrate the overestimation and underestimation of the predicted outputs, respectively. Moreover, the values of R2, WIA, and NSEC vary from 0 to 1. The value 1 refers to a perfect match between the predicted and actual outputs, whereas the value 0 demonstrates that the output cannot be predicted at all from available inputs [45,46]. Like the statistical performance indices, the scatter plot provides a comprehensive summary of a set of bivariate data (two variables) and is often employed to determine the potential associations between the variables. The resulting pattern, demonstrates the type and strength of the relationship between two variables. The more data points located in the neighborhood of the identity line (i.e., y = x), the more the data sets agree with each other. If the model output and actual data are the same, then all data points fall on the identity line. Based on the evaluated statistical performance indices of the test datasets, the proposed technique selects the fault location (regression problem) models. Conversely, the faulty section identification technique (classification problem) selects the efficient model based on the overall accuracy. However, after selection of the efficient models, the proposed ST based MLT approach diagnoses the faults based on the recorded three-phase current signals, as illustrated graphically in Figure 2.

3.4. Cross Validation

Cross-validation is a popular strategy for assessing how the results of statistical analysis techniques generalize to an independent data set [47]. To select an accurate model, the researchers developed several cross validation techniques, including the k-fold cross-validation, jackknife cross-validation, and independent data test [48,49,50,51,52,53]. This research work, employed the k-fold cross validation technique to select the appropriate machine learning predictor. The employed validation technique partitions the available data into k segments or folds (roughly equal) and employs k - 1 segments for training the predictor model, and the remaining segment for validation of the model.

4. Results and Discussions

The proposed fault location/faculty section identification techniques based on the ST and MLT, were tested on two different test distribution feeders. The following parts of this section, briefly introduce the test distribution feeders and provide essential discussions on the simulation results.

4.1. Example 1: Four-Node Test Distribution Feeder

The four-node test distribution feeder consists of two distribution transformers, a distribution line, a lumped load, and a source, as presented in Figure 3, whereas Table 2 presents the technical specifications of the test feeder. This research modeled the test distribution feeder for Example 1 in the MATLAB/SIMULINK environment, considering its unbalanced loading condition.

4.1.1. Fault Location with Un-Optimized MLT

The proposed fault location technique, generated 1200 cases of AG faults by varying the pre-fault loading condition, fault resistance, inception angle, and location. It recorded a three-phase current signal from the data measurement bus for each faulty case and employed the ST to extract the features from the recorded current signals. After selection of the previously mentioned thirty-six features for each faulty case, the proposed fault location technique fetched them as inputs to different machine learning tools, where their outputs were the corresponding fault locations. The proposed fault location scheme chose the SVM and ELM control parameters randomly, as summarized in Table 3 whereas the number of hidden neurons of the MLP-NN were chosen through a systematic trial and error approach. However, the proposed technique employed 70% of the available data for training and the rest of the data for testing purposes.
Figure 4 presents the targeted and the ST based MLT predicted fault locations, for forty randomly selected observations from the test dataset of AG faults, for comparison purpose. It visually appears that the MLP-NN technique adequately predicted the desired outputs in most of the cases. Conversely, the fault location accuracies of the other two techniques (the SVM and ELM) were poor, compared to the MLP-NN approach. Figure 5 presents the scatter plots of the test dataset, which depicts that the data points are closer to the identity line for the MLP-NN approach as expected for an efficient model. On the other hand, the scatter plots of the SVM and ELM models illustrated their inefficiency in locating faults in the four-node test distribution feeder. The dissatisfactory performance of these two models requires further investigation through the optimization of their key parameters, employing the well-known stochastic approaches to achieve better performance. However, like AG faulty cases, the proposed approach collected features for other types of faults including phase-B-to-ground (BG), phase-C-to-ground (CG), phase-A-phase-B-to-ground (ABG), phase-B-to-phase-C-ground (BCG), phase-C-to-phase-A-ground (CAG), and phase-A-to-phase-B-to-phase-C-to-ground (ABCG). Instead of presenting similar figures and discussions, this research summarized the operational times along with the selected statistical performance indices, for the test datasets of the ST based MLT techniques, for all seven types of faults in Table 4.
It showed the SVM technique required an almost three-fold time to get trained, compared to the MLP-NN technique, whereas the training process of the ELM technique was a hundred times faster than the MLP-NN approach. The test times for all the approaches were almost equal, and are less than few cycles of a 60 Hz power system network. Additionally, the RMSE, MAPE, and RSR values were quite low, whereas the R2, WIA, and NSCE were almost unity for the MLP-NN approach; indicating the strength and effectiveness of the MLP-NN approach in locating all types of faults. Furthermore, the PBIAS values were positive for a few cases and negative for the others, which demonstrates underestimations and overestimations of the fault distances, respectively. However, these values were very small and closer to zero; hence, the smaller underestimations and overestimations could be neglected for the ST based MLP-NN approach. On the contrary, the statistical performance indices of the other two approaches (the SVM and ELM), were not satisfactory for all seven types of faults. Consequently, they required further investigation through the optimization of their key parameters, employing the well-known stochastic approaches to achieve better accuracy.

4.1.2. Fault Location with Optimized MLT

The ST based MLP-NN technique, showed satisfactory performance in locating faults in four-node test feeders, but the performance of the other two MLT techniques were dissatisfactory. Consequently, this section optimizes their key parameters employing the constriction factor-based particle swarm optimization. The paper presented the detailed procedures of the CF-PSO, in Section 2.5. The objective function of the CF-PSO was to minimize the MAPE of the actual and the predicted fault locations, for the test datasets. After going through 100 generations with a population size of 20 individuals, the CF-PSO ended up with the optimized SVM control parameters. Similarly, the CF-PSO optimized the control parameters of the ELM in 100 generations, with a population size of 80 individuals. Table 5 summarizes the obtained optimal control parameters of both machine learning approaches. It is worth mentioning, that this paper neither optimized the number of hidden neurons nor the connecting weights and biases of the MLP-NN, due to its better performance with the systematically selected number of neurons.
Figure 6 presents the targeted and the ST based MLT predicted fault locations, for forty randomly selected observations from the test dataset of AG faults, where the SVM and ELM techniques used their CF-PSO optimized key parameters, as presented in Table 5. It visually appeared that all the MLT techniques adequately predict the desired outputs in almost every case. Furthermore, Figure 7 presents the scatter plots of the developed MLT for the test dataset, which depicts that the data points are closer to the identity line for the employed techniques. Hence, the presented results confirmed the superiority of the optimized MLT, over the MLT with randomly picked parameters. Table 6 summarizes the operational times, along with the selected statistical performance indices of the test datasets of the ST based optimized MLT approaches, for all seven types of faults. The training and testing times for the selected machine learning tools were like that of Table 4. As can be observed from the table, the RMSE, MAPE, and RSR values were quite low, whereas the R2, WIA, and NSCE values were almost unity, indicating the strength and effectiveness of the proposed approach in locating faults. However, the PBIAS values were positive for a few cases and negative for the others, which demonstrated underestimations and overestimations of the fault distances, respectively. However, these values were closer to zero; hence, the smaller underestimations and overestimations could be neglected.

4.1.3. Fault Location with Optimized MLT in the Presence of Measurement Noise

The presented results of the previous section, confirmed the superiority of the optimized MLT approaches over their non-optimized counter parts, where this section provides the results of the test datasets in the presence of measurement noise. Figure 8 shows the targeted and the ST based MLT models predicted fault locations, for forty randomly selected observations from the test dataset of AG faults, in the presences of measurement noise (30 dB SNR). It visually appeared that all ST based MLT approaches, estimated the fault distances with similar accuracy. Moreover, the scatter plots of Figure 9 of the developed MLT models for the test dataset, also depicts that the data points were closer to the identity line as expected for any efficient model. Additionally, Figure 10 presents the comparison of the predicted fault distances with the actual fault distances; and Figure 11 presents the scatter plot of the proposed MLT models, for the test dataset of AG faults, in the presence of 20 dB SNR. Figure 8 to Figure 11 illustrate the efficacy of proposed ST based MLT models in predicting the fault distances efficiently, even with the presence of measurement noise. Furthermore, Table 7 and Table 8 present the selected statistical performance indices, along with the required operational time of the proposed ST based MLT approaches, in the presence of 30 dB and 20 dB SNR, respectively. It can be observed from both tables, that the training and testing times, for the selected machine learning tools, are like that of Table 4.
Additionally, the RMSE, MAPE, RSR, and PBIAS were quite low, whereas the R2, WIA, and NSCE were almost unity for the optimized MLT approaches, indicating the strength and the effectiveness of the approaches, for all types of faults, even in the presence of measurement noise. However, the performance of the developed scheme degrades a bit in the presence of measurement noise, but it can still estimate fault locations with acceptable accuracy.

4.1.4. Validating the Developed Fault Location Technique

In order to validate the developed ST based fault location scheme for the four-node test distribution feeder, the following faults were applied on the distribution line and predicted the fault location employing the trained and tested machine learning tools. It is worth mentioning, the proposed approach generated the faulty cases by varying the pre-fault loading condition, presence of measurement noise, and fault resistance and inception angle. Then it employed the ST to extract the useful features and fetched them to the trained machine learning tools as inputs, to get a decision on the applied faults. As can be seen from Table 9, the developed signal processing based MLT approach located all applied faults with satisfactory accuracy (<1%), which validated the effectiveness of the employed approach.

4.2. Example 2: Thirteen-Node Test Distribution Feeder

The proposed ST based machine learning approach was also employed to identify the faulty sections of the IEEE 13-node test distribution feeder. The selected test distribution feeder operates at 4.16 kV and displays most of the characteristic features of the power distribution grids, as shown in Figure 12 [54]. Consequently, it is usually used to test common features of the distribution system analysis software. The highly loaded test feeder consists of a single voltage regulator at the substation, an in-line transformer, shunt capacitor banks, overhead distribution lines, and underground cables of various configurations, and several unbalanced spot and distributed loads. Moreover, the test feeder contained single-phase, double-phase, and three-phase laterals. This research modeled the test feeder in an RSCAD environment and implemented the developed model in a Real Time Digital Simulator (RTDS) machine. The proposed ST based faulty section identification approach generated 900 faulty cases by varying the pre-fault loading condition, fault resistance, inception angle, type, and location, through employing the ‘batch-mode’ operation option of the RSCAD software. After generation of different fault scenarios, the proposed approach recorded three-phase current signals from eight different branches/locations of the test feeder, as indicated in Figure 12, with a sampling frequency of 20 kHz. Then it processed the recorded signals, employing ST in the MATLAB/SIMULINK environment and extracted useful features, as discussed in Section 3.2. It is worth mentioning, that the proposed approach collected 288 features (~8 branches × 36 features per branches), for each fault scenario. Furthermore, this research divided the test distribution feeder into nine sections, as indicated by the rectangular boxes in Figure 12. Consequently, this research prepared the faulty section identification problem as a classification problem.

4.2.1. Faulty Section Identification in Noise-Free Environment

After collection of useful features, the proposed approach trained and tested the MLP-NN with a different number of neurons, where the best neural network consisted of eleven hidden neurons, in terms of minimum mean squared error and overall accuracy. In addition, the proposed approach systematically chose the tan-sigmoid as a squashing function and the resilient backpropagation technique as the training algorithm. Like the fault location technique of the four-node test distribution feeder, the faulty section identification approach also employed 70% of the available data for training purposes and the rest of the data for testing purposes. Table 10 summarizes the technical features related to the MLP neural networks employed for faulty section identification purposes. Table 11 presents a 9 × 9 confusion matrix obtained from the MLP neural networks, as they identified nine faulty sections employing the proposed approach in a noise-free environment. The diagonal and off-diagonal elements of the confusion matrix, represent the successful and unsuccessful identification of faulty sections, respectively. As can be observed from the table, the proposed technique identified the faulty sections with an accuracy of almost 100%, which demonstrated the successful implementation of the adopted ST based machine learning approach.

4.2.2. Faulty Section Identification in the Presence of Measurement Noise

The proposed technique added a different level of additive white gaussian noise to the recorded three-phase branch current signals, with a view to testing the efficacy of the proposed faulty section identification approach, in the presence of measurement noise. Then it extracted the same features from the noisy signals, and employed them as inputs to the MLP-NN. Table 12 summarizes the faulty section identification results of the ST based MLP-NN approach, in the presence of measurement noise. As can be seen, the ST based approach ended up with the overall accuracies of 99.876%, 99.741%, and 99.235% for the 40 dB, 30 dB, and 20 dB SNR, respectively. Consequently, the obtained results validated the efficacy of the proposed ST based faulty section identification technique, for both noise-free and noisy conditions. In addition, the developed approach confirmed its independence on the pre-fault loading condition, fault type, resistance, and inception angle.

4.2.3. Validating the Developed Faulty Section Identification Technique

To validate the developed ST based faulty section identification technique for the thirteen-node test distribution feeder, the following faults were applied on different locations of the feeder and predicted the faulty section employing the trained and tested MLP-NN. It is worth mentioning, the proposed approach generated the faulty cases by varying the pre-fault loading condition, presence of measurement noise, fault resistance, and inception angle. Then it employed the ST to extract the useful features and fetched them to the trained MLP-NN as inputs, to get a decision on the applied faults. As can be seen from Table 13, the developed signal processing based MLP-NN approach identified all applied faults accurately, which validated the effectiveness of the employed approach.

5. Conclusions and Future Scope

This paper presented advanced signal processing-based machine learning tools to locate faults and identify faulty sections in distribution grids. The proposed approach decomposed three-phase current signals recorded from different branches/locations of the distribution grids, employing the ST to extract the useful characteristic features of faulty cases. Then it employed the extracted features as inputs to different machine learning tools and trained them to locate the faults, and to identify the faulty section in the distribution grids. Additionally, this research employed the CF-PSO technique to optimize the parameters of the support vector and extreme learning machines; whilst it optimized the MLP-NN through a systematic approach to achieve better generalization performance. The presented results validated the efficacy of the developed fault diagnosis approach. Additionally, the optimized machine learning tools outperformed the non-optimized tools in locating distribution grid faults, in terms of the selected statistical performance indices, i.e., the RMSE, MAPE, PBIAS, RSR, R2, WIA, and NSEC. Moreover, the ST based ELM approach required lesser training time compared to the SVM and MLP-NN techniques. However, the trained MLT diagnosed the applied faults in a fraction of a second, signaling the real-time implementation of the developed fault diagnosis scheme. Furthermore, the developed faulty section identification approach, identified faulty sections with almost 100% accuracy. Finally, the presented results confirmed the independency of the proposed ST based MLT approach, in the fault inception angle and resistance, pre-fault loading conditions, and presence of measurement noise. As a future extension, the signal processing based MLT approach can be applied to diagnose faults in distribution grids, with the incorporation of renewable energy resources, considering their associated uncertainties. The proposed approach can also be extended to diagnose simultaneous faults at distribution level.

Author Contributions

All authors contributed to this work through collaboration. M.S. is the first author in this manuscript. All authors revised and approved for the publication.

Funding

This research received no external funding.

Acknowledgments

This work was supported by the King Abdulaziz City for Science and Technology through the Science and Technology Unit, King Fahd University of Petroleum & Minerals, under Project 14-ENE265-04, as a part of the National Science, Technology and Innovation Plan.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Bahmanyar, A.; Jamali, S.; Estebsari, A.; Bompard, E. A comparison framework for distribution system outage and fault location methods. Electr. Power Syst. Res. 2017, 145, 19–34. [Google Scholar] [CrossRef]
  2. Shafiullah, M.; Abido, M.A. A Review on Distribution Grid Fault Location Techniques. Electr. Power Components Syst. 2017, 45, 807–824. [Google Scholar] [CrossRef]
  3. Awalin, L.; Mokhlis, H.; Bakar, A. Recent developments in fault location methods for distribution networks. Prz. Elektrotech. 2012, 88, 206–212. [Google Scholar]
  4. Salim, R.; Resener, M.; Filomena, A.D.; De Oliveira, K.R.C.; Bretas, A.S. Extended fault-location formulation for power distribution systems. IEEE Trans. Power Deliv. 2009, 24, 508–516. [Google Scholar] [CrossRef]
  5. Filomena, A.D.; Resener, M.; Salim, R.H.; Bretas, A.S. Fault location for underground distribution feeders: An extended impedance-based formulation with capacitive current compensation. Int. J. Electr. Power Energy Syst. 2009, 31, 489–496. [Google Scholar] [CrossRef]
  6. Salim, R.H.; Salim, K.C.O.; Bretas, A.S. Further improvements on impedance-based fault location for power distribution systems. IET Gener. Transm. Distrib. 2011, 5, 467. [Google Scholar] [CrossRef]
  7. Girgis, A.A.; Fallon, C.M.; Lubkeman, D.L. A fault location technique for rural distribution feeders. IEEE Trans. Ind. Appl. 1993, 29, 1170–1175. [Google Scholar] [CrossRef]
  8. Dashti, R.; Sadeh, J. Accuracy improvement of impedance-based fault location method for power distribution network using distributed-parameter line model. Int. Trans. Electr. Energy Syst. 2014, 24, 318–334. [Google Scholar] [CrossRef]
  9. Das, S.; Karnik, N.; Santoso, S. Distribution Fault-Locating Algorithms Using Current Only. IEEE Trans. Power Deliv. 2012, 27, 1144–1153. [Google Scholar] [CrossRef]
  10. Tang, J.; Yin, X.; Zhang, Z. Traveling-Wave-Based Fault Location in Electrical Distribution Systems With Digital Simulations. Telecommun. Comput. Electron. Control 2014, 12, 297. [Google Scholar] [Green Version]
  11. Yan, F.; Liu, W.; Tian, L. Fault location for 10kV distribution line based on traveling wave-ANN theory. In Proceedings of the 2011 IEEE Power Engineering and Automation Conference, Wuhan, China, 8–9 September 2011. [Google Scholar]
  12. Sheng-nan, Y.; Yi-han, Y.; Hai, B. Study on fault location in distribution network based on C-type traveling-wave scheme. Relay 2007, 35, 1–5. [Google Scholar]
  13. Dehghani, F.; Nezami, H. A new fault location technique on radial distribution systems using artificial neural network. In Proceedings of the 22th International Conference and Exhibition on Electricity Distribution (CIRED), Stockholm, Sweden, 10–13 June 2013. [Google Scholar]
  14. Pourahmadi-Nakhli, M.; Safavi, A.A. Path Characteristic Frequency-Based Fault Locating in Radial Distribution Systems Using Wavelets and Neural Networks. IEEE Trans. Power Deliv. 2011, 26, 772–781. [Google Scholar] [CrossRef]
  15. Rafinia, A.; Moshtagh, J. A new approach to fault location in three-phase underground distribution system using combination of wavelet analysis with ANN and FLS. Int. J. Electr. Power Energy Syst. 2014, 55, 261–274. [Google Scholar] [CrossRef]
  16. Ray, P.; Mishra, D. Artificial Intelligence Based Fault Location in a Distribution System. In Proceedings of the 2014 International Conference on Information Technology, Bhubaneswar, India, 22–24 December 2014. [Google Scholar]
  17. Shafiullah, M.; Ijaz, M.; Abido, M.A.; Al-Hamouz, Z. Optimized support vector machine & wavelet transform for distribution grid fault location. In Proceedings of the 2017 11th IEEE International Conference on Compatibility, Power Electronics and Power Engineering (CPE-POWERENG), Cadiz, Spain, 4–6 April 2017. [Google Scholar]
  18. Stockwell, R.G.; Mansinha, L.; Lowe, R.P. Localization of the complex spectrum: The S transform. IEEE Trans. Signal Process. 1996, 44, 998–1001. [Google Scholar] [CrossRef]
  19. Mansinha, L.; Stockwell, R.G.; Lowe, R.P. Pattern analysis with two-dimensional spectral localisation: Applications of two-dimensional S transforms. Phys. A Stat. Mech. Its Appl. 1997, 239, 286–295. [Google Scholar] [CrossRef]
  20. Wang, Y. Efficient stockwell transform with applications to image processing. Ph.D. Thesis, University of Waterloo, Waterloo, ON, Canada, 20 May 2011. [Google Scholar]
  21. Shafiullah, M.; Abido, M.A. S-Transform Based FFNN Approach for Distribution Grids Fault Detection and Classification. IEEE Access 2018, 6, 8080–8088. [Google Scholar] [CrossRef]
  22. Khokhar, S.; Mohd Zin, A.A.B.; Mokhtar, A.S.B.; Pesaran, M. A comprehensive overview on signal processing and artificial intelligence techniques applications in classification of power quality disturbances. Renew. Sustain. Energy Rev. 2015, 51, 1650–1663. [Google Scholar] [CrossRef]
  23. Ijaz, M.; Shafiullah, M.; Abido, M.A. Classification of power quality disturbances using Wavelet Transform and Optimized ANN. In Proceedings of the 2015 18th International Conference on Intelligent System Application to Power Systems (ISAP), Porto, Portugal, 11–16 September 2015. [Google Scholar]
  24. Adewole, A.C.; Tzoneva, R.; Behardien, S. Distribution network fault section identification and fault location using wavelet entropy and neural networks. Appl. Soft Comput. 2016, 46, 296–306. [Google Scholar] [CrossRef]
  25. Sejdić, E.; Djurović, I.; Jiang, J. Time–frequency feature representation using energy concentration: An overview of recent advances. Digit. Signal Process. 2009, 19, 153–183. [Google Scholar] [CrossRef]
  26. Shafiullah, M.; Abido, M.A.; Al-Hamouz, Z. Wavelet-based extreme learning machine for distribution grid fault location. IET Gener. Transm. Distrib. 2017, 11, 4256–4263. [Google Scholar] [CrossRef]
  27. Mishra, S.; Bhende, C.N.; Panigrahi, B.K. Detection and Classification of Power Quality Disturbances Using S-Transform and Probabilistic Neural Network. IEEE Trans. Power Deliv. 2008, 23, 280–287. [Google Scholar] [CrossRef]
  28. Roy, N.; Bhattacharya, K. Detection, Classification, and Estimation of Fault Location on an Overhead Transmission Line Using S-transform and Neural Network. Electr. Power Components Syst. 2015, 43, 461–472. [Google Scholar] [CrossRef]
  29. Krishnanand, K.R.; Dash, P.K.; Naeem, M.H. Detection, classification, and location of faults in power transmission lines. Int. J. Electr. Power Energy Syst. 2015, 67, 76–86. [Google Scholar] [CrossRef]
  30. Masiur Rahman, S.; Khondaker, A.N.; Imtiaz Hossain, M.; Shafiullah, M.; Hasan, M.A. Neurogenetic modeling of energy demand in the United Arab Emirates, Saudi Arabia, and Qatar. Environ. Prog. Sustain. Energy 2017, 36, 1208–1216. [Google Scholar] [CrossRef]
  31. Ansari, H.R.; Zarei, M.J.; Sabbaghi, S.; Keshavarz, P. A new comprehensive model for relative viscosity of various nanofluids using feed-forward back-propagation MLP neural networks. Int. Commun. Heat Mass Transf. 2018, 91, 158–164. [Google Scholar] [CrossRef]
  32. Rana, M.J.; Shahriar, M.S.; Shafiullah, M. Levenberg–Marquardt neural network to estimate UPFC-coordinated PSS parameters to enhance power system stability. Neural Comput. Appl. 2017, 1–12. [Google Scholar] [CrossRef]
  33. Boser, B.E.; Guyon, I.M.; Vapnik, V.N. A training algorithm for optimal margin classifiers. In Proceedings of the fifth annual workshop on Computational learning theory (COLT), Pittsburgh, PA, USA, 27–29 July 1992. [Google Scholar]
  34. Vapnik, V.; Golowich, S.E.; Smola, A.J. Support Vector Method for Function Approximation, Regression Estimation, and Signal Processing. Adv. Neural Inf. Process. Syst. 1997, 9, 281–287. [Google Scholar]
  35. Canu, S.; Grandvalet, Y.; Guigue, V.; Rakotomamonjy, A. SVM and kernel methods matlab toolbox. Percept. Systmes Inf. 2003, 69, 70. [Google Scholar]
  36. Huang, G.-B.; Zhu, Q.-Y.; Siew, C.-K. Extreme learning machine: a new learning scheme of feedforward neural networks. In Proceedings of the 2004 IEEE International Joint Conference on Neural Networks (IEEE Cat. No.04CH37541), Budapest, Hungary, 25–29 July 2004. [Google Scholar]
  37. Cambria, E.; Huang, G.B.; Kasun, L.L.C.; Zhou, H.; Vong, C.M.; Lin, J.; Yin, J.; Cai, Z.; Liu, Q.; Li, K.; et al. Extreme Learning Machines [Trends & Controversies]. IEEE Intell. Syst. 2013, 28, 30–59. [Google Scholar]
  38. MATLAB Codes of ELM Algorithm. Available online: http://www.ntu.edu.sg/home/egbhuang/elm_kernel.html (accessed on 29 Feburary 2016).
  39. Eberhart, R.; Kennedy, J. A new optimizer using particle swarm theory. Proceedings of Sixth International Symposium on Micro Machine and Human Science (MHS), Nagoya, Japan, 4–6 October 1995. [Google Scholar]
  40. Clerc, M. The swarm and the queen: towards a deterministic and adaptive particle swarm optimization. In Proceedings of the 1999 Congress on Evolutionary Computation-CEC99, Washington, DC, USA, 6–9 July 1999. [Google Scholar]
  41. Malekpour, A.; Seifi, A. Application of constriction factor particle swarm optimization to optimum load shedding in power system. Mod. Appl. Sci. 2010, 4, 188–196. [Google Scholar] [CrossRef]
  42. Patwardhan, R.P.; Mhetre, S.L. Effect of constriction factor on minimization of transmission power loss using Particle Swarm Optimization. In Proceedings of the 2015 International Conference on Energy Systems and Applications, Pune, India, 30 October–1 November 2015. [Google Scholar]
  43. Alam, M.N.; Kumar, K.; Mathur, A. Economic load dispatch considering valve-point effects using time varying constriction factor based particle swarm optimization. In Proceedings of the 2015 IEEE UP Section Conference on Electrical Computer and Electronics (UPCON), Allahabad, India, 4–6 December 2015. [Google Scholar]
  44. Lewis, C.D. Industrial and Business Forecasting Methods: A Practical Guide to Exponential Smoothing and Curve Fitting; Butterworth–Heinemann: Oxford, London, 1982; ISBN 0408005599. [Google Scholar]
  45. Moriasi, D.N.; Arnold, J.G.; Van Liew, M.W.; Bingner, R.L.; Harmel, R.D.; Veith, T.L. Model evaluation guidelines for systematic quantification of accuracy in watershed simulations. Trans. ASABE 2007, 50, 885–900. [Google Scholar] [CrossRef]
  46. Willmott, C.J.; Robeson, S.M.; Matsuura, K. A refined index of model performance. Int. J. Climatol. 2012, 32, 2088–2094. [Google Scholar] [CrossRef]
  47. Arlot, S.; Celisse, A. A survey of cross-validation procedures for model selection. Stat. Surv. 2010, 4, 40–79. [Google Scholar] [CrossRef] [Green Version]
  48. Manavalan, B.; Shin, T.H.; Lee, G. PVP-SVM: Sequence-Based Prediction of Phage Virion Proteins Using a Support Vector Machine. Front. Microbiol. 2018, 9, 476. [Google Scholar] [CrossRef] [PubMed]
  49. Dao, F.-Y.; Yang, H.; Su, Z.-D.; Yang, W.; Wu, Y.; Hui, D.; Chen, W.; Tang, H.; Lin, H. Recent Advances in Conotoxin Classification by Using Machine Learning Methods. Molecules 2017, 22, 1057. [Google Scholar] [CrossRef] [PubMed]
  50. Manavalan, B.; Shin, T.H.; Kim, M.O.; Lee, G. AIPpred: Sequence-Based Prediction of Anti-inflammatory Peptides Using Random Forest. Front. Pharmacol. 2018, 9. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  51. Lin, H.; Liang, Z.-Y.; Tang, H.; Chen, W. Identifying sigma70 promoters with novel pseudo nucleotide composition. IEEE/ACM Trans. Comput. Biol. Bioinform. 2017, 1, 1–6. [Google Scholar] [CrossRef] [PubMed]
  52. Manavalan, B.; Subramaniyam, S.; Shin, T.H.; Kim, M.O.; Lee, G. Machine-Learning-Based Prediction of Cell-Penetrating Peptides and Their Uptake Efficiency with Improved Accuracy. J. Proteome Res. 2018, 17, 2715–2726. [Google Scholar] [CrossRef] [PubMed]
  53. Manavalan, B.; Shin, T.H.; Lee, G. DHSpred: support-vector-machine-based human DNase I hypersensitive sites prediction using the optimal features selected by random forest. Oncotarget 2018, 9, 1944–1956. [Google Scholar] [CrossRef] [PubMed]
  54. Distribution Test Feeders. Available online: https://ewh.ieee.org/soc/pes/dsacom/testfeeders/ (accessed on 6 January 2018).
Figure 1. The flowchart of the constriction factor particle swarm optimization.
Figure 1. The flowchart of the constriction factor particle swarm optimization.
Energies 11 02328 g001
Figure 2. Fault location/faulty section identification flowchart employing the ST based MLT approach.
Figure 2. Fault location/faulty section identification flowchart employing the ST based MLT approach.
Energies 11 02328 g002
Figure 3. Four-node test distribution feeder used for fault location purpose.
Figure 3. Four-node test distribution feeder used for fault location purpose.
Energies 11 02328 g003
Figure 4. Comparison of the Stockwell transform (ST) based non-optimized MLT predicted and actual fault distances, for the test dataset of applied phase-A-to-ground (AG) fault.
Figure 4. Comparison of the Stockwell transform (ST) based non-optimized MLT predicted and actual fault distances, for the test dataset of applied phase-A-to-ground (AG) fault.
Energies 11 02328 g004
Figure 5. Scatter plot of the targeted and the ST based non-optimized MLT predicted outputs, for the test dataset of AG fault.
Figure 5. Scatter plot of the targeted and the ST based non-optimized MLT predicted outputs, for the test dataset of AG fault.
Energies 11 02328 g005
Figure 6. Comparison of the ST based constriction factor-based particle swarm optimization (CF-PSO) optimized MLT predicted and actual fault distances for the test dataset of AG fault.
Figure 6. Comparison of the ST based constriction factor-based particle swarm optimization (CF-PSO) optimized MLT predicted and actual fault distances for the test dataset of AG fault.
Energies 11 02328 g006
Figure 7. Scatter plot of the targeted and the ST based CF-PSO optimized MLT predicted outputs for the test dataset of AG fault.
Figure 7. Scatter plot of the targeted and the ST based CF-PSO optimized MLT predicted outputs for the test dataset of AG fault.
Energies 11 02328 g007
Figure 8. Comparison of the ST based optimized MLT predicted and actual fault distances, for the test dataset, in the presence of 30 dB SNR.
Figure 8. Comparison of the ST based optimized MLT predicted and actual fault distances, for the test dataset, in the presence of 30 dB SNR.
Energies 11 02328 g008
Figure 9. Scatter plot of the targeted and the ST based optimized MLT predicted outputs, for the test dataset, in the presence of 30 dB SNR.
Figure 9. Scatter plot of the targeted and the ST based optimized MLT predicted outputs, for the test dataset, in the presence of 30 dB SNR.
Energies 11 02328 g009
Figure 10. Comparison of the ST based optimized MLT predicted and actual fault distances, for the test dataset, in the presence of 20 dB SNR.
Figure 10. Comparison of the ST based optimized MLT predicted and actual fault distances, for the test dataset, in the presence of 20 dB SNR.
Energies 11 02328 g010
Figure 11. Scatter plot of the targeted and the ST based optimized MLT predicted outputs, for the test dataset, in the presence of 20 dB SNR.
Figure 11. Scatter plot of the targeted and the ST based optimized MLT predicted outputs, for the test dataset, in the presence of 20 dB SNR.
Energies 11 02328 g011
Figure 12. IEEE thirteen-node test distribution feeder.
Figure 12. IEEE thirteen-node test distribution feeder.
Energies 11 02328 g012
Table 1. ST Extracted Feature Selection and Removal Process Summary.
Table 1. ST Extracted Feature Selection and Removal Process Summary.
Selected FeaturesCorrelation Factor/CommentAction
F5 & F61.00Remove feature F6
F9 & F131.00Remove feature F13
F21 & F241.00Remove feature F24
F25 & F281.00Remove feature F28
F21 & F250.9999Remove feature F25
F5 & F90.9997Remove feature F9
F11 & F150.9996Remove feature F15
F5 & F80.9965Remove feature F8
F22 & F260.9958Remove feature F26
F7 & F110.9957Remove feature F7
F8, F12 & F161.00Remove features F12 & F16
F17, F19 & F20Most of the cases they are zeroRemove all of them
F27It is constant for all casesRemove F27
The selected features: F1, F2, F3, F4, F5, F10, F11, F14, F18, F21, F22, and F23
Table 2. Four-node test distribution feeder specifications.
Table 2. Four-node test distribution feeder specifications.
ItemDetails
Transformer I12 MVA, 120/25 kV, Yg-D
Transformer II12 MVA, 25 kV/575 V, Yg-D
Length of distribution line30 km
System frequency60 Hz
Load (Unbalanced)Phase A (1 MW, 0.48 MVAR)
Phase B (2 MW, 0.65 MVAR)
Phase C (2.5 MW, 1.2 MVAR)
Load variations±10%
Table 3. Randomly selected machine learning tools (MLT) control parameters.
Table 3. Randomly selected machine learning tools (MLT) control parameters.
MLTValues of the Control Parameters
MLP-NNNumber of hidden neuron = 6
SVRC = 1000, λ = 10−8, ε = 0.1, KO = 100 and kernel = Gaussian RBF
ELMCR = 1010, Kp = 1.5 × 108 and kernel = Gaussian RBF
Table 4. Operation time and statistical performance indices for test datasets with randomly selected MLT parameters.
Table 4. Operation time and statistical performance indices for test datasets with randomly selected MLT parameters.
Fault TypeTechniqueTimeStatistical Performance Measures
TrainingTestingRMSEMAPERSRPBIASR2WIANSCE
AGANN11.54690.01560.33330.04420.03780.06420.99930.99960.9986
SVR28.60940.04692.64320.46120.29990.05690.95760.97750.9100
ELM0.06250.01562.27210.41520.2578−1.2380.96710.98340.9335
BGANN11.50000.01560.44190.08160.0517−0.1930.99870.99930.9973
SVR27.46880.03132.90870.52460.34001.55220.94430.97110.8844
ELM0.07810.06252.20180.44990.25740.59750.96640.98340.9338
CGANN12.23440.03130.33270.05890.0385−0.1510.99930.99960.9985
SVR28.76560.04692.20300.43760.25501.11590.96860.98370.9350
ELM0.12500.04691.67980.34930.19440.46690.98100.99050.9622
ABGANN11.00000.01560.26270.03370.02980.08490.99960.99980.9991
SVR33.42190.01563.15660.61230.35790.66870.97410.96800.8719
ELM0.09380.01561.85940.31390.2108−0.3840.97760.98890.9555
BCGANN12.37500.01561.92960.18430.22401.20460.97480.98750.9498
SVR29.60940.01563.00600.57060.34891.12420.97590.96960.8783
ELM0.04690.01561.80440.30470.20940.01520.97790.98900.9561
CAGANN11.14060.04690.20190.02610.0228−0.0300.99970.99990.9995
SVR31.65630.01563.12850.65070.35330.59500.97320.96880.8752
ELM0.12500.01561.93580.34510.21860.91080.97590.98810.9522
ABCGANN12.64060.03130.34960.03270.0425−0.0020.99910.99950.9982
SVR29.25000.04693.62390.72790.44091.73760.98670.95140.8056
ELM0.06250.03132.50050.43270.3042−0.5350.95290.97690.9075
Table 5. CF-PSO optimized MLT control parameters.
Table 5. CF-PSO optimized MLT control parameters.
MLTValues of the Control Parameters
MLP-NNNumber of hidden neuron = 6
SVRC = 1000, λ = 8.5 × 10−3, ε = 4.75×10−1, KO = 100 and kernel = Gaussian RBF
ELMCR = 5 × 1011, Kp = 3.5 × 109 and kernel = Gaussian RBF
Table 6. Operation time and statistical performance indices for test datasets with CF-PSO optimized MLT parameters.
Table 6. Operation time and statistical performance indices for test datasets with CF-PSO optimized MLT parameters.
Fault TypeTechniqueTimeStatistical Performance Measures
TrainingTestingRMSEMAPERSRPBIASR2WIANSCE
AGANN11.54690.01560.33330.04420.03780.06420.99930.99960.9986
SVR26.28130.01560.36460.00940.04140.03040.99910.99960.9983
ELM0.12500.04690.25170.02660.02860.03880.99960.99980.9992
BGANN12.50000.01560.44190.08160.0517−0.1930.99870.99930.9973
SVR26.32810.01560.22710.00580.02650.10450.99960.99980.9993
ELM0.07810.06250.31720.03350.0371−0.0350.99930.99970.9986
CGANN12.23440.03130.33270.05890.0385−0.1510.99930.99960.9985
SVR29.84380.01560.22630.00300.02620.20700.99970.99980.9993
ELM0.09380.03130.16680.02210.0193−0.0010.99980.99990.9996
ABGANN11.00000.01560.26270.03370.02980.08490.99960.99980.9991
SVR28.12500.01560.48980.08140.05550.09850.99920.99920.9969
ELM0.07810.01560.18990.02090.0215−0.0120.99980.99990.9995
BCGANN12.37500.01561.92960.18430.22401.20460.97480.98750.9498
SVR30.18750.03130.48680.07990.05650.18960.99910.99920.9968
ELM0.06250.04690.16380.02010.01900.03090.99980.99990.9996
CAGANN12.14060.04690.20190.02610.0228−0.0300.99970.99990.9995
SVR26.51560.03130.48780.08740.05510.01110.99910.99920.9970
ELM0.10940.03130.17140.01880.0194−0.0120.99980.99990.9996
ABCGANN11.64060.03130.34960.03270.0425−0.0010.99910.99950.9982
SVR30.78130.04690.51380.08520.06250.33960.99940.99900.9961
ELM0.06250.03130.29420.02550.03580.22540.99940.99970.9987
Table 7. Operation time and the statistical performance indices, for the test datasets with CF-PSO optimized MLT, in the presence of 30 dB SNR.
Table 7. Operation time and the statistical performance indices, for the test datasets with CF-PSO optimized MLT, in the presence of 30 dB SNR.
Fault TypeTechniqueTimeStatistical Performance Measures
TrainingTestingRMSEMAPERSRPBIASR2WIANSCE
AGANN12.39060.01560.47750.07610.0542−0.0750.99860.99930.9971
SVR30.46880.03130.41790.05640.0474−0.0060.99890.99940.9978
ELM0.12500.01560.49590.06060.05630.11730.99840.99920.9968
BGANN12.00000.01560.55510.09930.0649−0.4430.99790.99890.9958
SVR26.64060.03130.21840.00420.0255−0.0560.99970.99980.9993
ELM0.07810.01560.36410.05290.0426−0.2430.99910.99950.9982
CGANN12.28130.03130.27290.04950.0316−0.1450.99950.99980.9990
SVR27.34380.03130.70080.12910.08110.58640.99690.99840.9934
ELM0.07810.06250.28150.03570.03260.16620.99950.99970.9989
ABGANN12.95310.01560.22200.02640.02520.09210.99970.99980.9994
SVR31.40630.01560.48980.08140.05550.07580.99910.99920.9969
ELM0.06250.01560.25510.03070.02890.02610.99960.99980.9992
BCGANN11.93750.03130.14210.02290.0165−0.0940.99990.99990.9997
SVR33.10940.01560.48760.08020.05660.15670.99910.99920.9968
ELM0.07810.01560.22710.02440.02640.03010.99970.99980.9993
CAGANN12.28130.04690.15990.02650.01810.01640.99980.99990.9997
SVR28.98440.01560.48800.08760.05510.03130.99910.99920.9970
ELM0.09380.03130.22650.02410.02560.09120.99970.99980.9993
ABCGANN10.73440.01560.19600.02520.02380.04350.99970.99990.9994
SVR29.39060.01560.51380.08520.06250.32030.99930.99900.9961
ELM0.04690.04690.44370.04080.05400.12010.99850.99930.9971
Table 8. Operation time and the statistical performance indices, for the test datasets with CF-PSO optimized MLT, in the presence of 20 dB SNR.
Table 8. Operation time and the statistical performance indices, for the test datasets with CF-PSO optimized MLT, in the presence of 20 dB SNR.
Fault TypeTechniqueTimeStatistical Performance Measures
TrainingTestingRMSEMAPERSRPBIASR2WIANSCE
AGANN12.73440.03130.63080.10320.07160.18230.99750.99870.9949
SVR26.73440.03130.45180.06350.05130.03680.99870.99930.9974
ELM0.12500.03130.54130.06840.0614−0.1900.99810.99910.9962
BGANN11.34380.01560.74140.12150.08670.04100.99620.99810.9925
SVR31.57810.03130.56870.01440.06650.42960.99780.99890.9956
ELM0.10940.01560.67170.07750.07851.14190.99710.99850.9938
CGANN10.65630.03130.45700.06250.0529−0.1290.99860.99930.9972
SVR33.89060.03130.74560.13520.08630.14540.99640.99810.9926
ELM0.07810.03130.33660.04250.0390−0.1180.99920.99960.9985
ABGANN11.62500.03130.31260.03570.03540.06720.99940.99970.9987
SVR32.59380.01560.49890.08230.05660.14670.99910.99920.9968
ELM0.06250.01560.36780.03600.0417−0.0320.99910.99960.9983
BCGANN11.43750.03130.40050.07780.04650.04450.99890.99950.9978
SVR28.45310.19200.49340.08010.05730.17870.99900.99920.9967
ELM0.07810.03130.37470.03460.04350.03310.99910.99950.9981
CAGANN11.85940.01560.27530.04070.03110.00330.99950.99980.9990
SVR27.59380.48980.49720.08900.0562−0.1460.99900.99920.9968
ELM0.10940.19200.33880.03920.03830.13230.99930.99960.9985
ABCGANN10.32810.03130.41880.04530.05090.10530.99870.99940.9974
SVR29.12500.01560.51610.08540.06280.35740.99940.99900.9961
ELM0.09380.48980.67100.05040.08160.22330.99670.99830.9933
Table 9. Fault location results obtained employing the ST based machine learning tools, for the four-node test distribution feeder.
Table 9. Fault location results obtained employing the ST based machine learning tools, for the four-node test distribution feeder.
Applied Fault InformationEmployed
Technique
Predicted Fault Information
TypeLocation (km)Location (km)Accuracy (%)
AG9.0ANN8.99990.0003
SVR9.00710.0237
ELM8.99610.0130
BG27.5ANN27.30930.6357
SVR27.49080.0307
ELM27.50840.0280
CG17.5ANN17.52470.0823
SVR17.22490.09170
ELM17.53710.1237
ABG3.5ANN3.51690.0563
SVR3.50710.0237
ELM3.50000.0000
BCG21.5ANN21.52810.0937
SVR21.50710.0237
ELM21.50010.0003
CAG10.0ANN10.06910.2303
SVR10.00710.0237
ELM10.00020.0007
ABCG24.0ANN23.76090.7970
SVR23.72490.9170
ELM23.99990.0003
Table 10. Multilayer perceptron neural networks (MLP-NN) parameters for the faulty section identification problem.
Table 10. Multilayer perceptron neural networks (MLP-NN) parameters for the faulty section identification problem.
ItemSpecifications
Number of hidden neurons11
Squashing functionTan-sigmoid
Training algorithmResilient backpropagation
Objective functionMean squared error (MSE)
Training data70%
Testing data30%
Table 11. Faulty section identification results for noise-free measurement.
Table 11. Faulty section identification results for noise-free measurement.
SymbolS1S2S3S4S5S6S7S8S9
S190000000000
S209000000000
S301899000000
S400090000000
S500009000000
S600000900000
S700000089910
S800000009000
S900000000900
Overall accuracy = 99.975%
Table 12. Faulty section identification results under noisy measurement.
Table 12. Faulty section identification results under noisy measurement.
Faulty SectionSamples Classified Successfully
40 dB SNR30 dB SNR20 dB SNR
S1900897895
S2899896893
S3898899894
S4897896891
S5900900896
S6900895892
S7898899895
S8900900893
S9898895889
Overall accuracy (%)99.87699.74199.235
Table 13. Faulty section identification results obtained employing the ST based machine learning tools, for the IEEE thirteen-node test distribution feeder.
Table 13. Faulty section identification results obtained employing the ST based machine learning tools, for the IEEE thirteen-node test distribution feeder.
Applied Fault InformationST Based MLP-NN Predicted Faulty Section
TypeNodeFaulty Section
AG65233
BG67166
CG61111
ABG69255
BCG64688
CAG68044
ABCG65099

Share and Cite

MDPI and ACS Style

Shafiullah, M.; Abido, M.A.; Abdel-Fattah, T. Distribution Grids Fault Location employing ST based Optimized Machine Learning Approach. Energies 2018, 11, 2328. https://doi.org/10.3390/en11092328

AMA Style

Shafiullah M, Abido MA, Abdel-Fattah T. Distribution Grids Fault Location employing ST based Optimized Machine Learning Approach. Energies. 2018; 11(9):2328. https://doi.org/10.3390/en11092328

Chicago/Turabian Style

Shafiullah, Md, M. A. Abido, and Taher Abdel-Fattah. 2018. "Distribution Grids Fault Location employing ST based Optimized Machine Learning Approach" Energies 11, no. 9: 2328. https://doi.org/10.3390/en11092328

APA Style

Shafiullah, M., Abido, M. A., & Abdel-Fattah, T. (2018). Distribution Grids Fault Location employing ST based Optimized Machine Learning Approach. Energies, 11(9), 2328. https://doi.org/10.3390/en11092328

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop