Next Article in Journal
An Irregular Graph Based Network Code for Low-Latency Content Distribution
Next Article in Special Issue
Evaluation of the Possibility of Identifying a Complex Polygonal Tram Track Layout Using Multiple Satellite Measurements
Previous Article in Journal
Iterative Learning Control for a Soft Exoskeleton with Hip and Knee Joint Assistance
Previous Article in Special Issue
Window-Modulated Compounding Nakagami Parameter Ratio Approach for Assessing Muscle Perfusion with Contrast-Enhanced Ultrasound Imaging
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Distribution Transformer Parameters Detection Based on Low-Frequency Noise, Machine Learning Methods, and Evolutionary Algorithm

1
Department of Computer Science and Automatics, University of Bielsko-Biala, 43-309 Bielsko-Biala, Poland
2
Institute of Electric Power Engineering and Renewable Energy, Opole University of Technology, 45-758 Opole, Poland
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(15), 4332; https://doi.org/10.3390/s20154332
Submission received: 6 July 2020 / Revised: 29 July 2020 / Accepted: 1 August 2020 / Published: 4 August 2020

Abstract

:
The paper proposes a method of automatic detection of parameters of a distribution transformer (model, type, and power) from a distance, based on its low-frequency noise spectra. The spectra are registered by sensors and processed by a method based on evolutionary algorithms and machine learning. The method, as input data, uses the frequency spectra of sound pressure levels generated during operation by transformers in the real environment. The model also uses the background characteristic to take under consideration the changing working conditions of the transformers. The method searches for frequency intervals and its resolution using both a classic genetic algorithm and particle swarm optimization. The interval selection was verified using five state-of-the-art machine learning algorithms. The research was conducted on 16 different distribution transformers. As a result, a method was proposed that allows the detection of a specific transformer model, its type, and its power with an accuracy greater than 84%, 99%, and 87%, respectively. The proposed optimization process using the genetic algorithm increased the accuracy by up to 5%, at the same time reducing the input data set significantly (from 80% up to 98%). The machine learning algorithms were selected, which were proven efficient for this task.

1. Introduction

The transformer is a passive electrical device that transfers electrical energy from one electrical circuit to another, or multiple circuits. The transformer is composed of the main parts: The primary and secondary winding wound around the same core and the air or oil cooling system. Noise emitted by the transformer is a vibrio-acoustics problem. The acoustic vibrations of a transformer can be generated by the following main phenomena [1,2]:
  • Coil vibrations depending on the current amplitude and winding clamping compression;
  • Core vibration depending on magnetostriction or loosening of core clamping;
  • Air circulation caused by fans; and
  • Work of the pumps circulating the insulation oil.
The process of the generation of noise from the vibration of a transformer is shown in Figure 1.
In transformers, the load noise is predominantly produced by axial and radial vibration of the windings. Load noise can also be caused by vibrations in the transformer tank walls and magnetic shields due to the electromagnetic forces produced by the load currents. These electromagnetic forces are proportional to the square of the load currents. The frequency of load noise is usually twice the power frequency. In some cases, the natural mechanical frequency of winding clamping systems may tend to resonate with electromagnetic forces, thereby severely intensifying the load noise. Transformer cores are constructed by stacking layers of thin iron laminations, separated from its neighbors by a thin non-conducting layer of insulation. When the core becomes magnetized, the magnetic field acts between the adjacent plates, stretching and squeezing the adhesive and insulation between them. A transformer is magnetically excited by an alternating voltage and current so that it becomes extended and contracted twice during a full cycle of magnetization. This change in dimension is independent of the direction of magnetic flux, occurring at twice the line frequency. The main source of heat generation in transformers is caused by copper loss in the windings and core. This heat is often removed by cooling fans, which blow air over radiators or coolers. Noise produced by cooling fans usually contribute more to the total noise for transformers of a smaller rating and for low-induction transformers. Cooling equipment noise typically dominates the very low- and very high-frequency ends of the sound spectrum, whereas the core noise dominates in the intermediate range of frequencies between 100 and 600 Hz [3,4,5,6,7].
For a person with normal hearing, the human hearing range starts low at about 20 Hz. On the other side of the human hearing range, the highest possible frequency heard without discomfort is 20 kHz. Ultrasound is sound waves with frequencies higher than the upper audible limit of human hearing (above 20 kHz). In this description, we focus on the lower end of the frequency spectrum. We are interested in infrasound and low-frequency noise; see Figure 2. The range of analyzed frequencies is up to 200 Hz, taking under consideration body resonance infrasound.
The acoustic emission depends on the anomalies, age, and rated power of the machine. For this reason, the analysis of the generated sound, especially within the spectrum of low-frequency noise, can be useful in determining the rated parameters of the transformer and its diagnostic parameters. It is worth noting that the measurement of low-frequency signals is carried out non-invasively and during normal operation of the distribution transformer.
Our research was extended to remove drawbacks of the previous method [8], and to develop an automatic method for determining the technical parameters and diagnostics of distribution transformers based on the analysis of the characteristics of their low-frequency signals. Using low-frequency signal allows measurement of the device from a distance (50 m). This research presents a method for detecting the base parameters of distribution transformers (type and rated power).

2. State-Of-The-Art

Transformers can vary, from a miniature high-frequency audio transformer to a large power transformer, but the operating parameters are the same. These can be divided into eight groups and are posted on the nameplate of any transformer of significant size: Nominal apparent power (VA rating), cooling, transformer rating frequency, voltage, phase, connections, and taps [9]. Based on the low-frequency characteristics of the transformer, we are able to distinguish between the transformer’s operating status, its cooling type, and its apparent power.
In our first study [10], we tried to ascertain the emission level of low-frequency signals generated by distribution transformers at rated conditions. The medium-voltage devices (indoor and overhead type and variety apparent powers) were under study.
The results of the study showed that distribution transformers are a source of infrasound and low-frequency signals. The results demonstrated the similarity in the shape of waveforms of averaged amplitude spectra and time-frequency changes. The waveforms are characterized by relatively dynamic (exponentially) decreasing values of registered sound pressure, which occurs with an increase in the frequency in the range from 10 to 100 Hz [10,11].
Research of a similar nature was conducted for the problem of noise from electrical infrastructure. Piana et al. [12] claimed that the low-frequency disturbance occurs for the tested transformer at two harmonics of for each of two frequencies. Other scientists proposed prognostic and system health management (PHM) for power transformer fault diagnosis. Potential uses for PHM is a condition-based maintenance. This system presents opportunities for the detection of mechanical failures, or for system life cycle management. Li et al. [13] presented the study of a power transformer fault diagnosis using a machine learning-based method with a neural network model. The proposed method uses dissolved gas analysis (DGA) as input data. A frequently used method in the diagnosis of oil-filled power transformers is a partial discharge (PD) detection using an acoustic emission (AE) technique. Many cases of power transformer breakdowns are related to insulation system failures, which might have been caused by the high activity of partial discharges [14]. Kunicki et al. [15] proposed a method for detecting defects of power transformers. This method is based on machine learning classification of selected faults. In this case, input data is AE measurement. The classification process consists of two parts: The first part checks whether the source of the emitted signal is partial discharge or another AE source while the second part allows the identification of the specific AE source type.
The aforementioned research is focused on the life cycle of a distribution transformer measured in an isolated environment, where the external noise influence is minimalized. The classification is based on various machine learning algorithms (ML), which finds a pattern in data based on expert feedback. Therefore, in [12,13,14,15], the supervised learning [16] was used to train a model for various applications. However, in this research, the analyzed data were gathered from a significant distance, thus unsupervised learning was used [17] to find anomaly in the background noise, the source of which was unknown. In the case of maintenance operation issues [14], reinforcement learning [18] was used to find optimum actions for a given operation status of a transformer. In this paper, the proposed method utilizes both supervised and unsupervised learning. The supervised learning is used to build a classification model based on evolution strategy and state-of-the-art ML algorithms: k nearest neighbors (kNN) [19], naive Bayes classification (Bayes) [20], support vector machine (SVM) [21,22], random forests [23,24], and neural networks [25,26,27,28]. The unsupervised learning is applied to tackle the background noise.
The accuracy of the constructed model (using the ML algorithm) depends not only on the algorithm but also the type of input data and its representation. Thus, many models use a preprocessing method, which transfers data into a new variable space [17]. The transformation can be a simple operation as scale transformation (e.g., to the dB scale) or a more complex one, when the nature of data changes [29]. The principal component analysis [29,30,31,32] and canonical correlation analysis [33,34] are commonly used methods for TS.
The neural networks (especially deep ones), due to the variety of their structures, are both classification and pattern recognition tools [25,26]. Using multiple hidden layers allows the creation of linear and non-linear models. A drawback of this approach is that the training procedure requires a large amount of data and computation power to obtain a high model accuracy. Moreover, finding the optimum set of weights in the case of a multiple hidden layer structure is an NP-complete problem [35]. Thus, a substitute for classic multiple layer perceptron networks was proposed (deep neural network), in which the network is divided into layers with specific functions [28,32]. Significant results using deep neural networks have led them to be the most commonly employed classifiers in machine learning [35,36].
The proposed method is based on two sources of data (transformer sound and background sound gathered in various locations). Thus, these two sources of the sound are considered as independent ones. However, each registered series of sound according to [17,37] is characterized by significant sequential correlations and should be represented as temporal features. There are several methods fitting for this type of data: Hidden Markov [38], sliding window [39], Kalman filter [40], random fields [41], recurrent neural networks [42], and the Welch method. An extended analysis of the methods can be found in [43,44]. In this study, which is an extension of [8], the same Welch method [45] was used to retain the consistency of results. Furthermore, this method was used with success for noise analysis in [11,12].
The previous study [8] proposed a method to automatically detect a known working transformer in close vicinity (50 m). Low-frequency noise generated by transformers (two indoor and two overhead ones) was registered by a dedicated sensor from a distance of 50 m and then classified using the proposed machine learning method. It is worth noting that the research was performed in a real environment. The method used an exhaustive search and Bayesian optimization to find a frequency interval that gave the best detection results. The drawback of the previous method [8] was that only one interval could be searched at a time, while the results showed that depending on the method, various intervals were selected near the following frequencies: 2, 50, and 100 Hz. Despite the drawbacks of the method, a 99% accuracy was obtained for the state of the transformers (on/off) using the random forest, KNN, and naïve Bayes methods.

3. Proposed Method

In the research for this paper, the frequency was extended to 200 Hz as it was proven that most information was stored close to 100 Hz and a higher frequency was not examined. Furthermore, the research was performed for one interval at a time while several harmonics were found (1, 50, and 100 Hz). To provide sounder results, the transformer database was significantly extended (from 4 to 16) and several of its parameters were researched. Based on the initial observation and previous research [8], a method was proposed, which finds the optimum frequency representation for distribution transformer features: Model, type, and apparent power. The overview of the method, with a background profile analysis, is presented in Figure 3.

3.1. Preprocessing Stage

The input data of the method is a sound registered as a time series (X) by a dedicated sensor node. The sensor measures the transformer noise with background noise (Xt) or only background noise (Xb) if the transformer is not present in the vicinity or it is turned off. Each registered time series X is a sequence of sound pressure values xi, where i defines its order.
The samples, registered for each distribution transformer (Xt), are converted using the Welch method [8,42] to obtain its frequency representation ( F x t ). The obtained spectra F x t = [f1, f2, , fm] represent the low-frequency spectrum in the range from 2 to 200 Hz with the maximum considered resolution df equal to 0.125 Hz (thus m = 1585). The df value lower than 0.125 Hz would require a significant recording time (over 10 s a sample) and as it was proven in this research that the higher resolution did not increase the classification accuracy. The spectra samples are represented as Ft set (Ft = { F x t ,   x = 1 ,   ,   t m a x }, where tmax is a number of samples). Each vector F x t is described by basic transformer parameters (C = {c1,c2,c3}), i.e., transformer model (c1), its type (c2), and apparent power value (c3). The characteristic spectra registered in the vicinity of two transformers are presented in Figure 4.
Each plot in Figure 4a,b presents measurements of the noise characteristic in the vicinity of the same transformer. Measured values for the specific frequency (marked with dashed circles), single peaks (marked with crosses), as well as changes of the amplitude within the whole spectrum can vary significantly. This is caused by the measured distance of 50 m, where other sources of noise are registered as well. Moreover, the transformer load, as presented in [6], can also influence the value of noise, especially in the 50 Hz area. Therefore, an additional source of data was used to identify the anomalies, which originate from analyzed transformers. The background noise in multiple localizations was registered (denoted as F x b ,   x = 1 ,   ,   b m a x ), where the noise is not biased by a transformer noise. The registered background characterizes interiors, fields, forests, and the home environment. An example of the background is presented in Figure 5.
The analyzed background noise (without transformers in vicinity) shows that its characteristic are not constant and also change in time. It is characterized by the same type of variations; however, it is registered more frequently for specific frequencies (e.g., 75 Hz). In case of indoor background, registered in Figure 5b, additional noise was registered for the 50 Hz harmonics and near 10 Hz. This noise is generated by electric devices, which are part of the production areas. It is nearly impossible to gather background characteristic for each transformer, because they are part of the energy infrastructure and they cannot be turned off easily. Therefore, the gathered backgrounds (Figure 5) are characteristic for a specific area, not for a specific transformer. Moreover, as it is shown in Figure 4 and Figure 5, the background changes with time due to temporal occurrences like the influence of constructions, large objects, or vehicles.
Initial work was conducted to find a universal background using the following estimates: Mean, median average, and maximum and minimum value; however, this approach generated worse results. Therefore, based on the analysis, we propose to define several background characteristics and adopt them for each spectrum based on the similarity measure. The fuzzy c-means algorithm [46] was used to find characteristic backgrounds and then, based on the similarity to the analyzed spectra, the appropriate one was selected. To select the background Fb for transformer spectra F x t , the following procedure (Algorithm 1) is given:
Algorithm 1
  • V = cmean(Fb,vmax), where: cmean-fuzzy c-mean function [42], vmax-number of prototypes,
  • mt = mean( F x t );
  • mxw = 0; minVal = INF; minid = −1;
  • for j = 1:vmax
  • mx = mean(V(j)));
  • val1 = mean((( F x t −mt)−V(j)−mx).^2.^0.5);
  • val2 = mean( F x t )−mean(V(j));
  • if (val1 < minVal && val2 > 0)
  • minVal = val1;
  • minid = j;
  • mxw = mx;
  • end
  • end
  • result = F x t −V(minid).
Algorithm 1 generates a set of characteristic backgrounds denoted as V based on background data (line 1). Then, the most similar background noise V(j) is searched for in the analyzed F x t spectrum (lines 4–10). The V spectra are compared based on similarities for the same frequencies (val1) as well as its average value (val2). Figure 6 presents the similarity calculation process. At first, the average value for each potential background is calculated as presented in Figure 6a. The difference between average values (denoted as val2) is defined if the background noise is stronger than the registered transformer noise. Only those prototypes V(j) are considered, in which the average noise is lower than the analyzed spectrum, F x t . Next, the potential background spectra shapes are compared according to formula in line 6. The result of the subtraction is presented in Figure 6b. The obtained mean value of this substation is treated as a similarity measure. In the example presented in Figure 6, the V(1) prototype will not be selected because its mean noise level is higher (val2 = −50) and its similarity measure (val1) is lower than in the case of V(2). Thus, the V(2) prototype is selected as the F x t background. The background is subtracted from the spectrum (line 14). The operation is performed for every spectrum Ft = { F x t ,   x = 1 ,   ,   t m a x } . Then, the input set Ft with the subtracted background is used to find the optimum frequency interval representation in the next step.

3.2. Training Stage Using ML and Evolutionary Algorithms

The search for optimum data representation (frequency space and its resolution) is performed using the evolutionary algorithm. Two state-of-the-art evolution methods, genetic algorithm (ga) and particle swarm optimization (PSO), were applied for each of the C features separately.
The first method, due to the discrete search space, was based on integer programming. In this implementation, special creation, crossover, and mutation functions cause variables to be integers [47]. The implemented genetic algorithm attempts to minimize a penalty function, which is combined with binary tournament selection to select individuals for subsequent generations [48].
The PSO algorithm is based on the implementation proposed by Kennedy and Eberhart [49], using modifications suggested in Mezura-Montes and Coello Coello [50] and in Pedersen [51]. In contradiction to the ga algorithm, it creates the initial particles, and assigns initial velocities to them. Then, it evaluates the objective function at each particle location and determines the best (lowest) function value and the best location. Finally, it chooses new velocities, based on the current velocity, the particles’ individual best locations, and the best locations of their neighbors. Initially, the particles are uniformly distributed within bounds.
Both methods use a similar population/swarm size parameter equal to 200 as well as a total error function equal to 1 × 10−4.
The searched frequency space is described as an ordered set of frequency intervals described by a ti tuple. The initial number of intervals n is equal to 1 and its value increases with each iteration. Each tuple ti = [li, hi, si], i = 1, , n, where: li = [2, 200], hi = [2, 200], li < hi, and hi < li+1 define respectively the lower (li) and upper (hi) frequency bound and s i N —defines the sample resolution for the given tuple. The described constraint allows limitation of the searching space to a considered frequency interval [2, 200] and ensures that intervals will not overlap.
Using evolution algorithms, a population of potential vectors on = {ti, i = 1, , n} is selected. With each generation, the on is selected, which has the lowest error ratio calculated as the misclassification accuracy using cross-validation results. Additionally, the secondary aim of optimization is to find the minimal set of frequency, which did not influence the result significantly. Thus, the following fitness function (ff) was proposed:
f f ( F t , o n ) = min ( c v ( M L ( F t , o n ) ) + i = 1 n [ h i l i + 1 s i ] 1000000
where ML is a function performing training for data and cv-return cross-validation error in the range [0, 1].
Using this approach, we ensure that the result with the highest accuracy will be selected, while the results with less samples will be favored in case of a comparable accuracy level. Taking under consideration the maximum number of samples, the value fitness function will be changed by 0.0015.
Several optimization algorithms as well as ML methods were analyzed to find an optimum solution. In this section, all used algorithms will be described.
The M model is a result of the selected ML method and data provided as a Ft vector and interval set on. The following methods were researched to find the optimum classifier:
  • kNN model, where class is determined by the k closest vectors in a defined frequency space. As a measure of distance, the classic Euclidean distance was used with an initial k value equal to 5;
  • Bayes approach, where a family of probabilistic classifiers was applied according to [20], thus no additional parameters were needed. It was assumed that each frequency characteristic is independent (Bayes theorem);
  • Multivariable support vector machines, where binary learners were used to train the characteristics of each transformer and its parameters. In contrast to the radial basis function 45 applied in [8], in this research, multi-linear SVM was used, which finds a hyperplane that is a linear function of each input feature and the rest of the features. The implementation was adopted from [18,19,52];
  • Random forest, where 10 trees create a forest, was used make the method more robust [23]. The result of a class is determined by voting. The parameter value was selected empirically;
  • Neural network, where the multilayer perceptron network (MLP) was selected as the architecture. The MLP was selected to reduce the computation complexity of each iteration for the evolution algorithm. Several architectures were researched. Finally, 2 hidden layers and 20 neurons per layer were used to take the non-linear characteristic of the analyzed data into consideration [26].

3.3. Detection Stage

The detection stage is used to verify a proposed model and can also be applied for detecting transformer parameters C in the considered area. The registered noise is preprocessed in a similar way as data in the preprocessing stage. The sample is transformed to the frequency domain using the Walsh method and algorithm 1 is performed. However, in this case, the generated prototypes in the first stage are used. The preprocessed data are then processed in parallel by three ML models generated for each of three C features. As the result, the basic transformer parameters are determined.

4. Results and Discussion

Measurements were made using specialized equipment from Brüel & Kjær (company name details: Brüel & Kjær Sound & Vibration Measurement A/S, DK-2850 Nærum, Denmark). The system consists of a ½-inch free-field microphone type 4190, a preamplifier type 2669 L, and a digital signal meter with registration function LAN-XI type 3050-A-60 from Brüel & Kjær. The connected system and its block diagram are shown in Figure 7.
The microphone was designed for very precise measurements in the free field. Its lower cut-off frequency is 1.2 Hz and the transmission characteristic is linear within ±3 dB from 1.2 Hz to 20 kHz. This microphone is characterized by high sensitivity (50 mV/Pa) and a dynamic range from 15 to 148 dB. The lower range of the measured frequencies can be defined by the user using high-pass filters (0.7 or 7 Hz).
The system was managed from the personal computer using PULSE LabShop application version 15.1.0. This is dedicated software that defines all operating parameters, records the measured signals, and preprocesses and visualizes them after measurements in offline mode.
The measurements were taken in a continual process over several hours, with a sampling frequency of 51.2 kHz (sampling for the whole human hearing range). Additionally, all measurements were carried out far from major roads and motorways, which are considered to be sources of low-frequency noise. The minimal sample length was defined by the Welch window size and it was defined as 10 s.
The research was conducted using 16 distribution transformers and 6 backgrounds for various areas and types (indoor and overhead, dry-type, and oil-type transformers and their apparent power in the range of 100–2500 kVA). All tested transformers reduce the voltage from 15 to 0.4 kV with a mains frequency 50 Hz. A detailed description of the tested transformers is provided in Table 1.
Initial work was performed to find an optimum background representation. It was achieved by finding a set of background prototypes giving the best results in terms of classification. The classification considered the following basic transformer parameters: Specific transformer model (c1), transformer type (indoor/outdoor) (c2), and its apparent power (c3). The results are presented in Table 2. The values were obtained using the kNN classifier.
The results show that the background selection influences c3 more than c1 and c2. This is caused by the fact that background noise has a bigger impact on the magnitude of low-frequency noise registered and can influence the classification. The initial results confirmed that the characteristics of the background noise for air transformers and indoor transformers are significantly different and allow a decrease of the error by more than 1% even at the preprocessing stage. The indoor transformers are usually installed in urbanized areas, thus more sources of noise are present. Then, the proposed method was applied for the kNN classifier with both the pso and ga training methods. Additionally, the variant with and without background subtraction (application of Algorithm 1) was tested. The result of method training is presented in Figure 8.
The value n equal to 0 represents the value without interval optimization. The process of finding the optimum interval using the evolution strategy, from the first interval (n equal to 1) reduces the method error on average by 4% in the case of detecting a model of a transformer, five times in the case of the transformer type, and by 5% in the case of the transformer apparent power value. The further increase of the number of intervals steadily decreases the error up to n equal to 3. The best results were achieved for n equal to 6. Above this threshold, the method is over trained and its effectiveness starts to fall. The results confirmed that the used PSO method outperforms the tested ga algorithm by 2% on average for all analyzed features. Additionally, the background data decreases the error in the case of feature c3. Further analysis showed that the optimization of the data using the background gives a slightly worse result (0.5%) than raw data in the case of c1 feature. The difference is caused by the characteristic background of a specific area and the tendency to classify the transformer model based on the area background and not transformer noise. Similar training was performed for the remaining ML algorithms. The comparative result for the proposed method is presented in Figure 9 for the PSO algorithm and ML methods as shown in the legend.
The proposed algorithm of interval selection with background subtraction decreased the error in the case of all analyzed ML models. The method was effective in the case of all features; however, the biggest optimization (in percent scale) was noticed in the case of transformer model detection (c1) and transformer apparent power detection (c3). In the case of the type of transformer detection, the detection error was close to zero, thus a significant improvement was noticed only in the case of the Bayes classifier. The decreased error can be noticed at the first interval (n = 1), which supports the results obtained in previous research [8], but further decreases can be observed with the interval number increase, obtaining the optimum value at n = 5. This value allows optimal detection of all harmonics of 50 Hz generated by a transformer, and the infrasound interval, which proved to be vital frequencies for transformer classification. The biggest decrease in the error rate can be noticed in case of simple classifiers, such as kNN, Bayes, or random forest. On the other hand, in the case of a complex non-linear SVM model, the decrease is equal to 2% on average. The proposed model not only allows an increase of the detection accuracy but also allows a decrease of the data usage by selecting intervals and the resolution at which the data are vital for classification. The research shows that various ML models require various data sizes; nevertheless, some frequencies’ intervals are characteristic for all classifiers and were repeated in all models. Their ranges are presented in Figure 10.
The analysis definitely shows that in the case of transformer model detection, nearly all frequencies are vital to improve detection. Only frequencies near 70 and 170 Hz contain less information. The results confirm the observations shown in Figure 5, where the external background noise was strong for those intervals. It is worth noting that detection of a specific transformer is most biased by additional background noise (like other electric devices). In the case of type discrimination (c2 feature), the vital information can be obtained for frequencies until 25 Hz and near 200 Hz. These frequencies are sufficient to precisely define a transformer type and thus are the most robust to background noise. Finally, in the case of the power of the transformer, the vital information is stored near frequencies of 25, 50, 100, 150, and 200 Hz. This result is consistent with previous results for the frequency range 2–100 Hz [8] and the results found in [1,2,3,10,11]. Finally, in most cases, the frequency interval of 1 Hz is sufficient for precise classification for each feature, which strongly decreases the data usage. It is worth noting that more precision is required near 20 Hz (0.5 Hz). The final results for classification and data reduction using the proposed method are presented in Table 3.
The research presented in [12,13,14,15] proposes methods that precisely detect transformer flaws and discharge with an accuracy above 90%; nevertheless, the sensors have to be placed in close vicinity of a transformer. The proposed method, using low-frequency sound, allows the detection of basic parameters of various transformers from a distance of 50 m with an accuracy of at least 80%. It is worth noting that this type of sound has a strong influence on health [1,2] so this type of simple detection from a distance verifies the parameters of a transformer from a distance. The presented results show that various data sets need to be used depending on the task. The data reduction was achieved by decreasing the frequency resolution up to 1 Hz. An additional reduction is characteristic for a specific feature. In the case of indoor/outdoor transformer detection, only two frequency ranges were vital, thus the achieved reduction is high (98%). In the case of the specific transformer model, it was crucial to analyze a wide frequency range to find dissimilarities, thus the reduction was the smallest and depended on the interval resolution reduction. It is worth noting that the SVM ml MODEL requires bigger data precision (only 50% reduction); however, it obtained the best accuracy. The other classifiers obtained lower results; however, the data reduction was more significant, e.g., kNN obtained an 87% reduction at the cost of an accuracy decrease by 3%. In the case of the transformer type, all classifiers obtained a high 99% accuracy; however, the kNN outperformed the other classifiers in accuracy and data reduction. Finally, the power of a transformer can be precisely detected using kNN, SVM, and the MLP network. In the case of the SVM classifier, the accuracy was higher by 0.7%; however, this was at the cost of a lower data reduction. On the other hand, the MLP and kNN classifier obtained a comparable accuracy and data reduction.

5. Conclusions

The paper proposed a method to detect distribution transformer parameters from a distance, without the need of installing a multiple sensor array on the transformer. The proposed method uses genetic algorithms to find the optimum frequency representation for detecting a transformer model, its type, and rated power. In every case, the proposed method allowed an increase of the accuracy by 5% on average while finding the optimum intervals and their resolution and decreasing the input data set from 50% up to 98% depending on the task. The research confirmed that in the case of generated power, its harmonics are based on 50 Hz; however, an important interval can also be found near 20 Hz, which are infrasound signals. The specific transformer model can generate the noise in all of the analyzed spectrum; thus, in the case of all machine learning algorithms, the reduction of the interval caused a decrease in the detection accuracy. Finally, the type of the transformer due to its different characteristics (power level and interval construction) obtained a near 100% accuracy while a significant data reduction (over 90%) was achieved. The model, with the SVM classifier, can be applied for solutions requiring maximum accuracy, while models based on kNN and MLP can be applied in edge sensors, due to the significant data reduction.
Further research will be focused on anomaly detection during the operation of transformers based on various characteristics. This method will allow analysis of the technical condition of the transformer based on the measurement of its low-frequency signals made online (without switching off).

Author Contributions

Conceptualization, D.J.; Methodology, M.B.; Software, M.B., D.J., T.B.; Validation, T.B., D.J., M.B.; Formal Analysis, T.B.; Investigation, D.J., M.B., T.B.; Writing-Original Draft Preparation, M.B. and D.J.; Writing—Review and Editing, M.B., T.B. and D.J. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ying, L.; Wang, D.; Wang, J.; Wang, G.; Wu, X.; Liu, J. Power Transformer Spatial Acoustic Radiation Characteristics Analysis under Multiple Operating Conditions. Energies 2018, 11, 74. [Google Scholar] [CrossRef] [Green Version]
  2. Bartoletti, C.; Desiderio, M.; Di Carlo, D.; Fazio, G.; Muzi, F.; Sacerdoti, G.; Salvatori, F. Vibro-Acoustic Techniques to Diagnose Power Transformers. IEEE Trans. Power Deliv. 2004, 19, 221–229. [Google Scholar] [CrossRef]
  3. Zou, L.; Guo, Y.; Liu, H.; Zhang, L.; Zhao, T. A Method of Abnormal States Detection Based on Adaptive Extraction of Transformer Vibro-Acoustic Signals. Energies 2017, 10, 2076. [Google Scholar] [CrossRef] [Green Version]
  4. Bouayed, K.; Mebarek, L.; Lanfranchi, V.; Chazot, J.-D.; Marechal, R.; Hamdi, M.-A. Noise and vibration of a power transformer under an electrical excitation. Appl. Acoust. 2017, 128, 64–70. [Google Scholar] [CrossRef]
  5. Masti, R.S.; Desmet, W.; Heylen, W. On the influence of core laminations upon power transformer noise. In Proceedings of the International Conference on Noise and Vibration Engineering (ISMA), Leuven, Belgium, 20–22 September 2004; pp. 3851–3861. [Google Scholar]
  6. Girgis, R.S.; Bernesjo, M.; Anger, J. Comprehensive analysis of load noise of power transformers. In Proceedings of the 2009 IEEE Power & Energy Society General Meeting, Calgary, AB, Canada, 26–30 July 2009; pp. 1–7. [Google Scholar]
  7. Zawieska, W.M. The active control issues related to the noise generated by power transformers. Mechanics 2005, 24, 155–161. [Google Scholar]
  8. Jancarczyk, D.; Bernaś, M.; Boczar, T. Classification of Low-frequency Signals Emitted by Power Transformers Using Sensors and Machine Learning Methods. Sensors 2019, 19, 4909. [Google Scholar] [CrossRef] [Green Version]
  9. Orosz, T. Evolution and modern approaches of the power transformer cost optimization methods. Period. Polytech. Electr. Eng. Comput. Sci. 2019, 63, 37–50. [Google Scholar] [CrossRef] [Green Version]
  10. Jancarczyk, D.; Bernas, M.; Sidzina, M.; Janusz, J. Comparative Analysis of Infrasound Noise Emitted by Power Transformers. In Scientific Papers Nr 59; Faculty of Electrical and Control Engineering Gdansk University of Technology: Gdańsk, Poland, 2018. (In Polish) [Google Scholar]
  11. Jancarczyk, D. Research and Analysis of Infrasound Noise Emitted by Power Transformers. Electr. Eng. Pozn. Univ. Technol. Acad. J. 2018, 95, 153–161. (In Polish) [Google Scholar]
  12. Piana, E.A.; Roozen, N.B. On the Control of Low-Frequency Audible Noise from Electrical Substations: A Case Study. Appl. Sci. 2020, 10, 637. [Google Scholar] [CrossRef] [Green Version]
  13. Li, A.; Yang, X.; Dong, H.; Xie, Z.; Yang, C. Machine Learning-Based Sensor Data Modeling Methods for Power Transformer PHM. Sensors 2018, 18, 4430. [Google Scholar] [CrossRef] [Green Version]
  14. Sikorski, W. Development of Acoustic Emission Sensor Optimized for Partial Discharge Monitoring in Power Transformers. Sensors 2019, 19, 1865. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Kunicki, M.; Wotzka, D. A Classification Method for Select Defects in Power Transformers Based on the Acoustic Signals. Sensors 2019, 19, 5212. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Barber Bayesian, D. Reasoning and Machine Learning; Cambridge University Press: Cambridge, UK, 2012. [Google Scholar]
  17. Mahdavinejad, M.S.; Rezvan, M.; Barekatain, M.; Adibi, P.; Barnaghi, P.; Sheth, A.P. Machine learning for internet of things data analysis: A survey. Digit. Commun. Netw. 2018, 4, 161–175. [Google Scholar] [CrossRef]
  18. Murphy, K.P. Machine Learning: A Probabilistic Perspective; MIT Press: Cambridge, MA, USA, 2012. [Google Scholar]
  19. Jagadish, H.V.; Ooi, B.C.; Tan, K.L.; Yu, C.; Zhang, R. Idistance: An adaptive b+-tree based indexing method for nearest neighbor search. ACM Trans. Database Syst. (TODS) 2005, 30, 364–397. [Google Scholar] [CrossRef]
  20. Zhang, H. The optimality of naive bayes. Am. Assoc. Artif. Intell. 2004, 1, 3. [Google Scholar]
  21. Cortes, C.; Vapnik, V. Support-vector networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  22. Scholkopf, B.; Smola, A.J. Learning with Kernels: Support Vector Machines, Regularization, Optimization, and Beyond; MIT Press: Cambridge, MA, USA, 2001. [Google Scholar]
  23. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
  24. Breiman, L. Bagging predictors. Mach. Learn. 1996, 24, 123–140. [Google Scholar] [CrossRef] [Green Version]
  25. Glorot, X.; Bengio, Y. Understanding the difficulty of training deep feedforward neural networks. In Proceedings of the 13th International Conference Artificial Intelligence and Statistics (AISTATS), Sardinia, Italy, 13–15 May 2010; pp. 249–256. [Google Scholar]
  26. Eberhart, R.C. Neural Network PC Tools: A Practical Guide; Academic Press: Cambridge, MA, USA, 2014. [Google Scholar]
  27. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  28. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  29. Hotelling, H. Analysis of a complex of statistical variables into principal components. J. Educ. Psychol. 1933, 24, 417. [Google Scholar] [CrossRef]
  30. Jolliffe, I. Principal Component Analysis; Wiley Online Library: Hoboken, NJ, USA, 2002. [Google Scholar]
  31. Abdi, H.; Williams, L.J. Principal component analysis. Wiley Interdiscip. Rev. Comput. Stat. 2010, 2, 433–459. [Google Scholar] [CrossRef]
  32. Bro, R.; Smilde, A.K. Principal component analysis. Anal. Methods 2014, 6, 2812–2831. [Google Scholar] [CrossRef] [Green Version]
  33. Hotelling, H. Relations between two sets of variates. Biometrika 1936, 28, 321–377. [Google Scholar] [CrossRef]
  34. Bach, F.R.; Jordan, M.I. Kernel independent component analysis. J. Mach. Learn. Res. 2002, 3, 1–48. [Google Scholar]
  35. Blum, W.; Burghes, D.; Green, N.; Kaiser-Messmer, G. Teaching and learning of mathematics and its applications: First results from a comparative empirical study in england and Germany. Teach. Math. Appl. Int. J. IMA 1992, 11, 112–123. [Google Scholar] [CrossRef] [Green Version]
  36. Schmidhuber, J. Deep learning in neural networks: An overview. Neural Netw. 2015, 61, 85–117. [Google Scholar] [CrossRef] [Green Version]
  37. Bernas, M.; Placzek, B. Period-aware local modelling and data selection for time series prediction. Expert Syst. Appl. 2016, 59, 60–77. [Google Scholar] [CrossRef]
  38. Rabiner, L.R. A tutorial on hidden markov models and selected applications in speech recognition. Proc. IEEE 1989, 77, 257–286. [Google Scholar] [CrossRef]
  39. Sejnowski, T.J.; Rosenberg, C.R. Parallel networks that learn to pronounce english text. Complex Syst. 1987, 1, 145–168. [Google Scholar]
  40. Kalman, R.E.; Bucy, R.S. New results in linear filtering and prediction theory. J. Basic Eng. 1961, 83, 95–108. [Google Scholar] [CrossRef]
  41. Lafferty, J.; McCallum, A.; Pereira, F. Conditional random fields: Probabilistic models for segmenting and labeling sequence data. In Proceedings of the Eighteenth International Conference on Machine Learning, ICML, Williamstown, MA, USA, 28 June–1 July 2001; Volume 1, pp. 282–289. [Google Scholar]
  42. Williams, R.J.; Zipser, D. A learning algorithm for continually running fully recurrent neural networks. Neural Comput. 1989, 1, 270–280. [Google Scholar] [CrossRef]
  43. McCallum, A.; Freitag, D.; Pereira, F. Maximum Entropy Markov Models for Information Extraction and Segmentation. In Proceedings of the 17th International Conference on Machine Learning, ICML, Stanford, CA, USA, 29 June–2 July 2000; Volume 17, pp. 591–598. [Google Scholar]
  44. Baltrušaitis, T.; Ahuja, C.; Morency, L. Multimodal Machine Learning: A Survey and Taxonomy. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 41, 423–443. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  45. Welch, P. The use of fast Fourier transform for the estimation of power spectra: A method based on time averaging over short, modified periodograms. IEEE Trans. Audio Electr. 1967, 15, 70–73. [Google Scholar] [CrossRef] [Green Version]
  46. Bezdec, J.C. Pattern Recognition with Fuzzy Objective Function Algorithms; Plenum Press: New York, NY, USA, 1981. [Google Scholar]
  47. Deep, K.; Singh, K.P.; Kansal, M.L.; Mohan, C. A real coded genetic algorithm for solving integer and mixed integer optimization problems. Appl. Math. Comput. 2009, 212, 505–518. [Google Scholar] [CrossRef]
  48. Deb, K. An efficient constraint handling method for genetic algorithms. Comput. Methods Appl. Mech. Eng. 2000, 186, 311–338. [Google Scholar] [CrossRef]
  49. Kennedy, J.; Eberhart, R. "Particle Swarm Optimization". In Proceedings of the IEEE International Conference on Neural Networks, Perth, Australia, 27 November–1 December 1995; pp. 1942–1945. [Google Scholar]
  50. Mezura-Montes, E.; Coello, C.A.C. Constraint-handling in nature-inspired numerical optimization: Past, present and future. Swarm Evol. Comput. 2011, 1, 173–194. [Google Scholar] [CrossRef]
  51. Pedersen, M.E. Good Parameters for Particle Swarm Optimization; Hvass Laboratories: Luxembourg, 2010. [Google Scholar]
  52. Escalera, S.; Pujol, O.; Radeva, P. On the decoding process in ternary error-correcting output codes. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 32, 120–134. [Google Scholar] [CrossRef]
Figure 1. The process of the generation of noise from the vibration of a transformer.
Figure 1. The process of the generation of noise from the vibration of a transformer.
Sensors 20 04332 g001
Figure 2. The frequency spectrum of sound and its nomenclature.
Figure 2. The frequency spectrum of sound and its nomenclature.
Sensors 20 04332 g002
Figure 3. Proposed method for the detection of transformer parameters.
Figure 3. Proposed method for the detection of transformer parameters.
Sensors 20 04332 g003
Figure 4. The example of spectra registered in the vicinity of: (a) transformer air-type, and (b) interior type.
Figure 4. The example of spectra registered in the vicinity of: (a) transformer air-type, and (b) interior type.
Sensors 20 04332 g004
Figure 5. Noise background for various positions: (a) outdoor background, (b) indoor background.
Figure 5. Noise background for various positions: (a) outdoor background, (b) indoor background.
Sensors 20 04332 g005
Figure 6. Example of the c-means background cluster fitted to a transformer: (a) calculate val2 measure, (b) calculate val1 measure.
Figure 6. Example of the c-means background cluster fitted to a transformer: (a) calculate val2 measure, (b) calculate val1 measure.
Sensors 20 04332 g006
Figure 7. Used measurement system from Brüel & Kjær: (a) microphone, (b) digital meter, and (c) complete system.
Figure 7. Used measurement system from Brüel & Kjær: (a) microphone, (b) digital meter, and (c) complete system.
Sensors 20 04332 g007
Figure 8. The training results for the optimization algorithm number of intervals n and feature c1, c2, and c3.
Figure 8. The training results for the optimization algorithm number of intervals n and feature c1, c2, and c3.
Sensors 20 04332 g008
Figure 9. Training for all ML algorithms and analyzed features.
Figure 9. Training for all ML algorithms and analyzed features.
Sensors 20 04332 g009
Figure 10. The data usefulness for classification and detection.
Figure 10. The data usefulness for classification and detection.
Sensors 20 04332 g010
Table 1. List of the transformers under study.
Table 1. List of the transformers under study.
Transformer NumberManufacturerApparent PowerTransformer TypeCooling Type
Transformer 1Schneider Electric2500 kVAindoor typedry-type
Transformer 2Schneider Electric2000 kVAindoor typedry-type
Transformer 3Schneider Electric1600 kVAindoor typedry-type
Transformer 4Schneider Electric1250 kVAoverhead typeoil-type
Transformer 5Schneider Electric630 kVAoverhead typeoil-type
Transformer 6ABB400 kVAoverhead typeoil-type
Transformer 7ABB400 kVAoverhead typeoil-type
Transformer 8ABB250 kVAoverhead typeoil-type
Transformer 9ABB250 kVAoverhead typeoil-type
Transformer 10ABB250 kVAoverhead typeoil-type
Transformer 11ABB250 kVAoverhead typeoil-type
Transformer 12ABB250 kVAoverhead typeoil-type
Transformer 13ABB160 kVAoverhead typeoil-type
Transformer 14ABB100 kVAoverhead typeoil-type
Transformer 15ABB100 kVAoverhead typeoil-type
Transformer 16ABB63 kVAoverhead typeoil-type
Table 2. The result of finding background representation.
Table 2. The result of finding background representation.
Cross-Validation Error [%]
Error Rate for Number of Vectors (vmax)Specific Transformer (c1),Transformer Type (c2)Transformer Power (c3)
No background23.30.418
123.10.416.1
222.80.415.9
3230.517.22
4240.618
Table 3. Results of the classification for a specific algorithm.
Table 3. Results of the classification for a specific algorithm.
FeatureAlgorithmAccuracy (Based)Accuracy (Proposed Method)Data Reduction (Proposed Method)
Transformer model (c1)kNN77%81%87%
Bayes70%79%70%
Random Forest68%75%91%
SVM83%84%11%
MLP74%77%81%
Transformer type (c2)kNN98.9%99.99%98%
Bayes97%99.99%98.5%
Random Forest98.8%99.8%97.7%
SVM99.8%99.99%97.8%
MLP99.2%99.98%97.1%
Transformer Power (c3)kNN78%87%93%
Bayes73%78%93%
Random Forest74%79.1%95%
SVM85%87.7%50%
MLP79%86.8%92%

Share and Cite

MDPI and ACS Style

Jancarczyk, D.; Bernaś, M.; Boczar, T. Distribution Transformer Parameters Detection Based on Low-Frequency Noise, Machine Learning Methods, and Evolutionary Algorithm. Sensors 2020, 20, 4332. https://doi.org/10.3390/s20154332

AMA Style

Jancarczyk D, Bernaś M, Boczar T. Distribution Transformer Parameters Detection Based on Low-Frequency Noise, Machine Learning Methods, and Evolutionary Algorithm. Sensors. 2020; 20(15):4332. https://doi.org/10.3390/s20154332

Chicago/Turabian Style

Jancarczyk, Daniel, Marcin Bernaś, and Tomasz Boczar. 2020. "Distribution Transformer Parameters Detection Based on Low-Frequency Noise, Machine Learning Methods, and Evolutionary Algorithm" Sensors 20, no. 15: 4332. https://doi.org/10.3390/s20154332

APA Style

Jancarczyk, D., Bernaś, M., & Boczar, T. (2020). Distribution Transformer Parameters Detection Based on Low-Frequency Noise, Machine Learning Methods, and Evolutionary Algorithm. Sensors, 20(15), 4332. https://doi.org/10.3390/s20154332

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop