Next Article in Journal
Performance Evaluation of Deep CNN-Based Crack Detection and Localization Techniques for Concrete Structures
Previous Article in Journal
Identifying and Analyzing Dependencies in and among Complex Cyber Physical Systems
Previous Article in Special Issue
Real-Time Leak Detection for a Gas Pipeline Using a k-NN Classifier and Hybrid AE Features
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Fault Prediction and Early-Detection in Large PV Power Plants Based on Self-Organizing Maps

1
i-EM S.r.l. (Intelligence in Energy Management), 57121 Livorno, Italy
2
Department of Energy, Systems, Territory and Construction Engineering (DESTEC), University of Pisa, 56122 Pisa, Italy
*
Authors to whom correspondence should be addressed.
Sensors 2021, 21(5), 1687; https://doi.org/10.3390/s21051687
Submission received: 3 February 2021 / Revised: 21 February 2021 / Accepted: 23 February 2021 / Published: 1 March 2021
(This article belongs to the Special Issue Fault Detection and Localization Using Electromagnetic Sensors)

Abstract

:
In this paper, a novel and flexible solution for fault prediction based on data collected from Supervisory Control and Data Acquisition (SCADA) system is presented. Generic fault/status prediction is offered by means of a data driven approach based on a self-organizing map (SOM) and the definition of an original Key Performance Indicator (KPI). The model has been assessed on a park of three photovoltaic (PV) plants with installed capacity up to 10 MW, and on more than sixty inverter modules of three different technology brands. The results indicate that the proposed method is effective in predicting incipient generic faults in average up to 7 days in advance with true positives rate up to 95%. The model is easily deployable for on-line monitoring of anomalies on new PV plants and technologies, requiring only the availability of historical SCADA data, fault taxonomy and inverter electrical datasheet.

1. Introduction

1.1. Motivation

The implementation of accurate and systematic preventive maintenance strategies is emerging nowadays as an essential tool to maintain high technical and economic performance of solar photovoltaic (PV) plants over time [1]. Analytical monitoring systems have been installed worldwide to timely detect possible malfunctions through the assessment of PV system performance [2,3,4,5,6,7,8,9,10]. Due to the abundance of relevant data, and the difficulty in modeling many complex aspects of PV plants, statistical methods based on data mining and machine learning algorithms are recently emerging as a very promising approach both for fault prediction and early detection. In particular, the recent development of key enabling technologies and paradigms, most notably Internet-of-Things (IoT)-environments and machine learning algorithms to handle massive quantities of data, have been recently applied to monitoring the functioning of PV systems. A large number of scientific papers have been written to describe how they can be effectively used to timely detect possible malfunctions through the assessment of an indicator performance, and significant works on the topic include papers [11,12,13,14,15,16,17,18,19]. Besides, similar strategies have been also presented in works that tackle wind farms, see [20,21] with the objective of identifying equipment level failures, while in this case fewer works can be found for the counterpart for PV plants [22].

1.2. Paper Contribution

Such an abundance of scientific papers witnesses the interest of the scientific community on this research topic, and the practical importance of developing continuous monitoring algorithms. Indeed, in addition to high customization costs and the need of collecting and transmitting a large number of physical variables, there is a significant interest in developing automatic, non-supervised and accurate methodologies to perform such maintenance strategies, and this is the focus of the present paper. Furthermore, with respect to the aforementioned references, our work does not focus on small-size, usually roof-top-located, PV panels of few kW power, but rather on large-size PV plants that provide power at the scale of several MWs. Most importantly, differently from the previous strategies, we shall show how the maintenance strategies proposed here not only manage to identify possible malfunctioning conditions, but are also able to predict incipient faults a few days in advance from when they occur in practice. In particular, this paper describes a novel and flexible solution for inverter level fault prediction based on a data-driven approach. The model ability to predict or to recognize anomalous patterns and faulty operating conditions is here validated off-line for three different PV plants located in Romania and Greece, for a nominal power of up to 10 MW and a testing window of one year. As we show, the proposed approach has been used in the different plants of different sizes and technologies in the same fashion, and in each case it has proven to provide valuable and accurate failure predictions.

2. Case Studies and Methods

In the paper we shall consider three PV plants, called in the following as plants A, B, and C, respectively, with an installed capacity ranging between 3 and 10 MW, equipped with two different inverter technologies, labelled as 1 and 2, as shown in Table 1. Plant A is located in Romania, whereas plants B and C are in Greece, as shown in Figure 1. Globally, 67 inverter devices have been analysed. More details are given in the following subsections.

2.1. PV Plants Details

Plant A is located in Romania and has an installed capacity of around 10 MW, corresponding to 35 inverter modules with a rated output power either of 385 kW AC or 55 kW AC. In the plant both polycristalline and thin film solar panels are installed. The operating facility is able to produce around 15 million kWh per year, corresponding to the annual energy needs of more than 7500 households, thereby avoiding the emission of over 6800 tonnes of C O 2 into the atmosphere per year.
Plants B and C are located in Greece. Plant B is in the Xanthi region and is composed by strings of thin film solar panels connected to seven inverter modules with a rated ouput power of 385 kW AC, which globally corresponds to an installed capacity of 2.8 MW. On average, it is able to produce about 3.6 million kWh per year. The last considered PV facility is located in the Ilia region and it consists of polycristalline solar panels producing around 6.15 million kWh annually. It is equipped with 25 inverter modules with a rated output power of 183.4 kW AC, which corresponds to an overall installed capacity of 4.9 MW.

2.2. SCADA Data and Alarm Logbooks

The datasets of plants A and B consist of 10 measured signals, as explained in Table 2, with a sampling time ν of 5 min. Measured signals include both electrical (DC and AC sides) and environmental quantities (such as the solar irradiance that is acquired by pyranometers, and temperatures). For plant C, also the internal inverter temperature ( T i n t ) is available. The signals, collected by sensors installed in the PV facility, are stored in a Supervisory Control Furthermore, Data Acquisition (SCADA) system. Data are then transmitted to two higher supervision centers: the Local Control Room of the country of the PV plant, and the World Monitory Room at the headquarters of the power company managing the plant under investigation.
The lengths of the historical datasets are different for the three considered plants, as it is summarized in Table 3, which also shows the number of measured patterns for each dataset. In particular, it is possible to note that the testing period is always at least six months long, and in one case one year long, which is convenient to evaluate the performance of the proposed strategies over different seasons. As it will be described in the greater detail in the following sections, a careful pre-processing stage is required to clean the available data and make them more informative for the training stage.
In addition to the SCADA data, we also exploit the availability of Operation and Maintenance (O&M) logs. In this case, the relevant information consists of the start and end times at which single failure events have been experienced, as well as the specific fault type, and inverter device suffering the failure. Depending on the plant, this information was available either through automatic SCADA logbooks, or through manual logbooks where plant operators manually provided the relevant information. Accordingly, in our model we have also considered the O&M logs, together with the fault taxonomy which is required to associate the manufacturer code with the corresponding failure type, description and severity, which have been used a posteriori to assess the performance of the proposed strategies. In particular, the logbook has been embedded in our model by matching the fault classes listed in the fault taxonomy file to the fault occurrences recorded in the logbooks and associating them with the timestamp of SCADA data. More specifically, a fault of the kth type is assigned to timestamp t n if the following condition occurs:
t s t a r t , k t n t e n d , k
where t s t a r t , k ( t e n d , k ) are the initial (final) instant of the fault event. Once the O&M logs have been discretized consistently with the SCADA data time line, each timestamp t n has been labelled according to the fault code occurring in that instant. Simultaneous fault events at instant t n have been handled according to a prioritization rule, i.e., by labelling t n with the most severe fault code occurring at that instant and, if necessary, the most frequent fault in the day D, with t n D . As a consequence, the resulting time line is labelled with an integer number, one for each timestamp, corresponding to nominal behaviours (label equal to 0) or faulty events (label larger than 0). While in principle, the information of the specific occurred fault was available, yet in this work we only focus on a binary classification problem where the objective is to discriminate faulty and correct working conditions. Thus, we have trained our algorithms to recognize faulty working conditions, and not the specific type of occurred failure.

2.3. Data Pre-Processing

Due to the heterogeneity of the considered physical quantities, the pre-processing stage has been customized specifically for each tag. In particular, the AC power ( P A C ) depends mainly on the solar irradiance (GTI) striking on the PV panel plane and on the environmental temperature ( T a m b ). Statistical outliers corresponding to values of P A C significantly larger than 0 despite low values of the GTI, or viceversa, have been removed by implementing a first-order regression of the unknown underlying function P A C = P A C G T I and removing instances that lied far from a linear approximation:
P A C G T I · m + b G T I · m + b > η ,
where m and b are the slope and the intercept, respectively, of the linear approximation computed by means of a least-squares fitting, and η is the threshold set by a trial and error process, to identify unrealistically far samples. In addition, many signals exhibit a significant number of not regular data, such as missing or “frozen” samples (i.e., instances where the signal measured by the sensor does not change in time), or values out of physical or operative limits, or spikes. Accordingly, a classic procedure of data cleaning has been carried out to avoid training the algorithms with obviously wrong data. In particular, as many electrical and environmental signals exhibit daily patterns, days having a large percentage of missing data have been removed as a whole.

2.4. SCADA Imputation

Since the model, once deployed in practice on-site, must be obviously able to work also in situations of missing online instances, in a “best-effort” fashion (i.e., as well as possible given the obvious difficulties of wrong measurements or wrong data transmission), missing test samples have been imputed according to the classical k-Nearest Neighbors (k-NN) algorithm. More in detail, the training set has been used as the reference dataset, replacing missing data with the nearest neighbors according to the Euclidean distance [23,24].

2.5. Data Detrending and Scaling

As different electrical (e.g., P A C ) and environmental (e.g., GTI) signals exhibit seasonal trends, it is convenient to remove such seasonality trends to prevent biased predictions from occurring. In order to remove the season-dependent variability from input data, a detrending procedure has been applied by following tailored approaches for each variable. In particular, the training data of T m o d have been deseasonalized by means of the least-squares fitting method to infer the best line T f i t against T a m b and selecting only low samples with low GTI to remove the effect of the panel heating due to sunlight:
T ˜ m o d = T m o d T f i t T f i t G T I G T I t h r ,
where
T f i t = m T · T a m b + b T
is the fitting temperature, m T is the regression slope, b T is the intercept and G T I t h r = 100 W/m 2 is a heuristically determined threshold for the solar irradiance to identify “low values of the GTIs” that do not give rise to relevant panel heatings effects (it corresponds to the maximum value of irradiance which does not involve appreciable heating of the module with respect to the ambient temperature).
All the remaining input variables, apart from DC and AC voltages, have been de-trended according to a classical Moving Average (MA) smoothing method to compute the seasonal trend component and applying an additive model for time series decomposition [25,26].
Finally, input data normalization is performed to avoid unbalance between heterogeneous quantities. In particular, we used the standard normalization [26].

3. Methodology

The proposed approach consists in training a self-organizing map (SOM) [27,28] neural network with the aim to create a model of the nominal behaviour of the system. For this scope we use an historical dataset, that we denote as training dataset, containing only nominal observations, where faulty instances have been removed. The motivation under this choice is that, as is commonly the case in monitoring applications, most of the measured data correspond to nominal behaviours, and very few cases of faulty patterns are usually measured. The usage of supervised learning methodologies, such as Feed-Forward Neural Networks or Support Vector Machine methods, is not advisable in the case of strongly unbalanced distributions of correct and faulty patterns. On the other hand, unsupervised learning methods are more suitable to represent the structure and the distribution of nominal data. Among unsupervised learning methods, that include clustering and vector quantization algorithms, SOMs are very convenient as they operate a map from the original multi-dimensional space to a two-dimensional space preserving the same topology of the original data (i.e., points that were close to each other in the input space correspond to cells that are still close to each other in the two-dimensional output space). Accordingly, SOMs are an excellent candidate when it is necessary to provide an accurate model of a multivariate distribution of data, and the nonlinear map towards the output space allows us to introduce a number of very useful tools for data analysis, such as the measurement of cell occupancy that has been proposed in this work.
In fact, SOMs have been widely used for condition monitoring applications in other contexts [29,30]. In this manuscript, an original KPI based on the frequency of cells occupancy has been introduced on purpose for our specific application of interest.
In particular, the trained SOM is used to calculate a parameter for each cell of the map, denoted as probability of cell occupancy, which represents the number of training points that are mapped to that particular cell, normalized with respect to the total number of points. During the monitoring stage, new state observations are presented to the SOM and are classified as “in control" or “out-of-control". For this purpose, we calculate the probability of cell occupancy for all the instances measured during the last 24 h, and we compare it against the previously computed probability of cell occupancy. The procedure is now illustrated in more detail.

Self-Organizing Map Neural Network Based Key Performance Indicator: Monitoring of Cell Occupancy

The SOM output space consists of a fixed and ordered bi-dimensional grid of cells, identified by an index in the range 1 , , D , where a distance metric d ( c , i ) between any two cells of index c and i is defined [27]. Each cell of index i is associated with a model vector m i R 1 × n that lies in the same high-dimensional space of the input patterns r Δ , where the matrix Δ R N × n represents the training dataset to be analysed, containing N observations of row vectors r R 1 × n . After the training, the distribution of the model vectors resembles the distribution of the input data, with the additional feature of preserving the grid topology: model vectors that correspond to neighbouring cells shall be neighbours in the high-dimensional input space as well.
When a new input sample r is presented to the network, the SOM finds the best matching unit (BMU) c, whose model vector m c has the minimum Euclidean distance from r :
c = argmin i { r m i } .
In this case we say that the input pattern r is mapped to the cell c. In order to assess the condition of newly observed state patterns to be monitored, we introduce the following KPI:
K P I ( d ) = i = 1 D p i , d 1 | p i , T R A I N p i , d | 1 + | p i , T R A I N p i , d |
where d denotes a test day index, and the probability of cell occupancy during day d is defined as
p i , d = N i , d N d , i = 1 D ,
where N d = 24 · 60 / ν is the total number of samples in a day, and N i , d is the number of samples, within day d, that were mapped to cell i. In the same fashion the probability of cell occupancy in the training phase is defined as
p i , T R A I N = N i , T R A I N N , i = 1 D
where N i , T R A I N represents the number of training patterns that were mapped to cell i, while N is the total number of training samples. It is straightforward that
0 p i , d 1 , i = 1 D
and
0 p i , T R A I N 1 , i = 1 D .
As a result, the K P I ( d ) value defined in Equation (1) is calculated once a day, based on the analysis of the measurements of the previous 24 h.
If the test samples of the day d being monitored represent mainly nominal observations, then the corresponding p i , d values shall be close to p i , T R A I N values, that were calculated using nominal historical observations. In this case the resulting value of the K P I in Equation (1) tends to 1. Conversely, if the patterns of day d contain abnormal conditions, then the cell occupancy will be mainly altered, leading to a situation where for a significant number of cells | p i , T R A I N p i , d | tends to 1 ( p i , d close to 1 and p i , T R A I N close to 0 or viceversa). In this case the resulting K P I value tends to 0.
From a physical point of view, the proposed K P I is a robust indicator that is able to detect changes in the underlying non-linear dynamics of the plant. The normal status is represented by K P I = 1 , while decreasing values represent a deviation from healthy conditions.
In particular, we have accurately tuned a set of rules and thresholds based on the K P I values, in order to generate warning levels of different severity, as summarized in Table 4. The following two thresholds are defined as lower control limits:
t h r 1 = μ 3 σ ,
and
t h r 2 = μ 5 σ ,
where μ and σ represent, respectively, the mean value and the standard deviation of the KPI values calculated as in Equation (1) using all the training patterns.
The logic of the generation of the warnings takes into account the crossing of the thresholds, the persistence of the K P I values below the thresholds and the derivative of the K P I . In particular, a negative derivative, representing a progressive degradation of the health state of the plant, is a necessary condition for the generation of a warning. This choice has the effect to avoid the generation of warnings during the positive derivative of K P I , that usually correspond to a period where the plant is gradually returning to a normal state after a maintenance intervention. In this way the number of false positives is greatly reduced.

4. Results and Discussion

The proposed model has been trained on the training set as specified in Table 3, and in this section we discuss the outcome of the testing stage. A 20 × 20 map was used ( D = 400 ), with hexagonal cells int the topological space, and the sequential update rule was used to train the map [27]. Figure 2 shows the a-posteriori calibration of the trained map for the inverter A.2 of plant A. Each cell of the calibration map is associated with a label related to the fault class of the patterns that were mapped to that cell most times. Obviously, in the training calibration map only the nominal class is present, and the obtained map results to be divided in two clusters related to day (left side) and night (rigth side). The test calibration map was created using only faulty patterns, yielding the distribution shown in the right side of Figure 2, where it can be noticed that different kind of faults are mapped to nearby cells related to day and night regions.
As a result, our system was able to identify a significant amount of failure events, which we could validate using the available data, and a selection of the most interesting ones is discussed in more detail in this section.

4.1. Plant A

Table 5 lists the most relevant faults occurred on the inverter module A.2 of plant A in the test period 01 October 2014 to 30 September 2015, i.e., 1 year long. For each failure, the table reports the specific fault, the time interval until the problem was fixed, and its severity in a scale from 1 (most critical) to 5 (least critical), as defined by the inverter manufacturer. According to the alarm logbook, this plant experienced a number of thermal issues on several different devices which lead to a non-correct heat dissipation. Such damages led to a production loss estimated in some thousands of euros and required the replacement of many components of the inverter of plant A in August–September 2015.
Figure 3 illustrates, on the top part, the curve of the proposed daily KPI (in blue), as well as the warning levels triggered by the KPI, with different colours depending on the severity of the warning, ranging from green (warning level 1—least critical), to red (warning level 4—the most critical). In order to evaluate the ability of the proposed KPI to detect anomalous working conditions, we also show in black the normalized number of the true faulty instances N ^ f a u l t that were registered on each day. In particular, the normalized number of true faulty instances on the d th day is computed as:
N ^ f a u l t d = number of faulty instances in day d 1 N d · 100 % ,
Roughly speaking, Equation (2) may give rise to value between 0, i.e., no fault observed in the day, up to 100%, indicating abnormal days with all the N d instances labelled as faulty. The two thresholds t h r 1 and t h r 2 are also represented by dashed and dotted black curves, respectively.
Additionally, in order to make a quantitative performance evaluation, in the bottom of Figure 3, it is possible to observe the True Positive Rate (TPR), the False Negative Rate (FNR), and the False Positive Rate (FPR), as a function of the date. In particular, the TPR is defined as the ratio of the true positive (faults) with respect to the actual faults (i.e., TPR = TP/P); the FNR is defined as the ratio of the false negatives (i.e., faults that were not recognized as faults) with respect to the actual faults, so that FNR = FN/P; and FPR, which is defined as the ratio of the False Positives with respect to the non-faults cases (i.e., it corresponds to the rate of false alarms). In practice, a faulty sample has been classified as a TP if at least one warning is triggered in the previous 7 days, otherwise it is classified as FN. On the other hand, if an alert is raised and no faults occur in the following week, the corresponding sample is taken as a FP. As can be seen, a clear correlation between the warning alerts and the actual faults is observed, with the most severe warnings triggered in correspondence of the most critical days, i.e., those having a higher percentage of registered faulty instances. Additionally, although the daily generated KPI may introduce a delay in the generation of predictive alerts, it is effective in minimizing the amount of generated false positives and false negatives instances, thanks also to the monitoring of the trend derivative.
In particular, the first critical failure (AC Switch Open), that gives rise to almost 60% of device faults in a day, is observed on 10 October 2014. The model anticipates the failure triggering warnings of level 1 on both October 4 and on October 9, with a significant degradation of the KPI in correspondence of the fault. The same failure occurs again on 3 November 2014 for a more prolonged number of days (until 28 November 2014), and for 26 consecutive days the SCADA registers almost 100% of daily faults of the device, and almost no power generation at all. The SOM early detects the anomaly with a remarkable drop of the KPI from 4 November, triggered by an unexpected zero power generation for an almost fully sunny day, as shown in Figure 4. However the first sign of abnormal behaviour had been predicted almost 10 days before with warnings of level 1 occurring on the 24 and 25 October, and with the KPI well below the first warning threshold t h r 1 . Furthermore, during the prolonged fault, the KPI notifies the operators with a degree of criticality that progressively increases up to the maximum level 4, thus strongly advising the plant operator to proceed with the reactive maintenance action. In particular, as can be seen in Figure 5, application of the proposed method and timely maintenance interventions could have led to an energy gain up to roughly 20%.
Conversely, it is interesting to note that the proposed method does not trigger any alert in correspondence of the DC Insulation fault, i.e., an overvoltage across the DC capacitors, that occurs on 9 December, due to the positive value of the KPI derivative. However, the last warning of level 4 activated on 4 December would have allowed the O&M team to plan a maintenance intervention and solve this issue in time. Then, from the beginning of year 2015 the overall trend of the KPI exhibits a slow but almost constant increase, with some alerts up to level 2 that occur in correspondence, or even ahead, of some minor, yet actual, faulty events.
The second most severe failure starts on the 11 June 2015 due to an overvoltage across the bulk capacitors of the DC/DC converter. Remarkably, also in this case the SOM realizes of the anomalous behaviour already on 6 June, and triggers a first warning of level 1. Additionally, a sudden KPI drop is observed in correspondence of the failure, with warnings generated up to level 4. After this failure, the nominal behaviour is restored and the model does not generate any alert until the end of August 2015, when the model predicts an anomaly on 23 August, which is followed by an actual registered fault that occurs the following day.
The performance over the whole test set are remarkable, with a TPR exceeding 93% (FNR < 7%) and a FPR of almost 13%.

4.2. Plant B

Table 6 lists the most critical failures occurred on inverter B.1 installed in plant B in the test period of interest from 1 April 2015 to 29 February 2016, whereas in Figure 6 the proposed KPI, the warning levels, as well as the daily number of faults are plotted as a function of time for the same device. In the first part, the KPI is almost always above the safety threshold and does not detect the DSP communication error between the inverter internal control devices that occurs on 16 July 2015. After that, the KPI starts to decrease and realizes of an incoming failure on 26 July, generating alerts up to level 2. A real failure occurs indeed on 6 August, due to an internal sensor error in the measurement of the leakage current on the DC side. A consequent maintenance action is then scheduled to verify the issue.
Then a new fault is predicted on August 10, with warnings triggered up to level 2. An overvoltage across the bulk capacitors on the DC side occurs indeed on 13 August, and it lasts almost 13 days, causing also the replacement of the inverter. The warnings triggered by the indicator during the failure are thus explained with the lack of data in these days.
From 24 August, the KPI starts signaling new faulty conditions, with alerts that progressively become more critical, up to level 4, in correspondence of an internal sensor fault that is registered by the system on 7 September. The consequent inspection of the O&M operators confirmed the fault and caused the replacement of a cooling fan in the inverter. Similarly, from 14 September, the KPI starts again progressively to decrease with new warnings that get up to level 4, when a failure is again registered on 23 September. On that day, a new corrective intervention is scheduled which causes the substitution of the inverter cooling pump. Then the KPI recovers safe values, with some alerts generated in correspondence of minor faulty events, with very few missed detections or misdetections.
Furthermore, for the second PV plant, the KPI performs in an accurate fashion, with a TPR exceeding 98% (FNR < 2%) and a FPR equal to about 18%.

4.3. Plant C

Table 7 lists the most severe failures registered for inverter 3.5 of plant C in the testing period, from 1 February to 27 July 2016. As in the previous cases, Figure 7 shows the proposed KPI, the warning levels and the daily number of faults as a function of time for the same module. As can be seen in Table 5, the device does not experience particularly important failures until the last decade of May. Indeed, the registered failures are mainly due either to some parameters outside of the standard values or by scheduled maintenance actions.
Accordingly, the model does not detect any relevant issue until 21 May (see Figure 7), when the KPI suddenly drops triggering alerts up to level 4. Looking at the single signals (Figure 8), an obvious anomaly is given by the power generation that is equal to zero irrespectively of the sunny weather conditions. Similarly, an unmotivated drop in the internal inverter temperature occurs. The technical inspection on the plant confirms the issue, which was caused by an IGBT stack fault and led to a production loss estimated roughly in 16 MWh. The whole inverter is then replaced after the failure. Then the KPI comes back to take safe values, generating only some alerts around 27 June, in correspondence of minor grid failures caused by mains parameters out of range.
The KPI works in an accurate way also for plant C, as can be seen in the bottom plot of Figure 7. In fact, the TPR is almost 92% (FNR = 8%) and FPR is just roughly 1%.

5. Conclusions

In Table 8, we summarize the performance results of the proposed method obtained in the three case studies. In particular, excellent performances are obtained in terms of TPR values, and good results are also achieved in terms of FPR for all the case studies. The predictive capacity of the proposed method is summarized in Table 9 reporting the dates of the occurrence of the faults, and the dates when such faults had been predicted by the proposed KPI. On average, the KPI predicts incipient faults between 6 and 7 days before they are observed in practice. Furthermore, in addition to being able to predict the faults, the KPI also exhibits excellent early detection capabilities, by signaling with increasing warning levels as the faults evolve and reach more severe conditions.
The proposed SOM-based monitoring system is now being installed in PV plants for online condition monitoring and the preliminary feedback from plant operators is very positive. A full evaluation of the online system will be the subject of our future work. Furthermore, we are currently developing a supervised fault-classification tool that we plan to integrate in the system in order to predict the specific class of fault, in addition to recognizing a generic faulty condition, as in our presented work.

Author Contributions

Conceptualization, M.T. and A.B.; methodology, M.T. and A.B.; software, A.B. and A.P.; validation, A.B., A.P. and M.T.; data curation, A.B. and Antonio Piazzi; writing—original draft preparation, A.B.; writing—review and editing, E.C., M.T., D.T. and S.B.; visualization, S.B.; supervision, S.B. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Moser, D.; Del Buono, M.; Jahn, U.; Herz, M.; Richter, M.; De Brabandere, K. Identification of technical risks in the photovoltaic value chain and quantification of the economic impact. Prog. Photovolt. Res. Appl. 2017, 25, 592–604. [Google Scholar] [CrossRef]
  2. Lindig, S.; Louwen, A.; Moser, D. Outdoor PV System Monitoring—Input Data Quality, Data Imputation and Filtering Approaches. Energies 2020, 13, 5099. [Google Scholar] [CrossRef]
  3. Beránek, V.; Olšan, T.; Libra, M.; Poulek, V.; Sedláček, J.; Dang, M.Q.; Tyukhov, I.I. New monitoring system for photovoltaic power plants’ management. Energies 2018, 11, 2495. [Google Scholar] [CrossRef] [Green Version]
  4. Woyte, A.; Richter, M.; Moser, D.; Mau, S.; Reich, N.; Jahn, U. Monitoring of photovoltaic systems: Good practices and systematic analysis. In Proceedings of the 28th European Photovoltaic Solar Energy Conference, Villepinte, France, 30 September–4 October 2013; pp. 3686–3694. [Google Scholar]
  5. Moreno-Garcia, I.M.; Palacios-Garcia, E.J.; Pallares-Lopez, V.; Santiago, I.; Gonzalez-Redondo, M.J.; Varo-Martinez, M.; Real-Calvo, R.J. Real-time monitoring system for a utility-scale photovoltaic power plant. Sensors 2016, 16, 770. [Google Scholar] [CrossRef] [Green Version]
  6. Lazzaretti, A.E.; Costa, C.H.D.; Rodrigues, M.P.; Yamada, G.D.; Lexinoski, G.; Moritz, G.L.; Santos, R.B.D. A monitoring system for online fault detection and classification in photovoltaic plants. Sensors 2020, 20, 4688. [Google Scholar] [CrossRef]
  7. Okere, A.; Iqbal, M.T. A Review of Conventional Fault Detection Techniques in Solar PV Systems and a Proposal of Long Range (LoRa) Wireless Sensor Network for Module Level Monitoring and Fault Diagnosis in Large Solar PV Farms. Eur. J. Electr. Eng. Comput. Sci. 2020, 4. [Google Scholar] [CrossRef]
  8. Gimeno-Sales, F.J.; Orts-Grau, S.; Escribá-Aparisi, A.; González-Altozano, P.; Balbastre-Peralta, I.; Martínez-Márquez, C.I.; Seguí-Chilet, S. PV Monitoring System for a Water Pumping Scheme with a Lithium-Ion Battery Using Free Open-Source Software and IoT Technologies. Sustainability 2020, 12, 10651. [Google Scholar] [CrossRef]
  9. Betti, A.; Lo Trovato, M.; Leonardi, F.S.; Leotta, G.; Ruffini, F.; Lanzetta, C. Predictive Maintenance in Photovoltaic Plants with a Big Data Approach. In Proceedings of the 33rd European Photovoltaic Solar Energy Conference, Amsterdam, The Netherlands, 25–29 September 2017; pp. 1895–2000. [Google Scholar]
  10. Chine, W.; Mellit, A.; Pavan, A.M.; Kalogirou, S.A. Fault detection method for grid-connected photovoltaic plants. Renew. Energy 2014, 66, 99–110. [Google Scholar] [CrossRef]
  11. Dhimish, M.; Holmes, V. Fault detection algorithm for grid-connected photovoltaic plants. Sol. Energy 2016, 137, 236–245. [Google Scholar] [CrossRef]
  12. Pei, T.; Hao, X. A fault detection method for photovoltaic systems based on voltage and current observation and evaluation. Energies 2019, 12, 1712. [Google Scholar] [CrossRef] [Green Version]
  13. Navid, Q.; Hassan, A.; Fardoun, A.A.; Ramzan, R. An Online Novel Two-Layered Photovoltaic Fault Monitoring Technique Based Upon the Thermal Signatures. Sustainability 2020, 12, 9607. [Google Scholar] [CrossRef]
  14. Zhao, Q.; Shao, S.; Lu, L.; Liu, X.; Zhu, H. A new PV array fault diagnosis method using fuzzy C-mean clustering and fuzzy membership algorithm. Energies 2018, 11, 238. [Google Scholar] [CrossRef] [Green Version]
  15. Chen, Z.; Han, F.; Wu, L.; Yu, J.; Cheng, S.; Lin, P.; Chen, H. Random forest based intelligent fault diagnosis for PV arrays using array voltage and string currents. Energy Convers. Manag. 2018, 178, 250–264. [Google Scholar] [CrossRef]
  16. Zhu, H.; Lu, L.; Yao, J.; Dai, S.; Hu, Y. Fault diagnosis approach for photovoltaic arrays based on unsupervised sample clustering and probabilistic neural network model. Sol. Energy 2018, 176, 395–405. [Google Scholar] [CrossRef]
  17. Lu, X.; Lin, P.; Cheng, S.; Lin, Y.; Chen, Z.; Wu, L.; Zheng, Q. Fault diagnosis for photovoltaic array based on convolutional neural network and electrical time series graph. Energy Convers. Manag. 2019, 196, 950–965. [Google Scholar] [CrossRef]
  18. Chen, Z.; Chen, Y.; Wu, L.; Cheng, S.; Lin, P. Deep residual network based fault detection and diagnosis of photovoltaic arrays using current-voltage curves and ambient conditions. Energy Convers. Manag. 2019, 198, 111793. [Google Scholar] [CrossRef]
  19. Taghezouit, B.; Harrou, F.; Sun, Y.; Arab, A.H.; Larbes, C. Multivariate statistical monitoring of photovoltaic plant operation. Energy Convers. Manag. 2020, 205, 112317. [Google Scholar] [CrossRef]
  20. Kusiak, A.; Li, W. The prediction and diagnosis of wind turbine faults. Renew. Energy 2011, 36, 16–23. [Google Scholar] [CrossRef]
  21. Zaher, A.S.A.E.; McArthur, S.D.J.; Infield, D.G.; Patel, Y. Online wind turbine fault detection through automated SCADA data analysis. Wind Energy 2009, 12, 574–593. [Google Scholar] [CrossRef]
  22. Polo, F.A.O.; Bermejo, J.F.; Fernández, J.F.G.; Marquez, A.C. Assistance to Dynamic Maintenance Tasks by Ann-Based Models. In Advanced Maintenance Modelling for Asset Management; Crespo Márquez, A., González-Prida Díaz, V., Gómez Fernández, J., Eds.; Springer: Cham, Switzerland, 2018; pp. 387–411. [Google Scholar]
  23. Malarvizhi, M.R.; Thanamani, A.S. K-nearest neighbor in missing data imputation. Int. J. Eng. Res. Dev. 2012, 5, 5–7. [Google Scholar]
  24. Zhang, S. Nearest neighbor selection for iteratively kNN imputation. J. Syst. Softw. 2012, 85, 2541–2552. [Google Scholar] [CrossRef]
  25. Arianos, S.; Carbone, A. Detrending moving average algorithm: A closed-form approximation of the scaling law. Phys. A Stat. Mech. Appl. 2007, 382, 9–15. [Google Scholar] [CrossRef] [Green Version]
  26. Cowan, G. Statistical Data Analysis; Oxford University Press: Oxford, CA, USA, 1998. [Google Scholar]
  27. Kohonen, T. Self-Organizing Maps, 3rd ed.; Springer: Berlin/Heidelberg, Germany, 2001. [Google Scholar]
  28. Tucci, M.; Raugi, M. Adaptive FIR neural model for centroid learning in self-organizing maps. IEEE Trans. Neural Netw. 2010, 21, 948–960. [Google Scholar] [CrossRef] [PubMed]
  29. Jämsä-Jounela, S.L.; Vermasvuori, M.; Endén, P.; Haavisto, S. A process monitoring system based on the Kohonen self-organizing maps. Control. Eng. Pract. 2003, 11, 83–92. [Google Scholar] [CrossRef] [Green Version]
  30. Silva, R.G.; Wilcox, S.J. Feature evaluation and selection for condition monitoring using a self-organizing map and spatial statistics. Artif. Intell. Eng. Des. Anal. Manuf. 2019, 33, 1–10. [Google Scholar] [CrossRef]
Figure 1. Location of the three considered PV plants, which are denoted as A, B, C. The marker size is proportional to the installed capacity, which is shown in Table 1.
Figure 1. Location of the three considered PV plants, which are denoted as A, B, C. The marker size is proportional to the installed capacity, which is shown in Table 1.
Sensors 21 01687 g001
Figure 2. Calibration of the trained SOM, using training data and test data.
Figure 2. Calibration of the trained SOM, using training data and test data.
Sensors 21 01687 g002
Figure 3. Historical case studies for inverter A.2 of plant A. Top plot—Left axis: KPI, as well as the warning levels and the upper and lower thresholds are shown as a function of datetime; Right axis: time series of daily number of faults. Bottom plot: TPR, FNR and FPR as a function of datetime.
Figure 3. Historical case studies for inverter A.2 of plant A. Top plot—Left axis: KPI, as well as the warning levels and the upper and lower thresholds are shown as a function of datetime; Right axis: time series of daily number of faults. Bottom plot: TPR, FNR and FPR as a function of datetime.
Sensors 21 01687 g003
Figure 4. P a c of inverter A.2 (top) and GTI (bottom) as a function of datetime in the period 20 October 2014–30 November 2014 (plant A). The warning levels are superimposed for convenience.
Figure 4. P a c of inverter A.2 (top) and GTI (bottom) as a function of datetime in the period 20 October 2014–30 November 2014 (plant A). The warning levels are superimposed for convenience.
Sensors 21 01687 g004
Figure 5. Left axis: energy ratio with respect to the ideal case with and without the application of the SOM based model for inverter A.2 of Plant A. The green area represents the maximum energy gain achievable by enabling it. Right axis: time series of daily number of faults.
Figure 5. Left axis: energy ratio with respect to the ideal case with and without the application of the SOM based model for inverter A.2 of Plant A. The green area represents the maximum energy gain achievable by enabling it. Right axis: time series of daily number of faults.
Sensors 21 01687 g005
Figure 6. Historical case studies for inverter B.1 of plant B. Top plot—Left axis: KPI, as well as the warning levels and the upper and lower thresholds are shown as a function of datetime; Right axis: time series of daily number of faults. Bottom plot: TPR, FNR and FPR as a function of datetime.
Figure 6. Historical case studies for inverter B.1 of plant B. Top plot—Left axis: KPI, as well as the warning levels and the upper and lower thresholds are shown as a function of datetime; Right axis: time series of daily number of faults. Bottom plot: TPR, FNR and FPR as a function of datetime.
Sensors 21 01687 g006
Figure 7. Historical case studies for inverter 3.5 of plant C. Top plot—Left axis: KPI, as well as the warning levels and the upper and lower thresholds are shown as a function of datetime; Right axis: time series of daily number of faults. Bottom plot: TPR, FNR and FPR as a function of time.
Figure 7. Historical case studies for inverter 3.5 of plant C. Top plot—Left axis: KPI, as well as the warning levels and the upper and lower thresholds are shown as a function of datetime; Right axis: time series of daily number of faults. Bottom plot: TPR, FNR and FPR as a function of time.
Sensors 21 01687 g007
Figure 8. From top to bottom: P a c , GTI and T i n t of inverter 3.5 of plant C as a function of datetime in the period 8 May–7 June 2016. The warning levels are superimposed for convenience.
Figure 8. From top to bottom: P a c , GTI and T i n t of inverter 3.5 of plant C as a function of datetime in the period 8 May–7 June 2016. The warning levels are superimposed for convenience.
Sensors 21 01687 g008
Table 1. List of considered PV plants: plant A is located in Romania, whereas plants B and C in Greece.
Table 1. List of considered PV plants: plant A is located in Romania, whereas plants B and C in Greece.
Plant
Name
Number of
Inverter
Modules
Inverter
Manufacturer
Number
Max Active
Power
[kW]
Plant
Nominal
Power [MW]
A3513859.8
B713852.8
C252183.44.9
Table 2. List of electrical and enviromental signals used as input tags. T i n t is only available for plant C.
Table 2. List of electrical and enviromental signals used as input tags. T i n t is only available for plant C.
Signal NumberSignal TypeSignal NameVariable NameUnit
1ElectricalDC Current I D C [A]
2ElectricalDC Voltage V D C [V]
3ElectricalDC Power P D C [W]
4ElectricalAC Current I A C [A]
5ElectricalAC Voltage V A C [V]
6ElectricalAC Power P A C [W]
7EnvironmentalInternal Inverter
Temperature
T i n t [°C]
8EnvironmentalPanel Temperature T m o d [°C]
9EnvironmentalAmbient Temperature T a m b [°C]
10EnvironmentalGlobal Tilted
Irradiance
GTI[W/m 2 ]
11EnvironmentalGlobal Horizontal
Irradiance
GHI[W/m 2 ]
Table 3. Temporal extension of the data-sets and data used for training and for testing for each PV plant. The number of measured patterns is also shown (patterns are 10 dimensional for plants A and B, and 11 dimensional for plant C).
Table 3. Temporal extension of the data-sets and data used for training and for testing for each PV plant. The number of measured patterns is also shown (patterns are 10 dimensional for plants A and B, and 11 dimensional for plant C).
Plant NameTraining Period (dd/mm/yyyy)Test Period (dd/mm/yyyy)
Afrom 20/03/2014 to 30/09/2014
n° of patterns: 55,872
from 01/10/2014 to 30/09/2015
n° of patterns: 104,832
Bfrom 27/10/2014 to 31/03/2015
n° of patterns: 44,640
from 01/04/2015 to 29/02/2016
n° of patterns: 96,192
Cfrom 01/02/2015 to 31/01/2016
n° of patterns: 104,832
from 01/02/2016 to 27/07/2016
n° of patterns: 50,976
Table 4. Logic for the generation of the warning levels.
Table 4. Logic for the generation of the warning levels.
Warning LevelThresholds KPI DerivativePersistence
1 K P I < μ 3 σ <01 day
2 K P I < μ 3 σ <0≥2 days
3 K P I < μ 5 σ <01 day
4 K P I < μ 5 σ <0≥2 days
Table 5. Main failures occurred on inverter A.2 of plant A in the historical period.
Table 5. Main failures occurred on inverter A.2 of plant A in the historical period.
Fault NameSeverity (1 to 5)Start Date (dd/mm/yyyy)End Date (dd/mm/yyyy)
AC Switch Open210/10/201411/10/2014
AC Switch Open203/11/201428/11/2014
DC Insulation Fault209/12/201410/12/2014
DC Voltage High211/06/201523/06/2015
AC Switch Open224/08/201525/08/2015
Table 6. Main failures occurred on inverter B.1 of plant B during the testing period.
Table 6. Main failures occurred on inverter B.1 of plant B during the testing period.
Fault
Name
Severity
(1 to 5)
Start Date
(dd/mm/yyyy)
End Date
(dd/mm/yyyy)
Notes
Communication
Error
216/07/201516/07/2015None
Internal
sensor fault
206/08/201507/08/2015Fault Log
downloading
DC Voltage
High
213/08/201525/08/2015Device B.1
replaced
Internal
sensor fault
207/09/201507/09/2015Cooling fan
replaced
Internal
sensor fault
223/09/201523/09/2015Cooling pump
replaced
Table 7. Main failures occurring on inverter 3.5 of plant C during the testing period.
Table 7. Main failures occurring on inverter 3.5 of plant C during the testing period.
Fault
Name
Severity
(1 to 5)
Start Date
(dd/mm/yyyy)
End Date
(dd/mm/yyyy)
Notes
AC Voltage
out of range
307/03/201607/03/2016Grid fault
AC Voltage
out of range
309/03/201609/03/2016Grid fault
AC Voltage
out of range
312/04/201612/04/2016Grid fault
AC Voltage
out of range
315/05/201615/05/2016Scheduled
maintenance
AC Switch
Open
221/05/201607/06/2016Inverter 3.5
replaced
Table 8. Summary of the performance results on the three case studies.
Table 8. Summary of the performance results on the three case studies.
Test CaseTPRFNRFPR
Plant A, inv. A.293%7%13%
Plant B, inv. B.198%2%18%
Plant C, inv. 3.592%8%1%
Table 9. Summary of the predictive performance of the proposed method.
Table 9. Summary of the predictive performance of the proposed method.
Test CaseDate of Fault
Occurrence
(dd/mm/yyyy)
Date of Fault
Prediction
(dd/mm/yyyy)
Time in Advance
of Prediction
Plant A, inv. A.210/10/20144/10/20146 days
Plant A, inv. A.23/11/201424/10/201410 days
Plant A, inv. A.209/12/2014last warning on
04/12/2014
(5 days)
fault occurs during
plant maintenance
Plant A, inv. A.211/06/201506/06/20155 days
Plant A, inv. A.224/08/201523/08/20151 day
Plant B, inv. B.116/07/2015not detected-
minor fault
Plant B, inv. B.106/08/201526/07/201510 days
Plant B, inv. B.113/08/201510/08/20153 days
Plant B, inv. B.107/09/201524/08/201514 days
Plant B, inv. B.123/09/201514/09/20159 days
Plant C, inv. 3.521/05/201621/05/20160 days
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Betti, A.; Tucci, M.; Crisostomi, E.; Piazzi, A.; Barmada, S.; Thomopulos, D. Fault Prediction and Early-Detection in Large PV Power Plants Based on Self-Organizing Maps. Sensors 2021, 21, 1687. https://doi.org/10.3390/s21051687

AMA Style

Betti A, Tucci M, Crisostomi E, Piazzi A, Barmada S, Thomopulos D. Fault Prediction and Early-Detection in Large PV Power Plants Based on Self-Organizing Maps. Sensors. 2021; 21(5):1687. https://doi.org/10.3390/s21051687

Chicago/Turabian Style

Betti, Alessandro, Mauro Tucci, Emanuele Crisostomi, Antonio Piazzi, Sami Barmada, and Dimitri Thomopulos. 2021. "Fault Prediction and Early-Detection in Large PV Power Plants Based on Self-Organizing Maps" Sensors 21, no. 5: 1687. https://doi.org/10.3390/s21051687

APA Style

Betti, A., Tucci, M., Crisostomi, E., Piazzi, A., Barmada, S., & Thomopulos, D. (2021). Fault Prediction and Early-Detection in Large PV Power Plants Based on Self-Organizing Maps. Sensors, 21(5), 1687. https://doi.org/10.3390/s21051687

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop