Next Article in Journal
Comprehensive Analysis of Traffic Accidents in Seoul: Major Factors and Types Affecting Injury Severity
Next Article in Special Issue
Rolling Bearing Health Indicator Extraction and RUL Prediction Based on Multi-Scale Convolutional Autoencoder
Previous Article in Journal
Physiological Effects of Red-Colored Food-Derived Bioactive Compounds on Cardiovascular and Metabolic Diseases
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Generative Adversarial Network-Based Fault Detection Approach for Photovoltaic Panel

1
College of Computer Science and Technology, Shanghai University of Electric Power, Shanghai 200090, China
2
Department of Electronic Engineering, School of Electronic Information and Electrical Engineering, Shanghai Jiao Tong University, Shanghai 200240, China
3
College of Chemical Engineering, Zhejiang University of Technology, Hangzhou 310014, China
4
Zhijiang College, Zhejiang University of Technology, Shaoxing 312000, China
5
School of Economics, Fudan University, Shanghai 200010, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2022, 12(4), 1789; https://doi.org/10.3390/app12041789
Submission received: 19 December 2021 / Revised: 3 February 2022 / Accepted: 5 February 2022 / Published: 9 February 2022

Abstract

:
Photovoltaic (PV) panels are widely adopted and set up on residential rooftops and photovoltaic power plants. However, long-term exposure to ultraviolet rays, high temperature and humid environments accelerates the oxidation of PV panels, which finally results in functional failure. The traditional fault detection approach for photovoltaic panels mainly relies on manual inspection, which is inefficient. Lately, machine vision-based approaches for fault detection have emerged, but lack of negative samples usually results in low accuracy and hinders the wide adoption of machine vision-based approaches. To address this issue, we proposed a semi-supervised anomaly detection model based on the generative adversarial network. The proposed model uses the generator network to learn the data distribution of the normal PV panel dataset during training. When abnormal PV panel data are put into the model in the test phase, the reconstructed image generated by the model does not equal the input image. Since the abnormal PV panel data do not obey the data distribution learned by the generator, the difference between the original image and its reconstructed image exceeds the given threshold. So, the model can filter out the fault PV panel by checking the error value between the original image and its reconstructed image. The model adopts Gradient Centralization and SmoothL1 loss function to improve its generalization performance. Meanwhile, we use the convolutional block attention module (CBAM) to make the model pay more attention to the defective area and greatly improve the performance of the model. In this paper, the photovoltaic panels dataset is collected from a PV power plant located in Zhejiang, China. We compare the proposed approach with state-of-the-art semi-supervised and unsupervised approaches (i.e., AnoGAN (Anomaly Detection with Generative Adversarial Networks), Zhao’s method, GANomaly, and f-AnoGAN), and the result indicates that the Area Under Curve (AUC) increases by 0.06, 0.052, 0.041 and 0.035, respectively, significantly improving the accuracy of photovoltaic panel fault detection.

1. Introduction

Traditional thermal power generation uses heat energy generated by combustible materials (i.e., coal) and converts it into electric energy through power generation plants. The combustion of coal produces carbon dioxide, which is the main source of carbon emissions. Due to the rapid increase in carbon emissions, the concentration of global CO2 continues to rise, and the caused greenhouse effect results in global warming, which poses a great threat to the survival of human beings. Since 1992, the Intergovernmental Panel on Climate Change (IPCC) has signed a series of agreements, including the Paris Agreement. The overall goal of the Paris Agreement was to achieve a peak in global greenhouse gas emissions around 2020, and net zero greenhouse gas emissions in the second half of this century. Photovoltaic power generation is a technology that directly converts light energy into electric energy by using the photovoltaic effect at the semiconductor interface. Abundant solar radiant energy is an important energy that is inexhaustible, pollution-free, cheap and free to be used by human beings. Therefore, solar energy has become the focus of attention because of its unique advantages [1,2]. In the global promotion of clean and low-carbon energy transition, the adoption of photovoltaic (PV) panels in clean energy is increasing year by year [3]. According to International Energy Agency (IEA) statistics, since 2018, the cumulative installed capacity of global PV power generation has reached 480GW and is expected to reach 1721GW in 2030 [4].
In order to increase the exposure time of PV panels, large-scale PV power plants are located in areas without apparent shade, such as the roof of buildings, plains, hills and over fish ponds, as shown in Figure 1. Long-term exposure to ultraviolet rays, high temperature and humid environments accelerate the oxidation of PV panels, which results in functional failure [5]. Severe sealant delamination increases reflection, decreases irradiance, sinks moisture in the module, and accelerates panel oxidation [6]. The accumulation of dust directly reduces light transmittance and affects the efficiency of the PV power generation system, sometimes up to about 50% or, even worse, 80% [7]. Although snail trials have no direct effect on power generation efficiency, invisible cell cracks usually reduce their power output [8]. Therefore, it is necessary to inspect the status of photovoltaic panels regularly to find and replace defective photovoltaic panels in time. However, most of the photovoltaic panels are installed on roofs, fish ponds and hillsides with a wide area, so manual inspection cannot be carried out. Recently, several artificial intelligence algorithms (e.g., deep convolutional neural networks) were introduced to facilitate the process of PV panel fault detection [9,10,11,12,13]. Unfortunately, such fully supervised learning-based methods have been proven unsuitable for large-scale PV panel fault detection due to the lack of defective samples [14,15,16]. At present, semi-supervised learning-based approaches [17,18,19,20] are being explored by many researchers.
To solve these issues, by leveraging adversarial auto-encoders and conditional Generative Adversarial Networks (GAN), we propose a semi-supervised anomaly detection approach (named ppFDetector) to facilitate the process of fault detection in PV panels. The main contributions in this paper can be summarized thus:
(1)
A novel model-based semi-supervised generative adversarial network is proposed. Compared with the fully supervised learning model, the semi-supervised anomaly detection model does not require a large number of negative samples, which solves the problem of the anomaly detection model’s inability to be trained without negative samples.
(2)
The original optimization function is implemented with gradient centralization (GC) to regularize weight and output space to prevent the model from overfitting.
(3)
Convolutional block attention module (CBAM) is used to make the model pay more attention to the defective area, and improve the performance of the model.
(4)
SmoothL1loss is used to define the loss, which can combine the advantages of L1Loss and L2Loss to speed up the training of the model.
The remainder of this paper is organized as follows: Section 2 describes related literatures. Section 3 details the technical implementation of the ppFDetector. Experimental results using the ppFDetector on the PV panels dataset compared with state-of-the-art anomaly detection models are presented in Section 4. We finally conclude the paper in Section 5.

2. Related Work

2.1. Generative Adversarial Networks

The network structure of GAN is composed of a generation network and a discriminant network. The generator receives the random variable and generates fake sample data. The purpose of the generator is to try to make the generated samples the same as the real samples. The input of the discriminator consists of two parts, real data and data generated by the generator. The output is usually a probability value, which represents the probability that the input is a true distribution. If the input comes from real data, then output 1; otherwise, output 0. At the same time, the output of the discriminator will be fed back to guide’s training. Ideally, it cannot distinguish whether the input data are real or generated, i.e., each time the output probability is 1/2 (equivalent to random guessing), the data generated by the generator have reached real. At this point, the model is optimal. In practical applications, the generation network and discriminant network are usually implemented by deep neural networks. As the number of network layers increases, it is difficult for GAN networks to converge. In order to solve this problem, a wide variety of GAN variant models have emerged. Deep Convolutional Generative Adversarial Networks (DCGAN) [21] is the first major improvement on GAN architecture. Its generator is an upsampling process, i.e., deconvolution. The discriminator is a common convolutional network. It is more stable in training and produces higher quality samples.
With the development of deep learning, generative adversarial networks can also perform image anonymization to protect training data. Ashish Undirwade and Sujit Das [22] proposed applying optimized noise to the latent space representation (LSR) of the original image, and then converting the input original image into a new synthetic image through generative adversarial networks.
At present, GAN is a promising network model that can be applied in many fields, such as image super-resolution, image translation and style conversion. Karnewar [23] proposed a Multi-Scale Gradient Generative Adversarial Network (MSG-GAN), which is a simple but effective technique for addressing this by allowing the flow of gradients from the discriminator to the generator at multiple scales. This technology provides a stable method for high-resolution image synthesis, and can replace the commonly used progressive growth technology.

2.2. PV Panel Fault Detection

PV panel fault detection roughly goes through the following phases: an initial phase, the phase of physical approaches and the phase of intelligent detection.
In the initial phase, manual inspection is the most primitive method for detecting PV panel faults. The worker uses an image of the PV panel taken by a camera to check whether there is a defect. Traditional manual inspection evaluates a single PV panel just by sight, which is usually high cost, high error rate and low efficiency [24], and has severe subjective inclines for different defects.
In the phase of physical approaches, Sawyer DE et al. [25] used laser scanning technology to aid the fault-detection process. It displays the continuity of resistance in the crystalline silicon of the forward conversion of laser scanning. If there exists a crack, there will be a discontinuity of resistance. Tsuzuki K et al. [26] proposed the use of sound waves for fault detection, which makes PV panels vibrate to generate sound waves. Based on a comparative analysis with sound waves generated by non-defective solar cells, the fault detection can be realized. Physical detection [25,26,27] has high detection accuracy for specific types of PV panel faults, but it cannot effectively detect other faults.
In the intelligent phase of fault detection, the intelligent PV panel fault detection method based on machine vision is now being gradually adopted by users. By machine vision detection, Bastari et al. [28] proposed a method for texture analysis of dark defect areas in electroluminescence images to classify defective and non-defective cells. Quarter et al. [29] first reported the UAV inspection system to perform non-destructive inspection of large-scale PV power plants. However, the effect of fault diagnosis cannot meet the requirements of practical use, and many researchers are further studying fault diagnosis methods. Li et al. [30] first extracted features by calculating the first derivative of the Gaussian function on the collected images, and then performed defect detection through feature matching. This method verifies the effectiveness of the fault analysis algorithm based on image processing to a certain extent, but strictly requires good quality of the collected images. However, due to various reasons such as the wind effect, the obtained image resolution is low, and the performance of this method is significantly reduced. In addition, traditional pattern recognition algorithms often fail to extract fault features with acceptable complexity from the acquired aerial images.
With the evolution of deep convolutional neural networks, the introduction of artificial intelligence algorithms into the intelligent patrol system [9] can improve its robustness and reliability [10]. Li et al. [11] proposed an intelligent diagnosis method of aerial PV module image defects based on the deep convolutional neural network (CNN). This method uses CNN to classify multiple deep features and diverse states. Compared with traditional methods, it can flexibly and reliably solve the problem of PV module images with low quality and distortions. Tang [12] proposed using a generative adversarial network combined with data augmentation to increase training samples; this achieves high classification accuracy, but cannot precisely locate defects. Zhao [13] presented a deep learning-based automatic detection of multitype defects to fulfill inspection requirements of the production line. These studies used CNN to solve PV panel fault detection, and the models were based on supervised learning. However, due to the lack of abnormal panel samples, supervised learning is not suitable for this kind of problem. Schlegl et al. [15] proposed an unsupervised anomaly detection method based on the generative adversarial network (AnoGAN) only using standard panel samples; it uses continuously iterative optimization to find a certain point in latent space. However, due to the need for iterative optimization, the AnoGAN algorithm is bound to consume a lot of time. Zhao et al. [16] proposed using GAN and auto-encoder to reconstruct the defect image, and then use Local Binary Pattern (LBP) to extract the local contrast of the image to detect the defect. However, for images with complex backgrounds, Zhao’s method cannot reconstruct and repair defective images well for further fault detection. Also proposed was the f-AnoGAN [17], which can quickly map a picture to a certain point in hidden space and then use wasserstein GAN (WGAN) for anomaly detection. However, its quantitative evaluation of segmentation accuracy of anomaly detection only serves as a coarse indication.

3. ppFDetector Solution

In this paper, positive samples represent normal PV panel images without defect, and negative samples represent abnormal PV panel images with defect. As the diagram in Figure 2 shows, the workflow of ppFDetector is composed of GAN model training and anomaly detection strategy. During GAN model training, the generator network is designed to learn the data distribution of positive samples in the dataset of PV panel, and the discriminator network is designed for adversarial training, so as to reduce the error between the reconstructed image and its input original image. Then the generator can capture the training data distribution within both the normal PV panel image and its latent space vector. In the anomaly detection strategy, the test PV panel image (including positive samples and negative samples) enters the trained encoder–decoder to generate the latent space vector z and its reconstructed image. Next, the reconstructed image enters the encoder to generate its latent space vector z ˜ , and the error M ( z , z ˜ ) between latent space vectors z and z ˜ is computed. If M ( z , z ˜ ) > θ , the proposed model judges the test image as an abnormal PV panel, otherwise it is normal. The details of the model are described as follows.

3.1. GAN Model

The network structure of GAN model is composed of a generator network G and a discriminator network D . The generator and the discriminator are constructed by iterative confrontation training. Considering that the color and shape patterns of the PV panel image are usually single and regular, the generator network in GAN can better learn the data distribution of the PV panel image set and generate a more ideal reconstructed image.

3.1.1. Generator Network

As shown in Figure 3, the generator network is mainly composed of encoder network G E and decoder network G D . The details of the encoder parameters and decoder parameters are shown in Table 1 and Table 2, respectively. Firstly, the generator network G reads the input PV panel image x and passes it into the encoder network G E . After x is passed through the convolutional layer, batch normalization layer, the leaky linear rectification activation function layer and CBAM layer, and scaled down several times, x is compressed into a latent space vector z = G E ( x ) . The latent space vector z can theoretically be assumed as the best representation of x with the smallest dimension. Next, the decoder network G D enlarges z through the transposed convolution layer, batch normalization layer and linear rectification activation function layer. After upscaling, the latent space vector z is reconstructed into the reconstructed image x ˜ corresponding to the original PV panel image x , i.e., x ˜ = G ( x ) .

3.1.2. Encoder Subnet

The structure of the encoder subnet E consists of the convolutional layer, batch normalization layer, leaky linear rectification activation function layer and CBAM layer. Through the G E D , the reconstructed image data x ˜ is output. Put x ˜ into E to obtain the encoded latent space vector z ˜ = E ( x ˜ ) of the reconstructed image, and the fault detection of PV panels can be realized by capturing the error between z and z ˜ .

3.1.3. Discriminator Network

In ppFDetector, the original PV panel image x and its reconstructed image x ˜ are put into the discriminator network D . After the transposed convolution layer, the batch normalization layer and the leaky linear rectification activation function layer, the input image is downscaled by several times. After flattening, the compressed vector enters the fully connected layer to achieve the purpose of binary classification (by checking whether the input is the original PV panel image or the reconstructed one).

3.2. Loss Functions in GAN Model Training

In order to mathematicalize the above assumptions, three loss functions are defined and combined to establish the objective function of the model. Each loss function optimizes a single subnet. At the same time, this paper uses SmoothL1Loss instead of the combination of L1Loss and L2Loss. The formula of SmoothL1Loss is as follows:
s m o o t h L 1 ( t ) = { 0.5 t 2 | t | < 1 | t | 0.5 o t h e r w i s e
From the above formula, when the absolute value of t is less than 1, SmoothL1Loss is L2Loss. Because L2Loss will square the error, a smaller loss will be obtained, which is conducive to model convergence. Otherwise, SmoothL1Loss is the translation of L1Loss. L1Loss is not sensitive to outliers, and the magnitude of the gradient can be controlled so that it is not easy to explode during training. So SmoothL1Loss has the advantages of both L2Loss and L1Loss.

3.2.1. Adversarial Loss

The generator network G needs to update the training parameters according to the classification accuracy of discriminator network D (to determine whether the input is the original PV panel image or the reconstructed image). This section only refers to the generator part, but meanwhile, the discriminator is also trained and its weights are updated. In this paper, we update the generator network G based on the features of the intermediate layer of the discriminator denoted as φ ( x ) , as shown in Figure 3. Assuming that the independent variables x and x ˜ = G ( x ) both obey the positive sample data distribution P x , the adversarial loss function L a d v is the distance between the original PV panel image x and its reconstructed image x ˜ using s m o o t h L 1 distance, it is defined as follows:
L a d v = E x ~ P x | | φ ( x ) φ ( G ( x ) ) | | s m o o t h L 1
where P x is the data distribution of positive normal PV panel images; φ ( x ) is the output feature of the input PV panel image x in the intermediate layer of the discriminator; and G ( x ) is the reconstructed image of x in the generator. This loss is only activated with negative samples in the discriminator.

3.2.2. Contextual Loss

By optimizing the adversarial loss function L a d v in Equation (2), the generator G is able to produce images that are realistic enough to prevent discriminator D from discriminating between real and fake. However, for the reconstruction task, adversarial loss only allows the generator G to produce a reconstruction image x ˜ that obeys positive sample data distribution P x . But the reconstruction image x ˜ may not correspond to its original PV panel image x , and thus cannot achieve our purpose of reconstruction of the PV panel.
In this paper, we define the context loss L c o n by calculating the s m o o t h L 1 distance between the reconstructed image x ˜ and its original PV panel image x , and the equation is as follows:
L c o n = E x ~ P x | | x G ( x ) | | s m o o t h L 1
where G ( x ) is the reconstructed image in generator when input PV panel image x .

3.2.3. Encoder Loss

The two losses defined in Equations (2) and (3) can help the generator produce reconstruction images that are both realistic and relevant to the original PV panel image. We also use the latent space vector z to replace the original PV panel image x as the data for detecting anomalies. In order to reduce the difference between latent space vectors z and z ˜ , this paper defines the encoder loss L e n c , and its function equation is as follows:
L e n c = E x ~ P x | | G E ( x ) E ( G ( x ) ) | | s m o o t h L 1
where G E ( x ) is the latent space vector z output from encoder subnet of generator G E when the input is PV panel image x . E ( G ( x ) ) is the reconstructed latent space vector z ˜ when input PV panel image x firstly passes through the generator G and then passes through encoder subnet E .
Up to now, the objective loss function L of the GAN model can be defined as follows:
L = L a d v + λ L c o n + L e n c
where λ is a parameter for adjusting the sharpness of the reconstructed image.

3.3. Gradient Centralizaton

Since the gradient of neural network becomes more and more difficult to descend as the number of network layers increases, problems such as gradient disappearance, gradient explosion, and inability to converge will occur. So, optimization techniques are crucial for efficient training of neural networks.
In this paper, we merge the idea of gradient centralization (GC) [31] to optimize the gradient directly. Using weight space and output feature space regularization to avoid overfitting, this will also improve the generalization performance of neural networks. As shown in Figure 4a, W denotes the weight, L denotes the loss function, W L denotes the weight gradient, and Φ G C ( W L ) denotes the central gradient. As shown in Figure 4a, replace W L by Φ G C ( W L ) to implement embedding GC to the existing network optimizer. The GC is calculated as the mean of each column of the gradient matrix/tensor, and transfers the center of each column to zero mean as shown in Figure 4b.

3.3.1. GC Formula

As for the fully connected layer or convolutional layer, we assume that the gradients have been obtained by back propagation. As for the weight vector W i , its gradient is W i L ( i = 1 , 2 , , N ) , the subscript i represents the column i vector in the gradient matrix. The operator of GC, which is denoted by Φ G C , can be calculated as follows:
Φ G C ( W i L ) = W i L μ w i L
where μ W i L = 1 M j = 1 M W i , j L . As shown in Equation (6), we only need to calculate the mean of the column vector of weight matrix, then remove the mean from each column vector.

3.3.2. Properties of GC

The following is a theoretical analysis of how GC can improve the generalization ability of the model.
(1)
Weight space regularization: The projection of the weight gradient is able to constrain the weight space in a hyperplane, as shown in Figure 5, in which P is a projection matrix of hyperplane with normal vector e , and P W L is projection gradient. Firstly, projecting gradient W L to the hyperplane is determined by e T ( W W t ) = 0 , in which W t is the weight vector of the t times iteration, then weight W is updated along P W t L direction. It can be concluded that e T W t + 1 = e T W t = = e T W 0 , i.e., e T W is a constant during the training, so GC regularizes the solution space of W , thus decreasing the possibility of overfitting.
(2)
Output space regularization: After the introduction of GC, a constant intensity change in an input feature causes a change in the output activation, which is unrelated to the current weight vector. If the mean value of the initial weight vector converges to 0, then the output activation is insensitive to changes in the interference intensity of the input features, so the output feature space is more stable to the changes in the training samples.
Figure 5. The geometrical interpretation of GC.
Figure 5. The geometrical interpretation of GC.
Applsci 12 01789 g005
In this paper, the ppFDetector model uses Gradient Centralization to regularize the weight space and output space in order to avoid overfitting.

3.4. Convolutional Block Attention Module

Introducing an attention mechanism into a convolutional neural network can improve the performance of the model in classification tasks. Woo [32] proposed convolutional block attention module (CBAM). Considering that the traditional CNN network only uses convolution operation to extract different features in the images, we use the CBAM module to emphasize the meaningful features of channel and spatial axis. As shown in Figure 6, we use the channel attention module and the spatial attention module in turn, whose functions are to learn ‘what’ in the channel axis and ‘where’ in the spatial axis respectively.

3.4.1. Channel Attention

CBAM uses both average-pooled and max-pooled features, then feeds them sequentially into a weight-sharing Multilayer Perceptron (MLP), and finally uses element-wise summation to combine the output feature vectors. Channel attention mainly focuses on what features of the input image are meaningful. The channel attention module of CBAM is shown in Figure 7.

3.4.2. Spatial Attention

Adding the spatial attention module makes up for the lack of channel attention to a certain extent because spatial attention mainly focuses on which part of the input image has more effective information. In order to calculate spatial attention, a feature map is obtained through max pooling and average pooling, spliced into a 2D feature map, sent to the standard 7 × 7 convolution for parameter learning, and finally a 1D weight feature map is obtained. The spatial attention module of CBAM is shown in Figure 8.
In this paper, CBAM module is added to make the ppFDetector model pay more attention to the defect area, which can improve the performance of the proposed model.

3.5. Anomaly Detection Strategy

In the anomaly detection strategy, when the normal PV panel image is put into the trained model, the generator first encodes it into latent space vector and then decodes it into the reconstructed image. Because it obeys data distribution learned by the generator, the error between the input image and its reconstructed image is smaller than the threshold defined by the anomaly model. However, when the abnormal PV panel image is put into the model, the reconstructed image generated by the model is not equal to the input image. Since it doesn’t obey the data distribution learned by the generator, the error between the input image and its reconstructed image is bigger than the threshold. Therefore, the model detects abnormal PV panel by the error between the input image and its reconstructed image. However, the input image x is a three-channel color image that is stored in a matrix. If the error between the input image and its reconstructed image is directly calculated by the image matrix, the complexity of space and time is unacceptable. The latent space vector z can theoretically be assumed as the best representation of x with the smallest dimension. In order to reduce the time and space complexity of the proposed method, we used the error between latent space vectors z and z ˜ instead of the error between the input image and its reconstructed image.
In this paper, the error of the latent space vector denoted as M ( z , z ˜ ) can be calculated by the mean of the differences between the latent space vectors z and z ˜ . The relevant expressions are as follows:
M ( z , z ˜ ) = 1 N i = 1 N | z ( i ) z ˜ ( i ) | ,
where N is the length of vector z .
Finally, if M ( z , z ˜ ) > θ , the proposed model judges the test image as an abnormal PV panel, otherwise, it is normal.

4. Experiment

4.1. Dataset

The dataset is collected from a PV power plant located in Zhejiang province. The original image resolution is 3840 × 2048. In order to speed up the training process, we split the original images into a set of small images (i.e., 32 × 32 images as shown in Figure 9a). By segmenting the original image, 32,000 small images are obtained, of which 25,600 small images are included under the training dataset and 6400 under test set. An additional 3200 abnormal images are obtained from the negative samples by splitting defective images, as shown in Figure 9b.

4.2. Model Construction

The construction of the generator and the discriminator in the generative confrontation network is a key task. The generator includes an encoder and a decoder, and the discriminator is an encoder-like structure. The details of the encoder parameters and decoder parameters are shown in Table 1 and Table 2, respectively. The input image size of the encoder is 32 × 32 × 3, and a 4 × 4 convolution kernel is used. The first three layers of convolution use edge padding. The core step size is 2, and the batch normalization layer [33], the LeakyReLU activation function layer [34] and CBAM layer [32] are added after the convolution layer. The last layer of convolution directly outputs a hidden space vector with a size of 1 × 1 × 100. The decoder is symmetrical to the encoder structure, the first layer of transposed convolution has no padding and step size, and the last three layers of transposed convolution padding is 1, and the step size is 2. After the first three transposed convolutional layers, the batch normalization layer and the ReLU activation function layer are added, and the output size is 32 × 32 × 3 of the reconstruction image.
In this paper, three experiments are performed on where the CBAM layer is embedded. Firstly, the CBAM layer is embedded after the first convolution block of the encoder network G E . Secondly, the CBAM layer is embedded after both the first and second convolution blocks of G E . Thirdly, the CBAM layer is embedded after the first, second and third convolution blocks, as shown in Figure 3. The final results of the three different CBAM embedding methods are shown in Figure 10. It can be seen that the third embedding effect is the best, so the proposed ppFDetector model chooses the third CBAM embedding method, and the AUC reaches the maximum value.

4.3. Model Training

As shown in Figure 2, in the model training, only positive PV panel image x enters the encoder–decoder (S.2) to generate the reconstructed image x ˜ . Then it is divided into two branches. In the first branch, the reconstructed image x ˜ enters the discriminator through S.4, and then goes through S.5 to determine whether the result of the discriminator is correct or not. In the second branch, the reconstructed image x ˜ enters the encoder (S.3) to generate the latent space vector z ˜ of the reconstructed image. Then, the adversarial loss, encoder loss and contextual loss (S.7) are calculated and sent back to the generator. The generator and the discriminator are constructed by iterative confrontation training.

4.4. Model Validation

After the model training is completed, the model needs to be evaluated to ensure that efficient and accurate anomaly detection tasks are completed at minimum cost. Therefore, the model is evaluated from the following perspectives:
(1)
We evaluate the likelihood of the original image and its reconstructed image of the generator. As shown in Figure 11, in the training process, as the number of iterations increases, the difference between the original image and the reconstructed image gradually shrinks.
(2)
We evaluate the data distribution comparison between the original image and its reconstructed image for the positive sample and negative sample, respectively. Since only positive sample dataset X is fed into the generator G E D during the training process, G E D only learns by the data distribution P x of the normal PV panel images while the data distribution P y of the abnormal PV panel images is unknown. When a normal PV panel image x is put into the generator, the reconstructed image x ˜ generated by the generator is equal to the input image x because it obeys the data distribution learned by the generator, then its data distribution P x ˜ is infinitely close to P x as shown in Figure 12a. When an abnormal PV panel image y is put into the generator, G E D still encodes it into z y and then decodes it into y ˜ in a manner that obeys the data distribution P x . Due to the difference between P x and P y , y and y ˜ are different, and its data distribution P y ˜ is different from P y as shown in Figure 12b.
(3)
We evaluate the error between the latent space vectors of the original image and its reconstructed image. As shown in Figure 13a, the M ( z , z ˜ ) of the positive samples (in blue color) is distributed around 10, but the M ( z , z ˜ )   of the negative samples (in red color) is mainly distributed between 30 and 70. It can also be seen in Figure 13b. So, the M ( z , z ˜ )   of the positive samples and negative samples has a clear decision boundary.

4.5. Model Checking

As shown in Figure 2, in the anomaly detection process, the test PV panel image x R 256 × 256 is first split into a set of small images (i.e., 32 × 32 images as shown in Figure 9a). The small image x R 32 × 32 enters the encoder–decoder (T.2), to generate the latent space vector z and its reconstructed image. Next, the reconstructed image enters the encoder (T.3) to generate the latent space vector z ˜ . Then the error M ( z , z ˜ ) between latent space vectors z and z ˜ (T.4) is computed. If M ( z , z ˜ ) > θ , the test small image is judged as an abnormal PV panel, Otherwise, it is normal. According to the above method, each small image is judged whether it is abnormal or not in turn. Finally, we merge the 64 small image x R 32 × 32 back into the large original image x R 256 × 256 . As shown in Figure 14, a is the test image, and c is a line chart of the latent vector z of the test image and its reconstructed image. The horizontal axis represents the sequence number of the image block, and the total is 64. The vertical axis represents the value of latent vector z . Figure 14d is a scatter plot of the error M ( z , z ˜ ) of the test image, the green points are normal, but the red points are abnormal. It can be clearly seen that there is an obvious decision boundary between the normal point and the abnormal point. When we get the outliers in M ( z , z ˜ ) , and select the corresponding feature map as abnormal, the output result is shown in Figure 14b.

4.6. Model Evaluation

In order to verify the effectiveness of the proposed method, this paper compares the ppFDetector with AnoGAN [17], Zhao’s method [18], f-AnoGAN [19], GANomaly [20], and Pre-trained Vgg16. Considering that the minimum input size of Vgg16 is 48 × 48, we need to enlarge the image from 32 × 32 to 48 × 48. For a fully supervised model, pre-trained Vgg16, 10, 100 and 200 negative samples are respectively appended to the training set to explore the influence of different numbers of negative samples on the metric of accuracy.
We take ROC (Receiver Operating Characteristic curve) curve, precision, accuracy, F1 and sensitivity scores as the performance metrics. Precision represents the proportion of examples classified as positive examples that are actually so. The higher the index, the better the performance. Accuracy is the ratio of the number of samples correctly classified by the classifier to the total number of samples for a given test dataset. And the higher the score, the better the model. F1 score can be regarded as a weighted average of the model precision and recall. Its maximum value is 1, and its minimum value is 0. The larger the value, the better the model. Sensitivity refers to the proportion of all positive examples identified to all positive examples. The higher the sensitivity rate, the higher the probability that the actual abnormal sample is predicted. As shown in Table 3, the performance of a fully supervised learning method (i.e., Pre-trained Vgg16) improves by increasing the proportion of negative samples in the training process. Although the precision and accuracy performance metrics of vgg16(200) are higher than ppFDetector, ppFDetector does not require a negative sample during training, and so ppFDetector outperforms the Pre-trained Vgg16 method with limited negative samples. Table 4 shows the performance comparison between the ppFDetector, semi-supervised learning methods (i.e., Zhao’s method and GANomaly), and unsupervised learning methods (i.e., AnoGAN and f-AnoGAN). It shows that the ppFDetector only training with positive samples is generally competitive with state-of-the-art semi-supervised and unsupervised methods, and achieves higher accuracy, precision, F1 score and sensitivity.
Figure 15a illustrates the comparison of ROC curves of semi-supervised and unsupervised methods. The ppFDetector outperforms in PV panel fault detection. The AUC value of the ppFDetector is 0.943, which is 0.041 higher than GANomaly, and 0.052, 0.06 and 0.035 higher than Zhao’s method, AnoGAN and f-AnoGAN, respectively. Figure 15b illustrates the ROC curves of the ppFDetector and fully supervised method (i.e., Pre-trained Vgg16) with a different number of negative samples in the training process. The ppFDetector outperforms Pre-trained Vgg16 with limited negative samples, i.e., 10 and 100. But when the Pre-trained Vgg16 trained with more than 200 negative samples, its performance is close to ppFDetector.

4.7. Ablation Experiment

Figure 16a shows the comparison of the loss with GC and without GC in 50 rounds of training. The original network only takes Batch Normalization (BN) as optimization. As shown in Figure 16a, the training loss of BN+GC is reduced faster than BN, so GC can further speed up convergence of the model during training. Figure 16b compares SmoothL1Loss with the manual combination of L2loss and L1Loss. Using SmoothL1Loss is significantly lower than the one using the manual combination of L2loss and L1Loss. This proves that SmoothL1Loss can make the model more accurate. The GC is applied on conditional GAN of ppFDetector to regularize weight space and output feature space to prevent overfitting, so as to improve the generalization performance in the deep convolution network. Figure 16c is a comparison of the loss with and without the CBAM module. CBAM first extracts features based on the attention of the spatial domain, then extracts features based on the attention of the channel domain, and finally achieves the optimal attention extraction effect, thereby improving the performance of the model.

5. Conclusions

In this paper, with an aim to detect the fault of PV panels in a fast and accurate manner, a novel PV panel fault detection approach-ppFDetector is proposed. The ppFDetector does not require a large number of negative samples. It employs the generative adversarial network to extract image features to approximate the image data distribution of the positive samples. Thus, the model can generate a reconstructed image extremely similar to its positive samples. By calculating the differences between the original image and its reconstructed image, ppFDetector can accurately determine whether the PV panel is abnormal. This anomaly detection model is with strong transplantation. As there are a large number of positive samples, binary classifications can be performed after the model is trained; i.e., ppFDetector can solve the problem that models cannot train without negative samples. Moreover, ppFDetector accelerates the model training speed while increasing the generalization performance of the model through the application of gradient centralized. The ppFDetector also uses the CBAM and SmoothL1 loss function to further improve the accuracy and robustness of PV panel fault detection, which can replace manual inspection. In the future, we will work towards expanding the abnormal types that ppFDetector can detect by collecting abnormal datasets with various types of defects (such as dust cover, snail pattern and rust), and employ the CNN to classify the types of the defect after detecting abnormal areas.

Author Contributions

Conceptualization J.C.; Methodology, F.L. and R.N.; Software, R.N. and Z.Z.; Validation, R.N. and Z.Z.; Writing—original draft, F.L. and R.N.; Writing—review and editing, L.G. and J.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the Ministry of Education Humanities Social Sciences Research Project (No. 20YJC760101), the Educational Commission of Zhejiang Province of China (No. Y202147553), Zhejiang Provincial Natural Science Foundation of China (No. LGF20C050001) and the National Nature Science Foundation of China (No. 61805213).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Acknowledgments

We acknowledge a photovoltaic power plant located in Zhejiang province for providing data for our experiment.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Fan, Z.; Yang, Z.; Yang, L. Daylight performance assessment of atrium skylight with integrated semi-transparent photovoltaic for different climate zones in China. Build. Environ. 2021, 190, 107299. [Google Scholar] [CrossRef]
  2. Xuan, Q.; Li, G.; Jiang, B.; Zhao, X.; Zhao, B.; Ji, J.; Pei, G. Analysis and quantification of effects of the diffuse solar irradiance on the daylighting performance of the concentrating photovoltaic/daylighting system. Build. Environ. 2021, 193, 107654. [Google Scholar] [CrossRef]
  3. Liu, Y.; Zhang, R.; Huang, Z.; Cheng, Z.; López-Vicente, M.; Ma, X.; Wu, G. Solar photovoltaic panels significantly promote vegetation recovery by modifying the soil surface microhabitats in an arid sandy ecosystem. Land Degrad. Dev. 2019, 30, 2041–2047. [Google Scholar] [CrossRef]
  4. Du, W.Z.; Shen, H.Y. Information Technology: Investigators from Huazhong University of Science and Technology Target Information Technology (Defect Image Sample Generation with GAN for Improving Defect Recognition). Inf. Technol. Newsweekly 2020, 27, 24–33. [Google Scholar]
  5. Mohammed, S.; Boumediene, B.; Miloud, B. Assessment of PV modules degradation based on performances and visual inspection in Algerian Sahara. Int. J. Renew. Energy Res. 2016, 6, 106–116. [Google Scholar]
  6. Kontges, M. Review of Failures of Photovoltaic Modules; IEA-Photovoltaic Power Systems Programme T13-01; SUPSI, 2014; pp. 1–140. Available online: https://iea-pvps.org/wp-content/uploads/2020/01/IEA-PVPS_T13-01_2014_Review_of_Failures_of_Photovoltaic_Modules_Final.pdf (accessed on 3 February 2022).
  7. Tabanjat, A.; Becherif, M.; Hissel, D. Reconfiguration solution for shaded PV panels using switching control. Renew. Energy 2015, 82, 4–13. [Google Scholar] [CrossRef]
  8. Dolara, A.; Lazaroiu, G.; Leva, S.; Manzolini, G. Snail trails and cell micro-cracks impact on PV module maximum power and energy production. IEEE J. Photovolt. 2016, 6, 1269–1277. [Google Scholar] [CrossRef] [Green Version]
  9. Yang, T.; Zhao, L.Y.; Wang, C.S. Overview of application of Artificial Intelligence in Power System and Integrated Energy System. Autom. Electr. Power Syst. 2019, 43, 2–14. [Google Scholar]
  10. Zhou, N.C.; Liao, J.Q.; Wang, Q.G. Analysis and Prospect of Deep Learning Application in Smart Grid. Autom. Electr. Power Syst. 2019, 43, 180–191. [Google Scholar]
  11. Li, X.; Yang, Q.; Yan, W. Intelligent fault pattern recognition of aerial photovoltaic module images based on deep learning technique. Syst. Cybern. Inform. 2018, 16, 67–71. [Google Scholar]
  12. Tang, W.; Yang, Q.; Xiong, K.; Yan, W. Deep learning based automatic defect identification of photovoltaic module using electroluminescence images. Sol. Energy 2020, 201, 453–460. [Google Scholar] [CrossRef]
  13. Zhao, Y.; Zhan, K.; Wang, Z.; Shen, W. Deep learning-based automatic detection of multitype defects in photovoltaic modules and application in real production line. Prog. Photovolt. Res. Appl. 2021, 29, 471–484. [Google Scholar] [CrossRef]
  14. Mittal, S.; Dutta, M.K.; Issac, A. Non-destructive image processing based system for assessment of rice quality and defects for classification according to inferred commercial value. Measurement 2019, 148, 1792–1797. [Google Scholar] [CrossRef]
  15. Singh, A.P.; Kumar, A.; Akshayveer, O.P.S. Performance enhancement strategies of a hybrid solar chimney power plant integrated with photovoltaic panel. Energy Convers. Manag. 2020, 218, 113020. [Google Scholar] [CrossRef]
  16. Zhang, E.; Li, B.; Li, P.; Chen, Y. A Deep Learning Based Printing Defect Classification Method with Imbalanced Samples. Symmetry 2019, 11, 1440. [Google Scholar] [CrossRef] [Green Version]
  17. Schlegl, T.; Seebock, P.; Waldstein, S.M.; Schmidt-Erfurth, U.; Langs, G. Unsupervised anomaly detection with generative adversarial networks to guide marker discovery. In International Conference on Information Processing in Medical Imaging; Lecture Notes in Computer Science, 10265 LNCS; Springer: Cham, Switzerland, 2017; pp. 146–147. [Google Scholar]
  18. Zhao, Z.; Li, B.; Dong, R.; Zhao, P. A surface defect detection method based on positive samples. In Proceedings of the Pacific Rim International Conference on Artificial Intelligence (PRICAI), Nanjing, China, 28–31 August 2018; pp. 473–481. [Google Scholar]
  19. Schlegl, T.; Seeböck, P.; Waldstein, S.M.; Langs, G.; Schmidt-Erfurth, U. f-AnoGAN: Fast unsupervised anomaly detection with generative adversarial networks. Med. Image Anal. 2019, 54, 30–44. [Google Scholar] [CrossRef]
  20. Akcay, S.; Abarghouei, A.A.; Breckon, T.P. Ganomaly: Semi-supervised anomaly detection via adversarial training. In Proceedings of the ACCV, Perth, Australia, 2–6 December 2018; pp. 622–637. [Google Scholar]
  21. Radford, A.; Metz, L.; Chintala, S. Unsupervised Representation Learning with Deep Convolutional Generative Adversarial Networks. arXiv 2015, arXiv:1511.06434. [Google Scholar]
  22. Undirwade, A.; Das, S. Image anonymization using deep convolutional generative adversarial network. In Machine Learning Algorithms and Applications; Srinivas, M., Sucharitha, G., Matta, A., Chatterjee, P., Eds.; Crc Press: Boca Raton, FL, USA, 2021. [Google Scholar] [CrossRef]
  23. Karnewar, A.; Wang, O. MSG-GAN: Multi-Scale Gradients for Generative Adversarial Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 7796–7805. [Google Scholar]
  24. Tatapudi, S.; Libby, C.; Raupp, C. Defect and safety inspection of 6 PV technologies from 56000 modules representing 257000 modules in 4 climatic regions of the United States. In Proceedings of the 43rd IEEE Photovoltaic Specialists Conference, Portland, OR, USA, 5–10 June 2016; pp. 1747–1751. [Google Scholar]
  25. Sawyer, D.; Kessler, H.K. The solar cell is scanned by laser for displaying the working characteristics of the cell and detecting the defects of the cell. J. Electron. Equip. 1980, 27, 864–872. [Google Scholar]
  26. Tsuzuki, K.; Murakami, T.; Yoshino, T.; Takeyama, Y.; Shimizu, K. Inspection Method and Production Method of Solar Cell Modules. U.S. Patent 6,271,462, 7 August 2001. [Google Scholar]
  27. Tsai, D.M.; Wu, S.C.; Li, W.C. Defect detection of solar cells in electroluminescence images using Fourier image reconstruction. Sol. Energy Mater. Sol. Cells 2012, 99, 250–262. [Google Scholar] [CrossRef]
  28. Bastari, A.; Bruni, A.; Cristalli, C. Classification of silicon solar cells using electroluminescence texture analysis. In Proceedings of the IEEE International Symposium on Industrial Electronics, Bari, Italy, 4–7 July 2010; pp. 1722–1727. [Google Scholar]
  29. Quater, P.; Grimaccia, F.; Leva, S. Light unmanned aerial vehicles (UAVs) for cooperative inspection of PV plants. IEEE J. Photovolt. 2014, 4, 1107–1113. [Google Scholar] [CrossRef] [Green Version]
  30. Li, X.; Yang, Q.; Chen, Z. Visible defects detection based on UAV-based inspection in large-scale photovoltaic systems. IET Renew. Power Gener. 2017, 11, 1234–1244. [Google Scholar] [CrossRef]
  31. Yong, H.; Huang, J.; Hua, X.; Zhang, L. Gradient Centralization: A New Optimization Technique for Deep Neural Networks. In Proceedings of the ECCV, Cambridge, UK, 14–18 April 2020; pp. 635–652. [Google Scholar]
  32. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. CBAM: Convolutional Block Attention Module. In Proceedings of the ECCV, Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  33. Ioffe, S.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In Proceedings of the International Conference on International Conference on Machine Learning (ICICML), Lille, France, 6–11 July 2015; pp. 234–238. [Google Scholar]
  34. Xu, B.; Wang, N.; Chen, T. Empirical Evaluation of Rectified Activations in Convolutional Network. CoRR. 2015, pp. 14–23. Available online: http://arxiv.org/abs/1505.00853 (accessed on 3 February 2022).
Figure 1. The fishery light complementary large-scale PV power plant.
Figure 1. The fishery light complementary large-scale PV power plant.
Applsci 12 01789 g001
Figure 2. The integrated photovoltaic panel fault detection workflow of ppFDetector.
Figure 2. The integrated photovoltaic panel fault detection workflow of ppFDetector.
Applsci 12 01789 g002
Figure 3. Pipeline of the generative adversarial network for anomaly detection.
Figure 3. Pipeline of the generative adversarial network for anomaly detection.
Applsci 12 01789 g003
Figure 4. Schematic diagram of gradient centralization. (a) Schematic diagram of using GC; (b) GC operation of gradient matrix/weight tensor.
Figure 4. Schematic diagram of gradient centralization. (a) Schematic diagram of using GC; (b) GC operation of gradient matrix/weight tensor.
Applsci 12 01789 g004
Figure 6. Module of CBAM.
Figure 6. Module of CBAM.
Applsci 12 01789 g006
Figure 7. Channel attention module.
Figure 7. Channel attention module.
Applsci 12 01789 g007
Figure 8. Spatial attention module.
Figure 8. Spatial attention module.
Applsci 12 01789 g008
Figure 9. PV Segmentation images: (a) Positive sample; (b) Negative sample.
Figure 9. PV Segmentation images: (a) Positive sample; (b) Negative sample.
Applsci 12 01789 g009
Figure 10. ROC curves of different CBAM embedding methods.
Figure 10. ROC curves of different CBAM embedding methods.
Applsci 12 01789 g010
Figure 11. Original images and reconstructed images during training. (a) 10 iterations results; (b) 100 iterations results; (c) 200 iterations results.
Figure 11. Original images and reconstructed images during training. (a) 10 iterations results; (b) 100 iterations results; (c) 200 iterations results.
Applsci 12 01789 g011
Figure 12. Data distribution comparison between the original image and its reconstructed image. (a) Positive sample; (b) Negative sample.
Figure 12. Data distribution comparison between the original image and its reconstructed image. (a) Positive sample; (b) Negative sample.
Applsci 12 01789 g012
Figure 13. Plots of M ( z , z ˜ )   of the positive and negative samples. (a) Histogram; (b) Boxplot.
Figure 13. Plots of M ( z , z ˜ )   of the positive and negative samples. (a) Histogram; (b) Boxplot.
Applsci 12 01789 g013
Figure 14. Output results of abnormal detection. (a) Test image; (b) Output result of abnormal area; (c) Line chart of latent vector z ; (d) Scatter plot of Error M ( z , z ˜ ) .
Figure 14. Output results of abnormal detection. (a) Test image; (b) Output result of abnormal area; (c) Line chart of latent vector z ; (d) Scatter plot of Error M ( z , z ˜ ) .
Applsci 12 01789 g014
Figure 15. ROC curve comparison chart. (a) semi-supervised and unsupervised methods; (b) fully supervised methods.
Figure 15. ROC curve comparison chart. (a) semi-supervised and unsupervised methods; (b) fully supervised methods.
Applsci 12 01789 g015
Figure 16. Loss comparison chart. (a) BN, BN+GC loss curve; (b) L2Loss+L1Loss, SmoothL1Loss loss curve. (c) NO CBAM, CBAM loss curve.
Figure 16. Loss comparison chart. (a) BN, BN+GC loss curve; (b) L2Loss+L1Loss, SmoothL1Loss loss curve. (c) NO CBAM, CBAM loss curve.
Applsci 12 01789 g016
Table 1. Details of the encoder parameters.
Table 1. Details of the encoder parameters.
Layer 1InputFilter NumberKernelStridePaddingBatch NormalizationActivation FunctionAttention MechanismOutput
1(32,32,3)64(4,4)2Same (1)Batch NormLeakyReLU (0.2)CBAM(16,16,64)
2(16,16,64)128(4,4)2Same (1)Batch NormLeakyReLU (0.2)CBAM(8,8,128)
3(8,8,128)256(4,4)2Same (1)Batch NormLeakyReLU (0.2)CBAM(4,4,256)
4(4,4,256)100(4,4)1Valid (0)NoneNoneNone(1,1,100)
Table 2. Details of the decoder parameters.
Table 2. Details of the decoder parameters.
Layer 1InputFilter NumberKernelStridePaddingBatch NormalizationActivation FunctionOutput
1(1,1,100)256(4,4)1Valid (0)Batch NormReLU(4,4,256)
2(4,4,256)128(4,4)2Same (1)Batch NormReLU(8,8,128)
3(8,8,128)64(4,4)2Same (1)Batch NormReLU(16,16,64)
4(16,16,64)3(4,4)2Same (1)Nonetanh(32,32,3)
Table 3. Performance comparison with fully supervised model.
Table 3. Performance comparison with fully supervised model.
ModelPrecisionAccuracyF1Sensitivity
Pre-trained Vgg16 (10)0.3240.2650.3780.453
Pre-trained Vgg16 (100)0.7460.7500.7950.852
Pre-trained Vgg16 (200)0.9300.9600.9360.941
ppFDetector0.9290.9450.9440.960
Bold shows the best performance between the four models.
Table 4. Performance comparison with semi-supervised and unsupervised models.
Table 4. Performance comparison with semi-supervised and unsupervised models.
ModelPrecisionAccuracyF1Sensitivity
AnoGAN0.8780.7490.8520.827
Zhao’s method0.8690.8870.8990.931
GANomaly0.9080.8920.9290.952
f-AnoGAN0.9120.9300.9300.950
ppFDetector0.9290.9450.9440.960
Bold shows the best performance between the four models.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lu, F.; Niu, R.; Zhang, Z.; Guo, L.; Chen, J. A Generative Adversarial Network-Based Fault Detection Approach for Photovoltaic Panel. Appl. Sci. 2022, 12, 1789. https://doi.org/10.3390/app12041789

AMA Style

Lu F, Niu R, Zhang Z, Guo L, Chen J. A Generative Adversarial Network-Based Fault Detection Approach for Photovoltaic Panel. Applied Sciences. 2022; 12(4):1789. https://doi.org/10.3390/app12041789

Chicago/Turabian Style

Lu, Fangfang, Ran Niu, Zhihao Zhang, Lingling Guo, and Jingjing Chen. 2022. "A Generative Adversarial Network-Based Fault Detection Approach for Photovoltaic Panel" Applied Sciences 12, no. 4: 1789. https://doi.org/10.3390/app12041789

APA Style

Lu, F., Niu, R., Zhang, Z., Guo, L., & Chen, J. (2022). A Generative Adversarial Network-Based Fault Detection Approach for Photovoltaic Panel. Applied Sciences, 12(4), 1789. https://doi.org/10.3390/app12041789

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop