Next Article in Journal
Characterization of Superplastic Deformation Behavior for a Novel Al-Mg-Fe-Ni-Zr-Sc Alloy: Arrhenius-Based Modeling and Artificial Neural Network Approach
Next Article in Special Issue
Multimodal Unsupervised Speech Translation for Recognizing and Evaluating Second Language Speech
Previous Article in Journal
An Optimization Route Selection Method of Urban Oversize Cargo Transportation
Previous Article in Special Issue
Discovering Sentimental Interaction via Graph Convolutional Network for Visual Sentiment Prediction
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Rain Streak Removal for Single Images Using Conditional Generative Adversarial Networks

by
Prasad Hettiarachchi
1,2,
Rashmika Nawaratne
1,*,
Damminda Alahakoon
1,
Daswin De Silva
1 and
Naveen Chilamkurti
3
1
Research Centre for Data Analytics and Cognition, La Trobe University, Melbourne, VIC 3086, Australia
2
Department of Computer Science and Engineering, University of Moratuwa, Moratuwa 10400, Sri Lanka
3
Department of Computer Science and Information Technology, La Trobe University, Melbourne, VIC 3086, Australia
*
Author to whom correspondence should be addressed.
Appl. Sci. 2021, 11(5), 2214; https://doi.org/10.3390/app11052214
Submission received: 4 February 2021 / Revised: 24 February 2021 / Accepted: 25 February 2021 / Published: 3 March 2021
(This article belongs to the Special Issue Artificial Intelligence for Multimedia Signal Processing)

Abstract

:
Rapid developments in urbanization and smart city environments have accelerated the need to deliver safe, sustainable, and effective resource utilization and service provision and have thereby enhanced the need for intelligent, real-time video surveillance. Recent advances in machine learning and deep learning have the capability to detect and localize salient objects in surveillance video streams; however, several practical issues remain unaddressed, such as diverse weather conditions, recording conditions, and motion blur. In this context, image de-raining is an important issue that has been investigated extensively in recent years to provide accurate and quality surveillance in the smart city domain. Existing deep convolutional neural networks have obtained great success in image translation and other computer vision tasks; however, image de-raining is ill posed and has not been addressed in real-time, intelligent video surveillance systems. In this work, we propose to utilize the generative capabilities of recently introduced conditional generative adversarial networks (cGANs) as an image de-raining approach. We utilize the adversarial loss in GANs that provides an additional component to the loss function, which in turn regulates the final output and helps to yield better results. Experiments on both real and synthetic data show that the proposed method outperforms most of the existing state-of-the-art models in terms of quantitative evaluations and visual appearance.

1. Introduction

Rain is a common weather condition that negatively impacts computer vision systems. Raindrops appear as bright streaks in images due to their high velocity and light scattering. Since image recognition and detection algorithms are designed for clean inputs, it is essential to develop an effective mechanism for rain streak removal.
A number of research efforts have been reported in the literature focusing on restoring rain images, and different approaches have been taken. Some have attempted to remove rain streaks using video [1,2,3], while other researchers have focused on rain image recovery from a single image by considering the image as a signal separation task [4,5,6].
Since rain streaks overlap with background texture patterns, it is quite a challenging task to remove the rain streaks while maintaining the original texture in the background. Most of the times, this results in over-smoothed regions that are visible in the background after the de-raining process. De-raining algorithms [7,8] tend to over de-rain or under de-rain the original image. A key limitation in the traditional, handcrafted methods is that the feature learning is manual and designed to deal only with certain types of rain streaks, and they do not perform well with varying scales, shapes, orientations, and densities of raindrops [9,10]. In contrast, by using convolutional neural networks (CNNs), the feature learning process becomes an integral part of the algorithm and is able to unveil many hidden features. Convolutional neural network-based methods [11,12,13] have gained huge improvements in image de-raining during the last few years. These methods try to figure out a nonlinear mapping between the input rainy image and the expected ground truth image.
Still, there is potential for improvements and optimizations within CNN-based image de-raining algorithms, which could lead to more visually appealing and accurate results. Instead of being just constrained to characterizing rain streaks, visual quality should also be considered when defining the optimization functions, which will result in improving the visual appeal of test results. When defining the objective function, it should consider the fact that the performance of vision algorithms, such as classification/detection, should not be affected by the presence of rain streaks. The addition of this discriminative information ensures that the output is indistinguishable from its original counterpart.
Generative modeling is an unsupervised learning task in machine learning that involves automatically discovering and learning the patterns in input data in such a way that the model can be used to generate new examples that are indistinguishable from reality. The concept of generative adversarial networks (GANs) was originally presented in [14] and has gained a high level of interest, with several successful applications and directions reported within a short period in the machine learning community. Existing CNN-based mechanisms only consider either L 1 (Least Absolute Deviations) or L 2 (Least Square Errors) errors, whereas in conditional GANs, they have additional adversarial loss components, which result in very good, qualitative, visually appealing image outputs.
In our approach, we propose a conditional generative adversarial network-based framework for rain streak removal. Our model consists of a densely connected generator (G) network and a CNN-based discriminator (D) network. The generator network converts rainy images to de-rained images in such a way that it fools the discriminator network. In certain scenarios, traditional GANs tend to make output images more artificial and visually displeasing. To mitigate this issue, we have introduced a conditional CNN with skip connections for the generator. Skip connections guarantee better convergence by efficiently leveraging features from different layers of the network. The proposed model is based on the Pix2Pix framework by Isola et al. [15] and the conditional generative adversarial networks originally proposed by Fu et al. [16]. We have also used the source codes provided by authors of LPNet [17] and GMM [18] for quantitative and qualitative comparisons of the proposed model.
This paper makes the following contributions:
  • Propose a conditional, GAN-based deep learning architecture to remove rain streaks from images by adapting U-Net architecture-based CNN for single image de-raining.
  • Develop a classifier to identify whether the generated image is real or fake based on intra-convolutional “PatchGAN” architecture.
  • Due to the lack of access to the ground truth of rainy images, we present a new dataset synthesizing rainy images using real-world clean images, which are used as the ground truth counterpart in this research.
The paper is organized as follows: In Section 2, we provide an overview of related methods for image de-raining and the basic concepts behind cGANs. Section 3 describes the proposed model (CGANet—Conditional Generative Adversarial Network model) in detail with its architecture. Section 4 describes the experimental details with evaluation results. Section 5 provides the conclusion. Implementation details and the dataset used for the experiments are publicly available at GitHub (https://github.com/prasadmaduranga/CGANet (accessed on 11 December 2020)).

2. Related Work

In the past, numerous methods and research approaches have been proposed for image de-raining. These methods can be categorized as single image-based methods and video-based methods. With the evolution of neural networks, deep learning-based methods have become more dominant and efficient compared to past state-of-the-art methods.

2.1. Single Image-based Methods

Single image-based methods have limited access to information compared to video-based methods, which makes it more challenging to remove the rain streaks. Single image-based methods include low-rank approximations [3,19], dictionary learning [4,5,20], and kernel-based methods [21]. In [4], the authors decomposed the image into high- and low-frequency components and recognized the rain streaks by processing the high-frequency components. Other mechanisms have used gradients [22] and mixture models [18] to model and remove rain streaks. In [18], the authors introduced a patch-based prior for both clean and rainy layers using Gaussian mixture models (GMM). The GMM prior for rainy layers was learned from rainy images, while for the clean images, it was learned from natural images. Nonlocal mean filtering and kernel regression were used to identify rain streaks in [21].

2.2. Video-based Methods

With the availability of inter-frame information, video-based image de-raining is relatively more effective and easier compared to single image de-raining. Most research studies [1,23,24] have focused on detecting potential rain streaks using their physical characteristics and removing them using image restoration algorithms. In [25], the authors divided rain streaks into dense and sparse groups and removed the streaks using a matrix decomposition algorithm. Other methods have focused on de-raining in the Fourier domain [1] using Gaussian mixture models [23], matrix completions [24], and low-rank approximations [3].

2.3. Deep Learning based Methods

Deep learning-based methods have gained much popularity and success in a variety of high-level computer vision tasks in the recent past [26,27,28] as well as in image processing problems [29,30,31]. Deep learning was introduced for de-raining in [11] where a three-layer CNN was used for removing rain streaks and dirt spots in an image that had been taken through glass. In [12], a CNN was proposed for video-based de-raining, while a recurrent neural network was adopted by Liu in [32]. The authors in [33] proposed a residual-guide feature fusion network for single image de-raining. A pyramid of networks was proposed in [17], which used the domain-specific knowledge to reinforce the learning process.
CNNs learn to minimize a loss function, and the loss value itself decides the quality of output results. Significant design efforts and domain expertise are required to define an effective loss function. In other words, it is necessary to provide the CNN with what the user requires to minimize. Instead, if it is possible to set a high-level, general goal such as “make the output image indistinguishable from the target images”, then the CNN can automatically learn a loss function to satisfy the goal. This is the basic underlying concept behind generative adversarial networks (GANs).

2.4. Generative Adversarial Networks

Generative adversarial networks [14] are unsupervised generative models that contain two deep neural networks. The two neural networks are named as the generator (G) and discriminator (D) and are trained parallelly during the training process. GAN training can be considered to be a two-player min-max game where the generator and discriminator compete with each other to achieve each other’s targeted goal. The generator is trained to learn a mapping from a random noise vector (z) in latent space to an image (x) in a target domain: G(z) → x. The discriminator (D) learns to classify a given image as a real (output close to 1) image or a fake (output close to 0) image from the generator (G): D(x) → [0.1]. Both the generator and decimator can be considered as two separate neural networks trained from backpropagation, and they have separate loss functions. Figure 1 shows the high-level architecture of the proposed conditional GAN model. The generator will try to generate synthetic images that resemble real images to fool the discriminator. The discriminator learns how to identify the real images from the generated synthetic images from the generator.
The widest adaptation of GANs is for data augmentation, or that is to say, to learn from existing real-world samples and generate new samples consistent with the distribution. Generative modeling has been used in a wide range of application domains including computer vision, natural language processing, computer security, medicine, etc.
Xu et al. [34] used GANs for synthesizing image data to train and validate perception systems for autonomous vehicles. In addition to that, [35,36] used GANs for data fusion for developing image classification models while mitigating the issue of having smaller datasets. Furthermore, GANs were used for augmenting datasets for adversarial training in [37]. To increase the resolution of images, a super-resolution GAN was proposed by Ledig et al. [38], which took a low-resolution image as the input and generated a high-resolution image with 4× upscaling. To convert the image content from one domain to another, an image-to-image translation approach was proposed by Isola et al. [15] using CGANs. Roy et al. [39] proposed a TriGAN, which could solve the problem of image translation by adapting multiple source domains. Experiments showed that the SeqGAN proposed in [40] outperformed the traditional methods used for music and speech generation. In the computer security domain, Hu and Tan [41] proposed a GAN-based model to generate malware. For private product customization, Hwang et al. [42] proposed GANs to manufacture medical products.

3. Proposed Model

The proposed approach uses image-to-image translation for the image de-raining task. In a GAN, the generator produces the output based on the latent variable or the noise variable (z). However, in the proposed approach, it is necessary for a correlation to exist between the source image and the generator output image. We have applied the conditional GAN [16], which is a variant of the traditional GAN that takes additional information, y, as the input. In this case, we provide a source image with rain streaks as additional information for both the generator and discriminator. x represents the target image.
The objective of a conditional GAN is as follows:
LcGAN (G. D) = Ex~pdata(X)[log(D(x. y))] + Ez~pz(Z)[log(1 − D(G(z. y). y))
where pdata(X) denotes the real data probability distribution defined in the data space X, and pz(Z) denotes the probability distribution of the latent variable z defined on the latent space Z. Ex~pdata(X) and Ez~pz(Z) represent the expectations over the data spaces X and Z respectively. G(.) and D(.) represent the non-linear mappings of the generator and discriminator networks respectively.
In an image de-raining task, the higher-order color and texture information has to be preserved during the image translation. This has a significant impact on the visual performance of the output. Adversarial loss alone is not sufficient for this task. The loss function should be optimized so that it penalizes the perceptual differences between the output image and the target image.
Our implementation architecture is based on the work of Isola’s [15] Pix2Pix framework. It learns a mapping from an input image to an output image along with the objective function to train the model. In Pix2Pix, it suggests L 1 (mean absolute error) loss instead of L 2 (mean squared error) loss for the GAN objective function, since it encourages less blurring in the generator output. L1 loss averages the pixel level absolute difference between the target image and the generated image G(z. y) over the image space x.y.z.
L1(G) = Ex.y.z[‖ xG(z. y) ‖]
Finally, the loss function for this work is as follows:
L(G. D) = LcGAN (G. D) + λ L1(G)
Lambda (λ) is a hyperparameter that controls the weights of the terms. In this case, we kept lambda = 100 [15]. When training the model, lambda was increased to train a discriminator and minimized to train a generator. The final objective was to identify the generator G* by solving the following optimization problem:
G *   = arg   m i n G     m a x D   ( L c G A N ( G , D ) + λ L 1 ( G ) )

Model Overview

  • Generator Network
In image-to-image translations, it is necessary to map a high-resolution input grid to a high-resolution output grid. Though the input and output images differ in appearance, they share the same underlying structure, and as such, it is necessary to consider this factor when designing the generator architecture. Most previous work used the encoder-decoder network [43] for such scenarios. In encoder-decoder CNN, the input is progressively downsampled until the bottleneck layer, where the process gets reversed and starts to upsample the input data. Convolutional layers use 4 × 4 filters and strides with size 2 for downsampling. The same size of kernel is used for transpose convolution operation during upsampling. Each convolution/deconvolution operation is followed by batch normalization and Rectified Linear Unit (ReLU) activation. Weights of the generator are updated depending on the adversarial loss of the discriminator and the L 1 loss of the generator. Architecture details are shown in Table 1.
These networks require all the input information to pass through each of the middle layers. In most of the image-to-image translation problems, it is desirable to share the feature maps across the network since both input and output images represent the same underlying structure. For this purpose, we added a skip connection while following the general configuration of a “U-Net” [44]. Skip connections simply concatenate the channels at the ith layer with the channels at the (n–i)th layer.
  • Discriminator Network
We adapted PatchGAN architecture [45] for the discriminator, which penalized the structure at the scale of patches. It tried to classify each N × N patch as either real or fake. Final output of the discriminator (D) was calculated by averaging the received responses by running the discriminator convolutionally across the image. In this case, the patch was 30 × 30 in size, and each convolutional layer was followed by a ReLU activation and batch normalization. Zero-padding layers were used to preserve the edge details of the input feature maps during the convolution. Discriminator architecture is described in Table 2.

4. Experimental Details

This section discusses the experimental details of our proposed CGANet model and the quality matrices used to evaluate the performance of the proposed model. CGANet performance is compared with two other state-of-the-art methods: the Gaussian mixture model [18] and lightweight pyramid networks [17]. The algorithm implementation was conducted using Python and TensorFlow 2.0 [46]. CGANet was trained on a computer with a 2.2 GHz, 6-core Intel core i7 processor, 16 GB memory, and an AMD Radeon Pro 555X GPU.

4.1. Dataset

The training set consisting of 1500 images was chosen from a global road damage detection challenge dataset [47]. Rain streaks of different angles and intensities have been added to those images using Photoshop to create a synthesized rainy image set. Corresponding clean images become the target ground truth image set for the synthesized rainy image set. The test set consists of both synthesized and real-world rainy images. Three hundred synthesized images were chosen from the global road damage detection challenge dataset and pre-processed similarly when preparing the training set. Test dataset outputs are shown in Figure 2 as a comparison between the proposed CGANet model and the state-of-the-art de-raining methods. Real-world rainy images were taken from the internet, and they were considered only for demonstrating the effectiveness of the CGANet model. Since ground truth images were not available for the real-world rainy images, they were not taken into the account when training the model. Test results of real-world images are shown in Figure 3.

4.2. Evaluation Matrix and Results

The peak signal-to-noise ratio (PSNR) and structural similarity index (SSIM) [48] were used to evaluate and compare the performance of the model. PSNR measures how far the de-rained image is distorted from its real ground truth image by using the mean squared error at the pixel level. As shown in Table 1, the proposed CGANet model obtained the best PSNR value compared to the other two methods. The structural similarity index (SSIM) is a perception-based index that evaluates image degradation as the perceived difference in structural information while also incorporating both luminance masking and contrast masking terms. Table 3 shows the SSIM value comparison between the proposed CGANet model and the other two state-of-the art methods. By referring to this comparison, we could verify that the proposed method performed well compared to other de-raining mechanisms, and this is also visually verifiable in Figure 2 and Figure 3.

4.3. Parameter Settings

To optimize the proposed model, we followed the findings provided in the original GAN paper [14]. Instead of training the generator to minimize log(1 − D(x; G(x; z)), we trained it to maximize log D(x; G(x; z)). Since the discriminator could be trained much faster compared to the generator, we divided the discriminator loss by 2 while optimizing the discriminator. As such, the discriminator training speed slowed down compared to the generator. Both the discriminator and generator models were trained with an Adam optimizer [49] with a learning rate of 0.0002 and a momentum parameter β1 of 0.5 [15]. The model was trained using 150 epochs and updated after each image, and as such, the batch size was 1.

5. Conclusions

In this paper, we have proposed a single image de-raining model based on conditional generative adversarial networks and a Pix2Pix framework. The model consists of two neural networks: a generator network to map rainy images to de-rained images, and a discriminator network to classify real and generated de-rained images. Different performance matrices were used to evaluate the performance of the new model using both synthesized and real-world image data. The evaluations proved that the proposed CGANet model outperformed the state-of-the-art methods for image de-raining. The new CGANet model is presented as a high-potential approach for successful de-raining of images.
This paper is focused on image de-raining; however, the proposed model applies equally well to any other image translation problem in a different domain. In future developments, further analysis can be carried out to optimize the loss function by incorporating more comprehensive components with local and global perceptual information.

Author Contributions

Conceptualization, P.H., D.A., and R.N.; methodology, P.H., and R.N.; investigation, P.H.; data curation, P.H.; writing—original draft preparation, P.H.; writing—review and editing, R.N., D.A., D.D.S., and N.C.; supervision, D.A., D.D.S., and N.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data and source code for the experiments are publicly available at https://github.com/prasadmaduranga/CGANet (accessed on 11 December 2020).

Acknowledgments

This work was supported by a La Trobe University Postgraduate Research Scholarship.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Barnum, P.C.; Narasimhan, S.; Kanade, T. Analysis of Rain and Snow in Frequency Space. Int. J. Comput. Vis. 2009, 86, 256–274. [Google Scholar] [CrossRef]
  2. Brewer, N.; Liu, N. Using the Shape Characteristics of Rain to Identify and Remove Rain from Video. In Constructive Side-Channel Analysis and Secure Design; Springer: Berlin/Heidelberg, Germany, 2008; pp. 451–458. [Google Scholar]
  3. Chen, Y.-L.; Hsu, C.-T. A Generalized Low-Rank Appearance Model for Spatio-temporally Correlated Rain Streaks. In Proceedings of the 2013 IEEE International Conference on Computer Vision, Sydney, Australia, 1–8 December 2013; pp. 1968–1975. [Google Scholar]
  4. Kang, L.-W.; Lin, C.-W.; Fu, Y.-H. Automatic Single-Image-Based Rain Streaks Removal via Image Decomposition. IEEE Trans. Image Process. 2012, 21, 1742–1755. [Google Scholar] [CrossRef]
  5. Huang, D.-A.; Kang, L.-W.; Wang, Y.-C.F.; Lin, C.-W. Self-Learning Based Image Decomposition with Applications to Single Image Denoising. IEEE Trans. Multimedia 2013, 16, 83–93. [Google Scholar] [CrossRef]
  6. Sun, S.-H.; Fan, S.-P.; Wang, Y.-C.F. Exploiting image structural similarity for single image rain removal. In Proceedings of the 2014 IEEE International Conference on Image Processing (ICIP), Paris, France, 28 October 2014; pp. 4482–4486. [Google Scholar]
  7. Yang, W.; Tan, R.T.; Feng, J.; Liu, J.; Guo, Z.; Yan, S. Deep Joint Rain Detection and Removal from a Single Image. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 1685–1694. [Google Scholar]
  8. Fu, X.; Huang, J.; Zeng, D.; Huang, Y.; Ding, X.; Paisley, J. Removing Rain from Single Images via a Deep Detail Network. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 1715–1723. [Google Scholar]
  9. Zhang, X.; Li, H.; Qi, Y.; Leow, W.K.; Ng, T.K. Rain Removal in Video by Combining Temporal and Chromatic Properties. In Proceedings of the 2006 IEEE International Conference on Multimedia and Expo, Toronto, Canada, 9–12 July 2006; pp. 461–464. [Google Scholar]
  10. Liu, P.; Xu, J.; Liu, J.; Tang, X. Pixel Based Temporal Analysis Using Chromatic Property for Removing Rain from Videos. Comput. Inf. Sci. 2009, 2, 53. [Google Scholar] [CrossRef]
  11. Eigen, D.; Krishnan, D.; Fergus, R. Restoring an Image Taken through a Window Covered with Dirt or Rain. In Proceedings of the 2013 IEEE International Conference on Computer Vision, Sydney, Australia, 1–8 December 2013; pp. 633–640. [Google Scholar]
  12. Chen, J.; Tan, C.-H.; Hou, J.; Chau, L.-P.; Li, H. Robust Video Content Alignment and Compensation for Rain Removal in a CNN Framework. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 6286–6295. [Google Scholar]
  13. Fu, X.; Huang, J.; Ding, X.; Liao, Y.; Paisley, J. Clearing the Skies: A Deep Network Architecture for Single-Image Rain Removal. IEEE Trans. Image Process. 2017, 26, 2944–2956. [Google Scholar] [CrossRef] [Green Version]
  14. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 8–13 December 2014; pp. 2672–2680. [Google Scholar]
  15. Isola, P.; Zhu, J.-Y.; Zhou, T.; Efros, A.A. Image-to-Image Translation with Conditional Adversarial Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 5967–5976. [Google Scholar]
  16. Mirza, M.; Osindero, S. Conditional generative adversarial nets. arXiv 2014, arXiv:1411.1784. [Google Scholar]
  17. Fu, X.; Liang, B.; Huang, Y.; Ding, X.; Paisley, J. Lightweight pyramid networks for image deraining. arXiv 2019, arXiv:1805.06173. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  18. Li, Y.; Tan, R.T.; Guo, X.; Lu, J.; Brown, M.S. Rain Streak Removal Using Layer Priors. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 2736–2744. [Google Scholar]
  19. Chang, Y.; Yan, L.; Zhong, S. Transformed Low-Rank Model for Line Pattern Noise Removal. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 1735–1743. [Google Scholar]
  20. Luo, Y.; Xu, Y.; Ji, H. Removing Rain from a Single Image via Discriminative Sparse Coding. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 3397–3405. [Google Scholar]
  21. Kim, J.H.; Lee, C.; Sim, J.Y.; Kim, C.S. Single-image de-raining using an adaptive nonlocal means filter. In Proceedings of the 2013 IEEE International Conference on Image Processing, Melbourne, Australia, 15–18 September 2013; pp. 914–917. [Google Scholar]
  22. Zhu, L.; Fu, C.-W.; Lischinski, D.; Heng, P.-A. Joint Bi-layer Optimization for Single-Image Rain Streak Removal. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2545–2553. [Google Scholar]
  23. Bossu, J.; Hautière, N.; Tarel, J.-P. Rain or Snow Detection in Image Sequences through Use of a Histogram of Orientation of Streaks. Int. J. Comput. Vis. 2011, 93, 348–367. [Google Scholar] [CrossRef]
  24. Kim, J.H.; Sim, J.Y.; Kim, C.S. Video de-raining and desnowing using temporal correlation and low-rank matrix completion. IEEE Trans. Image Process. 2015, 24, 2658–2670. [Google Scholar] [CrossRef]
  25. Ren, W.; Tian, J.; Han, Z.; Chan, A.; Tang, Y. Video desnowing and de-raining based on matrix decomposition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4210–4219. [Google Scholar]
  26. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. In Proceedings of the Advances in Neural Information Processing Systems, Lake Tahoe, NV, USA, 3–6 December 2012; pp. 1097–1105. [Google Scholar]
  27. Gong, M.; Zhao, J.; Liu, J.; Miao, Q.; Jiao, L. Change Detection in Synthetic Aperture Radar Images Based on Deep Neural Networks. IEEE Trans. Neural Netw. Learn. Syst. 2016, 27, 125–138. [Google Scholar] [CrossRef]
  28. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. arXiv 2015, arXiv:1512.03385. [Google Scholar]
  29. Hou, W.; Gao, X.; Tao, D.; Li, X. Blind Image Quality Assessment via Deep Learning. IEEE Trans. Neural Netw. Learn. Syst. 2014, 26, 1275–1286. [Google Scholar] [CrossRef]
  30. Nawaratne, R.; Kahawala, S.; Nguyen, S.; De Silva, D. A Generative Latent Space Approach for Real-time Road Surveillance in Smart Cities. IEEE Trans. Ind. Inform. 2020, 1, 1. [Google Scholar] [CrossRef]
  31. Nawaratne, R.; Alahakoon, D.; De Silva, D.; Yu, X. Spatiotemporal Anomaly Detection Using Deep Learning for Real-Time Video Surveillance. IEEE Trans. Ind. Inform. 2020, 16, 393–402. [Google Scholar] [CrossRef]
  32. Liu, J.; Yang, W.; Yang, S.; Guo, Z. Erase or Fill? Deep Joint Recurrent Rain Removal and Reconstruction in Videos. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3233–3242. [Google Scholar]
  33. Fan, Z.; Wu, H.; Fu, X.; Hunag, Y.; Ding, X. Residual-guide feature fusion network for single image de-raining. arXiv 2018, arXiv:1804.07493. [Google Scholar]
  34. Xu, W.; Souly, N.; Brahma, P.P. Reliability of GAN Generated Data to Train and Validate Perception Systems for Autonomous Vehicles. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Online Conference. 9 January 2021; pp. 171–180. [Google Scholar]
  35. Saha, S.; Sheikh, N. Ultrasound Image Classification using ACGAN with Small Training Dataset. arXiv 2021, arXiv:2102.01539. [Google Scholar]
  36. Roy, S.; Sangineto, E.; Sebe, N.; Demir, B. Semantic-fusion gans for semi-supervised satellite image clas-sification. In Proceedings of the 2018 25th IEEE International Conference on Image Processing (ICIP), Athens, Greece, 7–10 October 2018; pp. 684–688. [Google Scholar]
  37. Cohen, J.; Rosenfeld, E.; Kolter, Z. Certified adversarial robustness via randomized smoothing. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 10–15 June 2019; pp. 1310–1320. [Google Scholar]
  38. Ledig, C.; Theis, L.; Huszar, F.; Caballero, J.; Cunningham, A.; Acosta, A.; Aitken, A.; Tejani, A.; Totz, J.; Wang, Z.; et al. Photo-Realistic Single Image Super-Resolution Using a Generative Adversarial Network. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 5892–5900. [Google Scholar]
  39. Roy, S.; Siarohin, A.; Sangineto, E.; Sebe, N.; Ricci, E. TriGAN: Image-to-image translation for multi-source domain adaptation. Mach. Vis. Appl. 2021, 32, 1–12. [Google Scholar] [CrossRef]
  40. Yu, L.; Zhang, W.; Wang, J.; Yu, Y. SeqGAN: Sequence generative adversarial nets with policy gradient. In Proceedings of the AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–10 February 2017; pp. 2852–2858. [Google Scholar]
  41. Hu, W.; Tan, Y. Generating Adversarial Malware Examples for Black-Box Attacks Based on GAN. Available online: https://arxiv.org/abs/1702.05983 (accessed on 11 December 2020).
  42. Hwang, J.-J.; Azernikov, S.; Efros, A.A.; Yu, S.X. Learning beyond human expertise with generative models for dental restorations. arXiv 2018, arXiv:1804.00064. [Google Scholar]
  43. Hinton, E.G.; Salakhutdinov, R.R. Reducing the dimensionality of data with neural networks. Science 2006, 313, 504–507. [Google Scholar] [CrossRef] [Green Version]
  44. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  45. Li, C.; Wand, M. Precomputed Real-Time Texture Synthesis with Markovian Generative Adversarial Networks. In Proceedings of the 14th European Conference on Computer Vision (ECCV) Amsterdam, The Netherlands, 11–14 October 2016; pp. 702–716. [Google Scholar]
  46. Abadi, M.; Barham, P.; Chen, J.; Chen, Z.; Davis, A.; Dean, J.; Zheng, X. Tensorflow: A system for large-scale machine learning. In Proceedings of the 12th {USENIX} Symposium on Operating Systems Design and Implementation, Savannah, GA, USA, 2–4 November 2016; pp. 265–283. [Google Scholar]
  47. Arya, D.; Maeda, H.; Ghosh, S.K.; Toshniwal, D.; Mraz, A.; Kashiyama, T.; Sekimoto, Y. Transfer learning-based road damage detection for multiple countries. arXiv 2020, arXiv:2008.13101. [Google Scholar]
  48. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  49. Da, K. A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
Figure 1. High-level architecture of the proposed model (CGANet).
Figure 1. High-level architecture of the proposed model (CGANet).
Applsci 11 02214 g001
Figure 2. Qualitative comparison between GMM, pyramid networks, and proposed CGANet methods.
Figure 2. Qualitative comparison between GMM, pyramid networks, and proposed CGANet methods.
Applsci 11 02214 g002
Figure 3. CGANet on real-world dataset ((Left) Input image; (Right) de-rained output).
Figure 3. CGANet on real-world dataset ((Left) Input image; (Right) de-rained output).
Applsci 11 02214 g003
Table 1. Generator architecture of the CGANet model.
Table 1. Generator architecture of the CGANet model.
Generator Architecture
Input(256 × 256), Num_c = 3
Downsampling: 4 × 4 Convolution + BN + ReLu, Output: 128 × 128, Num_c: 64
Downsampling 4 × 4 Convolution + BN + ReLu, Output: 64 × 64, Num_c: 128
Downsampling 4 × 4 Convolution + BN + ReLu, Output: 64 × 64, Num_c: 128
Downsampling: 4 × 4 Convolution + BN + ReLu, Output: 32 × 32, Num_c: 256
Downsampling: 4 × 4 Convolution + BN + ReLu, Output: 16 × 16, Num_c: 512
Downsampling 4 × 4 Convolution + BN + ReLu, Output: 128 × 128, Num_c: 512
Downsampling 4 × 4 Convolution + BN + ReLu, Output: 8×8, Num_c: 512
Downsampling 4 × 4 Convolution + BN + ReLu, Output: 4 × 4, Num_c: 512
Downsampling 4 × 4 Convolution + BN + ReLu, Output: 2 × 2, Num_c: 512
Downsampling 4 × 4 Convolution + BN + ReLu, Output: 1 × 1, Num_c: 512
Upsampling 4 × 4 Convolution + BN + ReLu, Output: 2 × 2, Num_c: 512
Concatenation: Input (2×2×512), (2 × 2 × 512), Output (2 × 2 × 1024)
Upsampling 4 × 4 Convolution + BN + ReLu, Output: 4 × 4, Num_c: 512
Concatenation: Input (4 × 4 × 512), (4 × 4 × 512), Output (4 × 4 × 1024)
Upsampling 4 × 4 Convolution + BN + ReLu, Output: 8 × 8, Num_c: 512
Concatenation: Input (8×8×512), (8×8×512), Output (8 × 8 × 1024)
Upsampling: 4 × 4 Transpose Convolution + BN + ReLu, Output: 16 × 16, Num_c: 512
Concatenation: Input (16 × 16 × 512), (16 × 16 × 512), Output (16 × 16 × 1024)
Upsampling: 4 × 4 Transpose Convolution + BN + ReLu, Output: 32 × 32, Num_c: 256
Concatenation: Input (32 × 32 × 256), (32 × 32 × 256), Output (32 × 32 × 512)
Upsampling: 4 × 4 Transpose Convolution + BN + ReLu, Output: 64 × 64, Num_c: 128
Concatenation: Input (64 × 64 × 128), (64 × 64 × 128), Output (64 × 64 × 256)
Upsampling: 4 × 4 Transpose Convolution + BN + ReLu, Output: 128 × 128, Num_c: 64
Concatenation: Input (128 × 128 × 64), (128 × 128 × 64), Output (128 × 128 × 128)
Upsampling: 4 × 4 Transpose Convolution, Output: 256×256, Num_c: 3
Table 2. Discriminator architecture of the CGANet model.
Table 2. Discriminator architecture of the CGANet model.
Discriminator Architecture
Input Image (256 × 256 × 3) + Target Image (256 × 256 × 3)
Concatenation: Input (256 × 256 × 3), (256 × 256 × 3), Output (2 × 2 × 1024)
Downsample 4 × 4 Convolution + BN + ReLu, Output: 128 × 128, Num_c: 64
Downsample 4 × 4 Convolution + BN + ReLu, Output: 64 × 64, Num_c: 128
Downsample 4 × 4 Convolution + BN + ReLu, Output: 32 × 32, Num_c: 256
Zero Padding 2D: Output: 34 × 34, Num_c: 256
Downsample 4 × 4 Convolution + BN + ReLu, Output: 31 × 31, Num_c: 512
Zero Padding 2D: Output: 33×33, Num_c: 512
Downsample 4 × 4 Convolution + BN + ReLu, Output: 30 × 30, Num_c: 1
Table 3. Quantitative comparison between different de-raining methods (mean ± STD).
Table 3. Quantitative comparison between different de-raining methods (mean ± STD).
IndexPyramidGMMCGANet
PSNR23.48 ± 2.0924.37 ± 2.1525.85 ± 1.57
SSIM0.731 ± 0.060.762 ± 0.060.768 ± 0.04
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Hettiarachchi, P.; Nawaratne, R.; Alahakoon, D.; De Silva, D.; Chilamkurti, N. Rain Streak Removal for Single Images Using Conditional Generative Adversarial Networks. Appl. Sci. 2021, 11, 2214. https://doi.org/10.3390/app11052214

AMA Style

Hettiarachchi P, Nawaratne R, Alahakoon D, De Silva D, Chilamkurti N. Rain Streak Removal for Single Images Using Conditional Generative Adversarial Networks. Applied Sciences. 2021; 11(5):2214. https://doi.org/10.3390/app11052214

Chicago/Turabian Style

Hettiarachchi, Prasad, Rashmika Nawaratne, Damminda Alahakoon, Daswin De Silva, and Naveen Chilamkurti. 2021. "Rain Streak Removal for Single Images Using Conditional Generative Adversarial Networks" Applied Sciences 11, no. 5: 2214. https://doi.org/10.3390/app11052214

APA Style

Hettiarachchi, P., Nawaratne, R., Alahakoon, D., De Silva, D., & Chilamkurti, N. (2021). Rain Streak Removal for Single Images Using Conditional Generative Adversarial Networks. Applied Sciences, 11(5), 2214. https://doi.org/10.3390/app11052214

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop