Next Article in Journal
A Method of Implementing a 4 × 4 Correlation Matrix for Evaluating the Uplink Channel Properties of MIMO Over-the-Air Apparatus
Previous Article in Journal
The Accuracy and Precision of Gait Spatio-Temporal Parameters Extracted from an Instrumented Sock during Treadmill and Overground Walking in Healthy Subjects and Patients with a Foot Impairment Secondary to Psoriatic Arthritis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Photo-Realistic Image Dehazing and Verifying Networks via Complementary Adversarial Learning

Department of Image, Chung-Ang University, Seoul 06974, Korea
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(18), 6182; https://doi.org/10.3390/s21186182
Submission received: 9 August 2021 / Revised: 6 September 2021 / Accepted: 13 September 2021 / Published: 15 September 2021
(This article belongs to the Topic Artificial Intelligence in Sensors)

Abstract

:
Physical model-based dehazing methods cannot, in general, avoid environmental variables and undesired artifacts such as non-collected illuminance, halo and saturation since it is difficult to accurately estimate the amount of the illuminance, light transmission and airlight. Furthermore, the haze model estimation process requires very high computational complexity. To solve this problem by directly estimating the radiance of the haze images, we present a novel dehazing and verifying network (DVNet). In the dehazing procedure, we enhanced the clean images by using a correction network (CNet), which uses the ground truth to learn the haze network. Haze images are then restored through a haze network (HNet). Furthermore, a verifying method verifies the error of both CNet and HNet using a self-supervised learning method. Finally, the proposed complementary adversarial learning method can produce results more naturally. Note that the proposed discriminator and generators (HNet & CNet) can be learned via an unpaired dataset. Overall, the proposed DVNet can generate a better dehazed result than state-of-the-art approaches under various hazy conditions. Experimental results show that the DVNet outperforms state-of-the-art dehazing methods in most cases.
Keywords:
dehazing; GAN; CNN

1. Introduction

In outdoor environments, acquired images lose important information such as contrast and salient edges because the particles attenuate the visible light. This degradation is referred to as hazy degradation, which distorts both spatial and color features and decreases visibility of the outdoor object. If the hazy degradation is not restored, we cannot expect a good performance of main image processing or image analysis methods such as object detection, image matching, and imaging systems [1,2,3,4], to name a few. Therefore, the common goal of dehazing algorithms is to enhance the edge and contrast while suppressing intensity or color saturation. To the best of the authors’ knowledge, Middleton and Edgar were the first to employ a physical haze model for the dehazing problem [5].
To generate the haze-free image using the physical model, atmospheric light and the corresponding transmission should be estimated. However, an accurate estimation of the atmospheric light and transmission map generally requires additional information, such as a pair of polarized images, multiple images under different weather conditions, distance maps, or user interactions [6,7,8,9]. For that reason, many state-of-the-art approaches try to find a better method to estimate the atmospheric light and the transmission map based on reasonable assumptions [10,11,12,13]. He et al. proposed a dark channel prior (DCP)-based haze removal method [14]. They assumed that pixels in the local patch of a clear image have at least one dark pixel. The DCP method works well in most regions that satisfy the DCP assumption, but fails in a white object region. Berman et al. estimated the transmission map using haze-line prior assumption that the pixel coordinates in the color space tend to become closer to the atmospheric light in a hazy image [15]. To find the lower bound of a haze-line, they used the 500 representative colors. While the Berman’s approach enhances color contrast, it is impossible to find representative colors in a severely degraded image by haze or fog. Shin et al. optimized the transmission estimation process using both radiance and reflectance components [16].
Recently, convolutional neural networks (CNN) are being applied not only to image classification, but also to variety of low-level image processing applications [17,18,19,20]. The CNN-based dehazing methods were also proposed in the literature to overcome the limitation of the transmission map estimation using a single image. Cai et al. estimated the transmission to restore a haze image using a DehazeNet [21]. Cai’s method falls in the end-to-end supervised learning approach using synthetic haze and clean patches. To overcome the limitation of haze feature estimation, Ren et al. presented a multi-scale CNN [22]. They also proposed a learning method using the pairs of the simulated haze image and true transmission [23].
To increase the training accuracy, Li et al. combined two CNN modules of the transmission and atmospheric light estimation via all-in-one dehazing network (AODNet) [24].
Zhang et al. proposed a densely connected pyramid dehazing network (DCPDN) optimized by a conditional adversarial learning method [25,26].
The depth information can be incorporated into the transmission estimation process using a supervised learning method. However, it is hard to reflect other quantities such as attenuation, atmospheric light, and illuminance at once because it is difficult to collect the data including the depth, attenuation, airlight, and ideal illuminance maps.
For example, Figure 1a shows a real-haze image provided by [27]. This type of haze in Figure 1a is different from what we have simulated, and degraded by multiple factors including the color attenuation, unbalanced light source and scattered light. Therefore, CNN-based estimation can not adaptively remove this real-haze as shown Figure 1b,c.
To overcome the dependency, a radiance estimation method can be applied to the dehazing process. Ren et al. estimated the haze-free radiance by using a mult-scale convolutional neural network and simulated haze dataset [22]. The mult-scale convolutional neural network can stably remove the simulated haze. Chen et al. estimated a physical haze model-based radiance image using a dilated convolution [18] and adaptive normalization [28]. It can approximate the DCP or non-local dehazing operators using low computational complexity. This radiance estimation method can accurately estimate the dehazed result without additional estimation steps, but it may generate the amplified noise and dehazing artifacts. To approach fusion method, Ren et al. removed the haze using derived inputs and gated fusion network [29], Shin et al. proposed the triple convolutional networks including dehazing, enhancement, and concatenating subnetworks to enhance the contrast without dehazing artifacts [30]. However, the separated subnetworks result in increasing computational complexity. To solve this problem, this paper presents a new dehazing and verifying network (DVNet). The proposed DVNet does not need the subnetworks in the prediction procedure. Instead, only correction subnetwork is used for the training process, and evaluates the dehazing error in the output using a complementary adversarial learning. Different from the transmission estimation-based method, the proposed DVNet successfully removed the real-haze without the noise, halo, or other undesired artifacts with low computational complexity. Since the proposed method can use more enhanced ground truth images, our DVNet can be effectively learned by using absolute-mean error and perceptual loss functions. Furthermore, our verifying network simultaneously estimates and reduces the error of the resulting images via self supervised learning and least square adversarial network. Therefore, experimental results show that the proposed DVNet outperforms existing state-of-the-art approaches in the sense of both robustness to various haze environment and computational efficiency. This paper is organized as follows: Section 2 summarized related works, and Section 3, respectively, describes the proposed DVNet and the corresponding training method. After summarizing experimental results in Section 4, we conclude the paper with some discussions in Section 5.

2. Related Works

A clear image is degraded by the physical haze model as [5]
x C ( p ) = t ( p ) J C ( p ) + 1 t ( p ) A C     for     C { R , G , B } ,
where J represents a haze-free, clean image, x the hazy, degraded version, p the two-dimensional pixel coordinate, t the light transmission map, and A the spatially-invariant atmospheric light. Superscripts in x, J, and A represent a color channel, and the transmission t ( p ) is independent of the color channel. To solve this equation, physical haze model-based methods estimate the major components such as t and A based on a proper assumptions. Recently, several deep learning techniques can make this formula solvable without estimating t or A estimations. Therefore, this section introduces various deep learning-based dehazing approaches.

2.1. Physical Haze Model-Based Dehazing

He et al. applied the dark channel prior (DCP) to estimate the transmission as [14]
t D C P ( p ) = 1 min q N ( p ) min C { R , G , B } x C ( q ) A C ,
where q is the 2D pixel coordinate in a local patch region around p, denoted as N ( p ) , in which the transmission is assumed to be constant. Berman et al. estimated the non-local (NL) transmission map using the geometric haze feature as [15]
t N L ( p ) = x ( p ) A J ( p ) A .
To solve for the feature in (3), Berman et al. used 500 representative colors and approximated the denominator using the k-nearest neighbor (k-NN) algorithm [31]. To minimize the dehazing artifacts such as noise and halo in the estimated transmission, either soft matting or weighted least squares [32,33] algorithm can be used as a regularization function. Shin et al. estimate the transmission by minimizing the radiance-reflectance combined cost as [16]
arg min t R R O , d J d J t R R O d I 2 2 ,
where d I , d J are the difference map between the atmospheric light and prior-images such as input and roughly restored input.

2.2. Radiance-Based Dehazing

Given N pairs of haze-free and its hazy version patches, CNN-based dehazing methods commonly train the network by minimizing the loss function as
L C N N ( Θ ) = 1 N i = 1 N F ( x i P ; Θ ) J i P ,
where J i P and x i P represent the i-th training patches of the haze-free and hazy images, respectively. Θ is a set of network parameters including weights and biases, and F ( · ) is the output of the network given an input hazy image patch and the set of parameters [28,34].

2.3. Adversarial Learning

To reduce the divergence between the generated and real images, the adversarial loss can be defined as [26,35,36,37]
arg min { G J } max { D } L 1 D ( G J ( I ) ) + L D ( J ) ,
where G J is the haze-free generator, D is a discriminator to discriminate a real or fake class, and L { · } denotes a sigmoid cross entropy operator. This adversarial learning can generate a haze-free image that is closer to the clean image.

3. Proposed Method

To remove haze, we present a new dehazing and verifying networks using dilated convolution layers and generative adversarial network. Deep learning-based dehazing methods require a serious of procedures including: Generation of dataset, configuration of a deep learning model, and training the model. In this section, we describe the data generation method in Section 3.1, the network architecture and learning functions of both correction and haze nets are given in Section 3.2 and Section 3.3. Section 3.4 presents the proposed training approaches including the verifying network and complementary adversarial learning.

3.1. Data Generation

To generate the pairs of the haze and clean images, we first generate the initial dehazed image from the input hazy image using a physical haze model given in (1). Let I ( p ) be the input hazy image, and t ^ ( p ) the estimated transmission using either (3) or (4), the initial clean image is computed as
I D ( p ) = A + I i n ( p ) A t ^ ( p ) .
Since (7) gives an one-step, closed-form estimation, the training pairs of the hazy and haze-free images can be easily created. In this paper, we used the result of the non-local dehazing (NL) and radiance-reflectance optimization method (RRO) given in (3) and (4) to generate the initial dehazed images. In addition, haze simulated images such as NYU-depth data [23] can also be used to generate I D and I i n pair based on physical haze-model. Overall, the generated data I D is used to input data of the correction network as shown in Figure 2. In the dehazing procedure, the input haze images are resoted by the haze network, which is learned by the corrected images. The verifying network imitates the natural images using self supervised learning, and the discriminator classifies the real or fake class between the natural image and generated images to reduce the statistical divergence.

3.2. Correction-Network (CNet)

We propose a correction network (CNet) to enhance the initial dehazed images by correcting both color and intensity values. To restore the missing information, we concatenate features of the haze network (HNet) using the dilated convolution and adaptive normalization [18,28] as
f ^ i k = g A k b i k + j f j k 1 * r k h i , j k ,
where f ^ i k and b i k , respectively, represent the i-th feature map and bias in the k-th layer, and h i , j k is   the   kernels   to   obtain   the i- th   feature   map   using   the   feature   maps   extracted   in   the k − 1 st   layer , f k 1 . The   operator “*rk represents   the   dilated   convolution   using   the   rate   of   the k- th   layer , rk. The   dilated   convolution   can   quickly   perform   filtering   in   a   wide receptive field without changing the scale. g is a leaky rectified linear unit (LReLU) [38] function defined as
g x = max x 5 , x .
A k (·) represents the adaptive normalization (AN) function in the k-th layer as
A k x = α k x + β k B N x ,
where B N ( · ) denotes the batch normalization function [39], α k and β k are the trainable parameters to control the relative portion of the batch normalization function. The adaptive normalization approach given in (10) can provide an enhanced restoration results [28]. In (8), f k 1 is concatenated as
f k 1 =   c o n c a t f ^ k 1 , f k 1 ,
where c o n c a t is a feature concatenation operator [40], f is the feature map in a HNet that will be described in Section 3.3. This connection plays an important role in coordinating the learning direction. For example, if the CNet is incorrectly learned without the upward connections, the HNet is also learned with different images and such erroneous cycles are repeated. To correctly propagate the learning direction, we concatenate the feature maps of the HNet to the upward feature maps of the CNet. Top of Figure 2 shows the CNet and the proposed upward connection scheme. In addition, the parameters of CNet can be optimized by self-supervised learning using the perceptual loss [41], and it can be defined by VGG16 network [42] which is pretrained using ImageNet data [43]. The perceptual loss in the CNet is referred to as correction loss, which is defined as
L C = 1 N i = 1 N F ( I D i ) F ( I C i ) 2 2 + I D i I C i 1 + λ I C i 1 ,
where N represents the batch size, I C the output of the C-Net, and F returns the feature maps of the VGG16 network model. We used relu1-2, relu2-2, relu3-3 and relu4-3 features in the VGG16. λ is a parameter to regularize 1 -norm of the gradient. This self-supervised CNet can correct color, intensity, and saturation in real-hazy dataset [27] as shown in Figure 3.

3.3. Haze-Network (HNet)

The HNet plays an important role in enhancing the degraded images. In addition, an efficient design of the H-Net can significantly reduce the processing time. For that reason, the HNet uses the dilated convolution and adaptive normalization [18,28] as,
f i k = g A k b i k + j f j k 1 * r s h i , j k ,
where f i k is a feature map of the H-Net in the k-th layer. b, h, and A k ( · ) , respectively, represent the bias, kernel and adaptive normalization operator. Since the HNet is learned using the results of the CNet, its result can also be corrected in an adaptive manner. The HNet can be optimized by minimizing the haze loss as:
L H = 1 N i = 1 N F I D i F I H i 2 2 + I C i I H i 1 + λ I H i 1 ,
where I H i is the output of the HNet.

3.4. Verifying Network

To make the outputs of the dehazing network (HNet, CNet) look more natural, we verify the errors, such as noise and halo artifact, using self-supervised learning with clean data [44]. The verifying loss of the self-supervised learning is defined as
L S = 1 N i = 1 N 2 F ( I N i ) ( F ( I V i ) + F ( I V ^ i ) ) 2 2 + 2 I N i ( I V i + I V ^ i ) 1 + λ I V i + I V ^ i 1 ,
where I N i , I V ^ i , and I V i , respectively, represent the clean image, results of the CNet, and HNet. Note that the self-supervised terms are designed by considering the errors, which means that the pixels and features in output images of both CNet and HNet are closed to the real natural images when the input images are ideally clean [30]. If input images are the clean images, the ideal haze model should generate the same natural images as in the left-bottom of Figure 2. Therefore this self supervised loss should be separately applied to optimize the networks as Algorithms 1 and 2. In this context, the self-supervised learning based on the loss in (15) using a clean image can minimize the dehazing artifacts as shown in Figure 4d. Futhermore, to reduce the statistical divergence between the generated and real images, the proposed DVNet can be optimized based on the least square adversarial cost [36]
min D V D = E I N P D ( I N ) D I N 1 2 + E I i n P G ( I i n ) D H N e t I i n 2 + E I D P G ( I D ) D C N e t I D 2 + E I N P G ( I N ) D H N e t I N 2 ,
and
min G V G = E I i n P G ( I i n ) D H N e t I i n 1 2 + E I D P G ( I D ) D C N e t I D 1 2 + E I N P G ( I N ) D H N e t I N 1 2 ,
where D is a convolutional neural net based dicriminator as shown in right-bottom of Figure 2, which returns a probablity value of the input image I * using a binary softmax algorithm. G is the generative networks including HNet and CNet. The input data of the discriminator is the ideally natural data I N , and the random noise is replaced to real-haze image I i n , the initial dehazed image I D , and natural image I N to engage our HNet and CNet.
In this adversarial learning method, the proposed network can be learned to reduce the probability divergence between the clean image I N and the result of the proposed network ( I H , I C , I V ) using unfair images. To implement the adversarial cost, we will describe about the optimal parameters in Appendix A.
Therefore, the resulting images ( I H , I C , I V ) can be improved as the visibility is similar to the clean images ( I N ). Figure 4e shows the performance of the proposed DVNet. More specifically, the resulting images in Figure 4 show that our DVNet can better enhance the hazy images [45] in the sense of both details and contrast without the undesired dehazing artifacts.
Algorithm 1: Training procedures of the proposed DVNet
Input: I i n , I D , I N
Output: w d
for iteration from 1 to 15 K do
  1: [features, I H ] = HNet( I i n , w d )
  2: I C = CNet( I D , features, w g )
  3: I V , I V ^ = VNet( I N , w d , w g )
  4: P * = Discriminator( I C , I V , I H , I N , w a d v )
  5: update model by minimizing (14) + (12)
  6: update model by minimizing (15)
  7: update model by minimizing (16)
  8: update model by minimizing (17)
end for
Algorithm 2: Testing procedures of the proposed DVNet
Input: I i n , w d
Output: I H
  1: I H = HNet( I i n , w d )

3.5. Implementation

For the implementation, we split our method into the training and testing procedures. The training procedure consists of eight steps: (i) Feature extraction using HNet, (ii) feature concatenation using the CNet and generation of the corrected clean image, (iii) error verification using the same network architecture and natural image [44], (iv) differentiation of the real and fake images using discriminator, (v) minimizing (14) + (12), (vi) minimizing (15), (vii) maximizing and minimizing adversarial costs V ( D ) and V ( G ) , (viii) repeat the above seven steps until the optimal CNN weights are obtained. The test procedure is simpler than the training procedure, and applies the optimal HNet to remove haze. Table 1 shows the pseudo-code of training and testing procedures of the proposed method. In Table 2 and Table 3, the parameters of the proposed DVNet and discriminator are given for the implementation. To optimize the cost functions, we used an adaptive moment estimation (ADAM) optimization algorithm proposed by [46]. Learning rate values of the DNet and VNet were, respectively, set to 1 × 10 4 and 4 × 10 4 . We used 500 real-haze images from the dataset provided by [27], which are engaged to the DVNet with high quality images from NITRE 2017 dataset [44]. Initial clean images were created using the NL, RRO, and NYU-depth data [15,16,23] using five hundred training images. We trained the proposed DVNet 10,000 times. Table 1 shows conventions for the important variables and parameters for the implementation.

4. Experimental Results

For the experiment, we selected three benchmark datasets of size 512 × 512 including I-Haze, O-Haze, and 100 real hazy images [27,47,48,49]. Especially for the comparative experiment, we tested existing dehazing methods including: Haze-line prior-based nonlocal dehazing method (NL), densely connected pyramid dehazing net (DCPDN), radiance-reflectance optimization based dehazing (RRO), the region-based haze image enhancement method by using triple convolution network(TCN) [15,16,25,30]. Both NL and RRO were implemented in Matlab 2016b and tested on i7 CPU equipped with 64 GB of RAM. On the other hand, DCPDN, TCN and the proposed method were tested using NVIDIA RTX 2080ti graphics processing unit (GPU) and implemented in Python version 3.6 and Tensorflow. This section includes similarity evaluation in Section 4.1, visual quality evaluation in Section 4.2, and ablation study in Section 4.3.

4.1. Similarity Evaluation

For the similarity evaluation, we used three benchmarking datasets including: I-Haze (30), and O-Haze (45) [47,48].
For the quantitative evaluation, we measured the peak signal to noise ratio (PSNR), structural similarity index measure (SSIM), and CIE color difference formula 2000 (CIED) [50,51] as shown in Figure 5 and Figure 6 and Table 4, where the best and second best scores are, respectively, shown with blue and cyan colored text. The proposed DVNet is trained by non-local dehazing or radiacne-reflectance optimization-based restoration results or NYU-depth dataset based haze-clean pair.
Both DVNet-RRO and DVNet-NL outperform than state-of-the-art approaches in term of both SSIM, and CIED in I-Haze dataset, which has the ideal illumination because each image was acquired in the indoor environment. However, the performance of DVNet-NYU was slightly lower than TCN-RRO in term of PSNR and SSIM because simulated dataset can not reflect various environments such as airlight and illuminance. It means that the DVNet-NYU can generate intensity saturation as shown in Figure 5h.
Since adaptive normalization used in the TCN and our DVNet stretches the intensity, both DVNet and TCN can change the background color. Therefore, the PSNR of the DVNet-RRO is similar to that of TCN. Note that the DVNet does not only remove the haze but also change the illumination. So the resulting image has a different illuminance from the ground truth image. For that reason, the DVNets and TCN produce a lower similarity in the O-Haze dataset than the NL and RRO approaches.
However the DVNet-RRO performs better than other CNN-based methods such as DCPDN and TCN in term of SSIM.

4.2. Visual Qaulity Assessment

To verify the performance of the DVNets in the real haze conditions, we used 100-FADE test sets provided by [27]. For the objective evaluation, we select no-reference measures including: Contrast to noise ratio (CNR), natural image quality evaluation (NIQE), entropy to evaluate amount of information in a single image such as intensity distribution, and intensity saturation [27,52,53]. A high-quality image has high CNR and entropy values, whereas it should have a low NIQE and saturation values for stable enhancement. The average scores of the proposed DVNet-NL are higher than those of state-of-the-art approaches in terms of the CNR and saturation as shown in Table 5. The ranking of the DVNet-NYU was the best score in terms of CNR, entropy, and NIQE. However, due to highly saturated pixels, the color of resultant image of DVNet-NYU can be distorted as shown in Figure 7h. Note that the DVNet-NL has high score in terms of the NIQE with a very small difference from the first NL. The DVNet-RRO also has a similar score in term of NIQE compared with RRO. However, the saturation score of the DVNets are lower than NL and RRO because our DVNets verifies the errors of the NL, RRO, and NYU-depth dataset. In summary, the proposed DVNet can successfully remove various types of haze in various environment [27] as shown in Figure 7 and Figure 8.

4.3. Additional Study

To demonstrate the effect the proposed contributions, we conducted the additional studies using the I-Haze and O-Haze datasets. We also used version of the DVNet-NL for the ablation study. In Table 6, HNet and CNet represent the baseline of the proposed dehazing network, DVNet the optimized version of the proposed method with the natural image and self-supervised learning, GAN the optimized version of the proposed method using the proposed adversarial learning method.
Note that the combined HNet and CNet model without VNet returns only similar images to those of physical model-based dehazing method, which also imitates the error such as noise and saturation. Our DNet ( H N e t + C N e t ) can reduce the intensity distortion caused by initial dehazed image I D . The SSIM values the DVNet increased at the cost of a slight PSNR reduction. This means that our verifying process can prevent the noise and halo at the cost of slightly reduced dehazing performance. However, since the proposed adversarial network complements the dehazing performance, the PSNR values outperform the vanilla DVNet. In addition, Table 7 shows the processing time of the proposed DVNet with various image sizes. In evaluation procedure, the proposed DVNets only use a single network(HNet). Therefore, the DVNets can more reduce the computational time over 5–10 times than the TCN and DCPDN, which have several subnetworks.

5. Conclusions

To estimate a high-quality, clean radiance image without the dehazing artifacts, we proposed a novel dehazing network followed by a verifying network, which generates the radiance images to verify the dehazing errors. To estimate an ideally clean image pair, we concatenate feature maps using adaptive normalization and upward connections from the HNet to the CNet. In addition, an unpaired natural image and the discriminator can help minimizing the noise and dehazing artifacts without the performance degradation. The DVNet can be adaptively remove the haze without addtional estimation processes. Therefore, the proposed approach can efficiently remove various types of haze with low conputational complexity. More specifically, three experiments were conducted to verify the performance of the DVNet and the effect of the individual contributions. As a result, the DVNet can provide high-quality dehazing results under various types of haze environments. However, the DVNet may depend on the based training data. In the future work, we plan to combine the DVNet with the data augmentation method, and expand it to video dehazing.

Author Contributions

Conceptualization, J.S. and J.P.; methodology, J.S.; software, J.S. and J.P.; validation, J.S. and J.P.; formal analysis, J.S. and J.P.; investigation, J.S. and J.P.; resources, J.P.; data curation, J.S. and J.P.; writing—original draft preparation, J.S. and J.P.; writing—review and editing, J.P.; visualization, J.S.; supervision, J.S. and J.P.; project administration, J.P.; funding acquisition, J.P. All authors have read and agreed to the published version of the manuscript.

Funding

This work was partly supported by Institute of Information & communications Technology Planning & Evaluation (IITP) grant funded by the Korea government(MSIT) (No. 2021-0-01341), and by the National R&D Program through the National Research Foundation of Korea(NRF) funded by Ministry of Science and ICT(2020M3F6A1110350), and in part by the Institute for Information & Communications Technology Promotion (IITP) through the Korea Government (MSIT)—Intelligent Defense Boundary Surveillance Technology Using Collaborative Reinforced Learning of Embedded Edge Camera and Image Analysis under Grant 2017-0-00250.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Optimal Parameters

In proposed method, the least square adversarial cost functions are defined as [36]
max D V ( D ) = E I N P D ( I N ) ( D ( I N ) b ) 2 + E I i n P G ( I i n ) ( D ( H N e t ( I i n ) ) a ) 2 + E I D P G ( I D ) ( D ( C N e t ( I D ) ) a ) 2 + E I N P G ( I N ) ( D ( H N e t ( I N ) ) a ) 2 ,
and
min G V G = E I i n P G ( I i n ) D H N e t I i n c 2 + E I D P G ( I D ) D C N e t I D c 2 + E I N P G ( I N ) D H N e t I N c 2 .
To expand the adversarial cost, (A2) can be modified as
min G V G = E I N P D ( I N ) D I N c 2 + E I i n P G ( I i n ) D H N e t I i n c 2 + E I D P G ( I D ) D C N e t I C c 2 + E I N P G ( I N ) D H N e t I N c 2 ,
where D returns the probability values via the discriminator using the soft-max algorithm, and G represents the proposed generator model including H and CNet. To find the optimal point of the discriminator, V ( D ) in (A1) can be expressed as
V D = E x P D ( I N ) D x b 2 + E x P G ( I i n ) D x a 2 + E x P G ( I D ) D x a 2 + E x P G ( I N ) D x a 2 = 1 2 x P D x D x b 2 + ( 3 P G x ) D x a 2 .
The optimal point of the discriminator D * can be obtained when its partial derivative with respect to D is equal to zero, such as
D P D x D x b 2 + ( 3 P G ) D x a 2 = 0 .
Therefore, the optimal point D * can be defined as
D * x = b P D ( I N ) + a 3 P G P D ( I N ) + 3 P G ,
which can be simplified by defining the real and fake distributions, respectively, denoted as P 1 = P D and P 2 = 3 P G ,
D * ( x ) = b P 1 + a P 2 P 1 + P 2 ,
(A2) is expressed as
V G = E x P 1 D * x c 2 + E x P 2 D * x c 2 ,
V G = E x P 1 b P 2 x + a P 2 x P 1 x + P 2 x c 2 + E x P 2 b P 1 x + a P 2 x P 1 x + P 2 x c 2 ,
V G = x P 1 x b c P 1 x + a c P 2 x P 1 x + P 2 x 2 d x + x p 2 x b c P 1 x + a c P 2 x P 1 x + P 2 x 2 d x ,
and
V G = x b c p 1 x + p 2 x b a p 2 x 2 p 1 + p 2 d x .
If we set conditions as: b c = 1 , b a = 4 3 , and P 1 1 3 P 2 , then V G will converge. Therefore, (A10) is re-written as
V G = x 4 3 P 2 x P 1 x + P 2 x 2 P 1 + P 2 d x ,
= χ P 2 P 1 + P 2 4 3 P 2 ,
where χ P 2 represents Pearson- χ 2 divergence [36]. It means that when the above conditions are satisfied, χ 2 divergence can minimize the distance between P 1 + P 2 and 4 3 P 2 . So, above equation can be expressed as
V G = χ P 2 P D + 3 P G 4 P G .
If all conditions are satisfied, then P D = P G . Therefore, the optimal parameters can be defined as a = 4 / 3 , b = 0 , and c = 1 . However, since the maximum value of D is equal to 1, the proposed parameters are applied as
max D V D = E I N P D ( I N ) D I N 2 + E I i n P G ( I i n ) D H N e t I i n 1 2 + E I D P G ( I D ) D C N e t I D 1 2 + E I N P G ( I N ) D H N e t I N 1 2 ,
and
min G V G = E I i n P G ( I i n ) D H N e t I i n 1 2 + E I D P G ( I D ) D C N e t I D 1 2 + E I N P G ( I N ) D H N e t I N 1 2 .
To convert the minimizing problem, (A15) can be rewritted as
min D V D = E I N P D ( I N ) D I N 1 2 + E I i n P G ( I i n ) D H N e t I i n 2 + E I D P G ( I D ) D C N e t I D 2 + E I N P G ( I N ) D H N e t I N 2 .

References

  1. Shin, J.; Koo, B.; Kim, Y.; Paik, J. Deep Binary Classification via Multi-Resolution Network and Stochastic Orthogonality for Subcompact Vehicle Recognition. Sensors 2020, 20, 2715. [Google Scholar] [CrossRef]
  2. Kim, Y.; Shin, J.; Park, H.; Paik, J. Real-Time Visual Tracking with Variational Structure Attention Network. Sensors 2019, 19, 4904. [Google Scholar] [CrossRef] [Green Version]
  3. Jeon, J.; Yoon, I.; Kim, D.; Lee, J.; Paik, J. Fully digital auto-focusing system with automatic focusing region selection and point spread function estimation. IEEE Trans. Consum. Electron. 2010, 56, 1204–1210. [Google Scholar] [CrossRef]
  4. Im, J.; Jeon, J.; Hayes, M.H.; Paik, J. Single image-based ghost-free high dynamic range imaging using local histogram stretching and spatially-adaptive denoising. IEEE Trans. Consum. Electron. 2011, 57, 1478–1484. [Google Scholar] [CrossRef]
  5. Middleton, W.E.K. Vision through the Atmosphere; University of Toronto Press: Toronto, ON, Canada, 1952. [Google Scholar]
  6. Schechner, Y.Y.; Narasimhan, S.G.; Nayar, S.K. Instant dehazing of images using polarization. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Kauai, HI, USA, 8–14 December 2001. [Google Scholar]
  7. Narasimhan, S.G.; Nayar, S.K. Contrast restoration of weather degraded images. IEEE Trans. Pattern Anal. Mach. Intell. 2003, 25, 713–724. [Google Scholar] [CrossRef] [Green Version]
  8. Oakley, J.P.; Satherley, B.L. Improving image quality in poor visibility conditions using a physical model for contrast degradation. IEEE Trans. Image Process. 1998, 7, 167–179. [Google Scholar] [CrossRef]
  9. Narasimhan, S.G.; Nayar, S.K. Interactive (de) weathering of an image using physical models. In Proceedings of the IEEE Workshop on Color and Photometric Methods in Computer Vision, Nice, France, 12 October 2003; Volume 6, p. 1. [Google Scholar]
  10. Fattal, R. Single image dehazing. ACM Trans. Graph. 2008, 27, 72. [Google Scholar] [CrossRef]
  11. Jeong, K.; Song, B. Fog Detection and Fog Synthesis for Effective Quantitative Evaluation of Fog–detection-and-removal Algorithms. IEIE Trans. Smart Process. Comput. 2018, 7, 350–360. [Google Scholar] [CrossRef]
  12. Shin, J.; Park, H.; Park, J.; Ha, J.; Paik, J. Variational Low-light Image Enhancement based on a Haze Model. IEIE Trans. Smart Process. Comput. 2018, 7, 325–331. [Google Scholar] [CrossRef]
  13. Ha, E.; Shin, J.; Paik, J. Gated Dehazing Network via Least Square Adversarial Learning. Sensors 2020, 20, 6311. [Google Scholar] [CrossRef] [PubMed]
  14. He, K.; Sun, J.; Tang, X. Single image haze removal using dark channel prior. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 33, 2341–2353. [Google Scholar] [PubMed]
  15. Berman, D.; Avidan, S. Non-local image dehazing. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1674–1682. [Google Scholar]
  16. Shin, J.; Kim, M.; Paik, J.; Lee, S. Radiance–Reflectance Combined Optimization and Structure-Guided 0-Norm for Single Image Dehazing. IEEE Trans. Multimed. 2020, 22, 30–44. [Google Scholar] [CrossRef]
  17. Chen, Y.; Lai, Y.K.; Liu, Y.J. CartoonGAN: Generative Adversarial Networks for Photo Cartoonization. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 9465–9474. [Google Scholar]
  18. Yu, F.; Koltun, V. Multi-scale context aggregation by dilated convolutions. arXiv 2015, arXiv:1511.07122. [Google Scholar]
  19. Shamsolmoali, P.; Zareapoor, M.; Zhang, J.; Yang, J. Image super resolution by dilated dense progressive network. Image Vis. Comput. 2019, 88, 9–18. [Google Scholar] [CrossRef]
  20. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  21. Cai, B.; Xu, X.; Jia, K.; Qing, C.; Tao, D. Dehazenet: An end-to-end system for single image haze removal. IEEE Trans. Image Process. 2016, 25, 5187–5198. [Google Scholar] [CrossRef] [Green Version]
  22. Ren, W.; Liu, S.; Zhang, H.; Pan, J.; Cao, X.; Yang, M.H. Single image dehazing via multi-scale convolutional neural networks. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 8–16 October 2016; pp. 154–169. [Google Scholar]
  23. Silberman, N.; Hoiem, D.; Kohli, P.; Fergus, R. Indoor segmentation and support inference from rgbd images. In Proceedings of the European Conference on Computer Vision, Florence, Italy, 7–13 October 2012; pp. 746–760. [Google Scholar]
  24. Li, B.; Peng, X.; Wang, Z.; Xu, J.; Dan, F. AOD-Net: All-in-One Dehazing Network. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017. [Google Scholar]
  25. Zhang, H.; Patel, V.M. Densely connected pyramid dehazing network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  26. Mirza, M.; Osindero, S. Conditional Generative Advarsarial Nets. arXiv 2014, arXiv:1411.1784. [Google Scholar]
  27. Choi, L.; Yu, J.; Conrad, V.A. Referenceless Prediction of Perceptual Fog Density and Perceptual Image Defogging. IEEE Trans. Image Process. 2015, 24, 3888–3901. [Google Scholar] [CrossRef] [PubMed]
  28. Chen, Q.; Xu, J.; Koltun, V. Fast image processing with fully-convolutional networks. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; Volume 9, pp. 2516–2525. [Google Scholar]
  29. Ren, W.; Ma, L.; Zhang, J.; Pan, J.; Cao, X.; Liu, W.; Yang, M. Gated Fusion Network for Single Image Dehazing. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 3253–3261. [Google Scholar] [CrossRef] [Green Version]
  30. Shin, J.; Park, H.; Paik, J. Region-Based Dehazing via Dual-Supervised Triple-Convolutional Network. IEEE Trans. Multimed. 2021. [Google Scholar] [CrossRef]
  31. Duda, R.O.; Hart, P.E. Pattern classification and scene analysis. In A Wiley-Interscience Publication; Wiley: New York, NY, USA, 1973. [Google Scholar]
  32. Levin, A.; Lischinski, D.; Weiss, Y. A closed form solution to natural image matting. In Proceedings of the 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, New York, NY, USA, 17–22 June 2006; Volume 1, pp. 61–68. [Google Scholar]
  33. Farbman, Z.; Fattal, R.; Lischinski, D.; Szeliski, R. Edge-preserving decompositions for multi-scale tone and detail manipulation. ACM Trans. Graph. 2008, 27, 67. [Google Scholar] [CrossRef]
  34. Liu, Z.; Xiao, B.; Alrabeiah, M.; Wang, K.; Chen, J. Single Image Dehazing with a Generic Model-Agnostic Convolutional Neural Network. IEEE Signal Process. Lett. 2019, 26, 833–837. [Google Scholar] [CrossRef]
  35. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. Adv. Neural Inf. Process. Syst. 2014, 27, 2672–2680. [Google Scholar]
  36. Mao, X.; Li, Q.; Xie, H.; Lau, R.Y.; Wang, Z.; Smolley, S.P. Least squares generative adversarial networks. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2813–2821. [Google Scholar]
  37. Ouyang, Y. Total variation constraint GAN for dynamic scene deblurring. Image Vis. Comput. 2019, 88, 113–119. [Google Scholar] [CrossRef]
  38. Maas, A.L.; Hannun, A.Y.; Ng, A.Y. Rectifier nonlinearities improve neural network acoustic models. In Proceedings of the ICML, Atlanta, GA, USA, 16–21 June 2013; Volume 30, p. 3. [Google Scholar]
  39. Ioffe, S.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. arXiv 2015, arXiv:1502.03167. [Google Scholar]
  40. Huang, G.; Liu, Z.; van der Maaten, L.; Weinberger, K.Q. Densely Connected Convolutional Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 2261–2269. [Google Scholar]
  41. Johnson, J.; Alahi, A.; Fei-Fei, L. Perceptual losses for real-time style transfer and super-resolution. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 8–16 October 2016; pp. 694–711. [Google Scholar]
  42. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  43. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. Imagenet large scale visual recognition challenge. Int. J. Comput. Vis. 2015, 115, 211–252. [Google Scholar] [CrossRef] [Green Version]
  44. Agustsson, E.; Timofte, R. Ntire 2017 challenge on single image super-resolution: Dataset and study. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, Honolulu, HI, USA, 22–25 July 2017; Volume 3, p. 2. [Google Scholar]
  45. Li, B.; Ren, W.; Fu, D.; Tao, D.; Feng, D.; Zeng, W.; Wang, Z. Benchmarking Single-Image Dehazing and Beyond. IEEE Trans. Image Process. 2019, 28, 492–505. [Google Scholar] [CrossRef] [Green Version]
  46. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  47. Codruta O., A.; Cosmin, A.; Radu, T.; Christophe De, V. I-HAZE: A dehazing benchmark with real hazy and haze-free indoor images. arXiv 2018, arXiv:1804.05091v1. [Google Scholar]
  48. Codruta O., A.; Cosmin, A.; Radu, T.; Christophe De, V. O-HAZE: A dehazing benchmark with real hazy and haze-free outdoor images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshop (CVPRW), Salt Lake City, UT, USA, 18–22 June 2018. [Google Scholar]
  49. Li, Y.; You, S.; Brown, M.S.; Tan, R.T. Haze visibility enhancement: A Survey and quantitative benchmarking. Comput. Vis. Image Underst. 2017, 165, 1–16. [Google Scholar] [CrossRef] [Green Version]
  50. Wang, Z.; Conrad, V.A.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  51. Sharma, G.; Wu, W.; Dalal, E.N. The CIEDE2000 Color-Difference Formula: Implementation Notes, Mathematical Observations. Color Res. Appl. 2005, 30, 21–30. [Google Scholar] [CrossRef]
  52. Mittal, A.; Soundararajan, R.; Bovik, A.C. Making a “Completely Blind” Image Quality Analyzer. IEEE Signal Process. Lett. 2013, 20, 209–212. [Google Scholar] [CrossRef]
  53. Hautière, N.; Tarel, J.P.; Aubert, D.; Dumont, E. Blind contrast enhancement assessment by gradient ratioing at visible edges. Image Anal. Stereol. 2011, 27, 87–95. [Google Scholar] [CrossRef]
Figure 1. An analysis of limitation of CNN based methods: (a) One of real-haze images, (b) dehazed result using AODNet, (c) dehazed result using DCPDN, and (d) dehazed result using our DVNet. Note that the proposed method can restore the most naturally looking image by removing real-haze based on the direct estimation of the radiance map.
Figure 1. An analysis of limitation of CNN based methods: (a) One of real-haze images, (b) dehazed result using AODNet, (c) dehazed result using DCPDN, and (d) dehazed result using our DVNet. Note that the proposed method can restore the most naturally looking image by removing real-haze based on the direct estimation of the radiance map.
Sensors 21 06182 g001
Figure 2. The architecture of the proposed DVNet. Note that the blue and black arrows are used for the training the DVNet. In the prediction procedure, only black arrows are efficiently applied.
Figure 2. The architecture of the proposed DVNet. Note that the blue and black arrows are used for the training the DVNet. In the prediction procedure, only black arrows are efficiently applied.
Sensors 21 06182 g002
Figure 3. Performance of the proposed CNet. The proposed CNet can adaptively correct the intensity and contrast via HNet features and adaptive normalization. The HNet uses this corrected data as ground truth for complementary learning.
Figure 3. Performance of the proposed CNet. The proposed CNet can adaptively correct the intensity and contrast via HNet features and adaptive normalization. The HNet uses this corrected data as ground truth for complementary learning.
Sensors 21 06182 g003
Figure 4. Performance of the proposed DVNet using several challenging example of haze images.
Figure 4. Performance of the proposed DVNet using several challenging example of haze images.
Sensors 21 06182 g004
Figure 5. Comparison of dehazed image using I-Haze: (a) Haze input, (b) NL, (c) DCPDN, (d) RRO, (e) TCN, (f) DVNet-NL, (g) DVNet-RRO, (h) DVNet-NYU, and (i) Ground Truth.
Figure 5. Comparison of dehazed image using I-Haze: (a) Haze input, (b) NL, (c) DCPDN, (d) RRO, (e) TCN, (f) DVNet-NL, (g) DVNet-RRO, (h) DVNet-NYU, and (i) Ground Truth.
Sensors 21 06182 g005
Figure 6. Comparison of dehazed image using O-Haze: (a) Haze input, (b) NL, (c) DCPDN, (d) RRO, (e) TCN, (f) DVNet-NL, (g) DVNet-RRO, (h) DVNet-NYU, and (i) Ground Truth.
Figure 6. Comparison of dehazed image using O-Haze: (a) Haze input, (b) NL, (c) DCPDN, (d) RRO, (e) TCN, (f) DVNet-NL, (g) DVNet-RRO, (h) DVNet-NYU, and (i) Ground Truth.
Sensors 21 06182 g006
Figure 7. Dehazing results using real-world dataset: The red desh-shaped box denotes a zoomed region in the red box.
Figure 7. Dehazing results using real-world dataset: The red desh-shaped box denotes a zoomed region in the red box.
Sensors 21 06182 g007
Figure 8. Dehazing results using real-world dataset: The red desh-shaped box denotes a zoomed region in the red box.
Figure 8. Dehazing results using real-world dataset: The red desh-shaped box denotes a zoomed region in the red box.
Sensors 21 06182 g008
Table 1. Conventions of image types.
Table 1. Conventions of image types.
I i n input haze image
I C result of the C-Net
I H generated dehazed image using H-Net
I N natural images for the VNet
I V output of the VNet
I D inintial dehazed image using NL or RRO or NYU
Table 2. Details of the proposed DVNet, where Conv denotes a convolution operator, K3 is kernel size of 3 × 3, R is dilation rate, I input channel, O output channel, AN Adaptive normalization, and lrelu is leaky relu.
Table 2. Details of the proposed DVNet, where Conv denotes a convolution operator, K3 is kernel size of 3 × 3, R is dilation rate, I input channel, O output channel, AN Adaptive normalization, and lrelu is leaky relu.
HNetCNet
Input I i n , I N Input I D , I N
Conv(K3, R1, I3, O24), AN, lreluConv(K3, R1, I3, O24), AN, lrelu
Concat
Conv(K3, R1, I24, O24), AN, lreluConv(K3, R1, I48, O24), AN, lrelu
Concat
Conv(K3, R1, I24, O24), AN, lreluConv(K3, R1, I48, O24), AN, lrelu
Concat
Conv(K3, R2, I24, O24), AN, lreluConv(K3, R2, I48, O24), AN, lrelu
Concat
Conv(K3, R4, I24, O24), AN, lreluConv(K3, R4, I48, O24), AN, lrelu
Concat
Conv(K3, R8, I24, O24), AN, lreluConv(K3, R8, I48, O24), AN, lrelu
Concat
Conv(K3, R16, I24, O24), AN, lreluConv(K3, R16, I48, O24), AN, lrelu
Concat
Conv(K3, R1, I24, O24), AN, lreluConv(K3, R1, I48, O24), AN, lrelu
Concat
Conv(K3, R1, I24, O3)Conv(K3, R1, I48, O3)
Output I H , I V Output I C , I V ^
Table 3. Details of the proposed Discriminator, where Conv denotes a convolution operator, K3 is kernel size of 3 × 3, R is dilation rate, I input channel, O output channel, BN Batch normalization, and lrelu is leaky relu.
Table 3. Details of the proposed Discriminator, where Conv denotes a convolution operator, K3 is kernel size of 3 × 3, R is dilation rate, I input channel, O output channel, BN Batch normalization, and lrelu is leaky relu.
Discriminator
Input I N , I H , I V , I C
Conv (K3, R1, I3, O64), BN, lrelu
Conv (K3, R1, I64, O128), BN, lrelu
Conv (K3, R1, I128, O256), BN, lrelu
Conv (K3, R1, I256, O512), BN, lrelu
FC (I8192, O100), BN, lrelu
FC (I100, O2), Softmax
Table 4. Comparison with state-of-the-art dehazing method using various benchmark dataset, where blue and cyan colored numbers are the best and secondly best scores.
Table 4. Comparison with state-of-the-art dehazing method using various benchmark dataset, where blue and cyan colored numbers are the best and secondly best scores.
-I-HazeO-Haze
Method PSNR SSIM CIED PSNR SSIM CIED
NL [15]16.000.768614.216.760.784216.61
DCPDN [25]14.760.775815.7613.200.744923.79
RRO [16]14.960.766815.5117.230.781316.51
TCN [30]17.150.792114.0415.470.762917.04
DVNet-NL16.760.798513.6215.180.765716.93
DVNet-RRO17.080.801913.6715.210.770717.31
DVNet-NYU16.970.790713.8115.030.756818.16
Table 5. Visual quality evaluation using CNR, Entropy, NIQE, and saturation, where blue and cyan colored numbers are the best and secondly best scores.
Table 5. Visual quality evaluation using CNR, Entropy, NIQE, and saturation, where blue and cyan colored numbers are the best and secondly best scores.
MethodInputNLDCPDNRROTCNDVNet-NLDVNet-RRODVNet-NYU
CNR129.41149.03138.27148.16148.16154.29147.56151.06
Entropy7.026.957.327.167.447.507.507.62
NIQE19.3118.5318.8818.6319.2118.5718.6918.52
Saturation0.798.22%3.66%3.02%1.33%1.29%1.84%2.34%
Table 6. Ablation Study, where bold numbers are best scores.
Table 6. Ablation Study, where bold numbers are best scores.
Ablation StudyI-HazeO-Haze
HNet CNet DVNet GAN PSNR SSIM PSNR SSIM
OXXX15.910.694414.970.6799
OOXX16.380.696415.370.6776
OOOX16.280.790414.500.7519
OOOO16.760.798515.180.7657
Table 7. Processing time (s) according to image size.
Table 7. Processing time (s) according to image size.
Width & Height Size2565127681024
DVNet (gpu)0.0050.0180.0390.065
TCN (gpu)0.010.050.180.74
DCPDN (gpu)-0.05--
RRO (cpu)0.712.424.918.30
NL (cpu)3.133.714.716.80
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Shin, J.; Paik, J. Photo-Realistic Image Dehazing and Verifying Networks via Complementary Adversarial Learning. Sensors 2021, 21, 6182. https://doi.org/10.3390/s21186182

AMA Style

Shin J, Paik J. Photo-Realistic Image Dehazing and Verifying Networks via Complementary Adversarial Learning. Sensors. 2021; 21(18):6182. https://doi.org/10.3390/s21186182

Chicago/Turabian Style

Shin, Joongchol, and Joonki Paik. 2021. "Photo-Realistic Image Dehazing and Verifying Networks via Complementary Adversarial Learning" Sensors 21, no. 18: 6182. https://doi.org/10.3390/s21186182

APA Style

Shin, J., & Paik, J. (2021). Photo-Realistic Image Dehazing and Verifying Networks via Complementary Adversarial Learning. Sensors, 21(18), 6182. https://doi.org/10.3390/s21186182

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop