Next Article in Journal
A Review of Mobile Mapping Systems: From Sensors to Applications
Next Article in Special Issue
Skin Lesion Classification Using Collective Intelligence of Multiple Neural Networks
Previous Article in Journal
Upgrading of the L-P Band Cryogenic Receiver of the Sardinia Radio Telescope: A Feasibility Study
Previous Article in Special Issue
A Novel Method to Inspect 3D Ball Joint Socket Products Using 2D Convolutional Neural Network with Spatial and Channel Attention
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Noise2Kernel: Adaptive Self-Supervised Blind Denoising Using a Dilated Convolutional Kernel Architecture

Department of Computer Science and Engineering, Korea University, Seoul 02841, Korea
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(11), 4255; https://doi.org/10.3390/s22114255
Submission received: 22 April 2022 / Revised: 27 May 2022 / Accepted: 29 May 2022 / Published: 2 June 2022
(This article belongs to the Special Issue Image Processing and Pattern Recognition Based on Deep Learning)

Abstract

:
With the advent of unsupervised learning, efficient training of a deep network for image denoising without pairs of noisy and clean images has become feasible. Most current unsupervised denoising methods are built on self-supervised loss with the assumption of zero-mean noise under the signal-independent condition, which causes brightness-shifting artifacts on unconventional noise statistics (i.e., different from commonly used noise models). Moreover, most blind denoising methods require a random masking scheme for training to ensure the invariance of the denoising process. In this study, we propose a dilated convolutional network that satisfies an invariant property, allowing efficient kernel-based training without random masking. We also propose an adaptive self-supervision loss to increase the tolerance for unconventional noise, which is specifically effective in removing salt-and-pepper or hybrid noise where prior knowledge of noise statistics is not readily available. We demonstrate the efficacy of the proposed method by comparing it with state-of-the-art denoising methods using various examples.

1. Introduction

Denoising is one of the actively studied topics in computer vision and image processing. Images generated from various devices are prone to noise and corruption due to limited imaging environments (e.g., low light, slow shutter speed, etc.). Conventional denoising methods usually rely on known noise models based on specific noise distributions. For instance, image prior-based approaches, such as self-similarity [1,2,3,4,5], require a specific property of pre-defined noise statistics or prior knowledge of a target image. However, there exist many real examples that pre-defined noise statistics do not fit, such as the coherent random noise observed in the transparent films used in electron microscopy (EM) imaging [6]. In such cases, conventional denoising methods may not work well.
In recent years, the supervised learning of convolutional neural networks (CNNs) which have been widely used in many tasks [7,8,9] using clean-noisy image pairs has achieved superior denoising performance [10,11]. Due to the difficulty in obtaining clean–noisy image pairs in real examples, Lehtinen et al. [12], in their seminal study, introduced unsupervised learning of a denoiser (Noise2Noise (N2N)) using only noisy images. Even though N2N proposed the general approach, it still suffers from the acquisition of noisy-noisy image pairs under known noise statistics.
More recently, several new unsupervised image denoising methods [13,14,15,16] have shown promising results with denoisers that can be trained in a self-supervised fashion. For example, Noise2Void (N2V) [14] and Noise2Self (N2S) [15] require only the assumption of zero-mean noise without prior knowledge of noise statistics. These methods performed the denoising task successfully using only noisy images under a zero-mean noise condition. Self2Self (S2S) [16] proposed a novel framework with dropout using Bernoulli-sampled instances of a single input image. Moreover, using only a single training image, S2S outperformed on several famous noise distributions. Despite their potential, these approaches have several drawbacks. First, these self-supervised methods approximate the optimal denoiser with a noisy distribution based on the blind-spot scheme (i.e., random masking of pixels during training). The blind-spot scheme damages the original noisy image, and the large masking rate leads to poor performance. Second is the weakness of the general self-supervision loss function. Because general self-supervision training depends on only the noisy signal, excessive noise causes CNNs to learn poorly and incorrectly. We discovered that state-of-the-art blind denoising methods are prone to predicting the wrong brightness level or shape if noisy images are corrupted by impulse noise (e.g., salt-and-pepper noise) or unconventional noise (e.g., fusion noise). Even though S2S successfully removed the pepper noise, it also predicted different brightness for salt-and-pepper noise, as shown in Figure 1. As a result, for state-of-the-art blind denoising methods, the brightness-shifting artifact always appears in the case of corrupted salt-and-pepper noise.
To address the above issues, we introduce a novel unsupervised-learning-based denoising network. The combination of dilated convolution layers and donut-shaped kernels means it can build a specific function that satisfies the J - i n v a r i a n t property [15]. In addition to the novel network architecture, we propose a novel adaptive self-supervision (ADSS) loss to restore the clean signal on the highly corrupted noisy image without brightness shifting.
The main contributions of our study are as follows:
  • We propose a dilated convolutional invariant network using a donut-shaped kernel and dilated convolutional layers. We no longer need a special training scheme (e.g., random masking) for blind denoising with self-supervision loss.
  • We propose an adaptive self-supervision loss, which is the pixel-level nonlinear energy, to suppress incorrect learning from unconventional noise. We demonstrate that the proposed adaptive loss is highly effective on corrupted noisy images (for example, images with speckle noise, salt-and-pepper noise, and fusion noise) without any prior knowledge of the noise model.
  • We demonstrate that the total variation regularization term can help to restore the pixel-wise artifact, which is a drawback of the proposed method.
To the best of our knowledge, the proposed method is the first fully blind denoising method that can prevent brightness shifting for images highly corrupted by speckle noise, salt-and-pepper noise, and fusion noise without noise statistics and clean–noisy pairs.

2. Related Work

2.1. Conventional Denoising Methods

Total variation (TV), known as TV regularization, is a widely used denoising technique [17,18,19,20] that adopts prior sparsity gradients in image denoising. Filtering methods [1,2,21] based on spatial information or nonlocal self-similarity achieve better performance than TV-based methods. The block-matching and 3D filtering (BM3D) algorithm [2] still performs well enough to be used as a comparison for deep learning. The structure of BM3D is actively applied to various noise types, including salt-and-pepper noise and speckle noise [22,23]. With efficient training data, learning-based methods eventually perform better than non-learning-based ones. Before deep learning, which involves the use of large training data, dictionary learning (DL) and convolutional sparse coding (CSC) [24] were used to restore the original signal using a sparse representation prior with a learned dictionary [25,26,27,28].

2.2. Non-Blind Denoising Methods

In recent years, with advances in deep learning and the related equipment, supervised deep learning over CNNs [10,29,30] has shown great promise with its denoising performance. However, it is not suitable to apply this method in practice because most supervised learning methods require noise statistics to generate the training data on a clean dataset.
Recently, Noise2Noise (N2N) [12] proved that training a deep learning model is feasible and that the expected value of noisy inputs could be equal to the clean target. However, in situations where the noise statistics are unknown, N2N is impractical because of the difficulty in collecting a noisy pair for the same target. With only noise statistics, these works [31,32] perform as well as or slightly better than supervised learning. For instance, Ref. [31] suggest concrete self-supervision losses suitable for each noise statistic, but it is difficult to apply the proposed loss in cases with unknown noise statistics. Ref. [31] also presented a new blind-spot network that contributes similar to the architecture presented in the present study. However, we take a different approach to enabling self-supervision learning using the J - i n v a r i a n t property. Similarly, Noisy-As-Clean (NAC) [32] suggested a training scheme with pairs of noisy images x and x + n s where n s is a simulated noise. The researchers demonstrated that loss function L ( f ( x + n s ) , x ) can be embedded into supervised learning. Noisier2Noise [33] presented a novel training approach with only a single noisy realization and noise statistics. It also overcomes the drawback of N2N that is the requirement of a prior of noise distribution. Moreover, the Noisier2Noise approach is applicable to spatially structured noise, one of the main disadvantages of a blind denoising method.

2.3. Blind Denoising Methods

Blind denoising approaches assume that the prior noise distribution is unknown. To restore the clean signal without noise statistics, for instance, deep image prior (DIP) [13] tries to use a handcrafted prior to the image processing tasks. In other words, DIP shows that image prior can be learned by a random-initialized neural network without a specific condition. However, the internal-image-prior-based approach has the two drawbacks of excessive testing time and inadequate performance.
The external-image-prior-based approaches, such as N2V [14] and N2S [15], employ the blind-spot scheme to prevent being an identity mapping function by self-supervision loss. Furthermore, two state-of-the-art methods take the self-supervision perspective to train the deep learning model using only noisy images. Two methods achieved significant shortening of testing time through the external image prior. In addition, N2S [15] suggested the J - i n v a r i a n t property to prove that self-supervision loss can substitute for general loss of supervised learning.
Recently, S2S [16] proposed a novel framework based on Bernoulli dropout, a new masking scheme in the training step, to avoid increasing the variance based on internal image prior because a single training sample causes large variances for denoising models such as a Bayes estimator. Even though only a single noisy image is a training sample, S2S outperforms existing blind denoising methods based on the external image prior.

3. Method

In this section, we introduce a novel deep neural network architecture that satisfies the J - i n v a r i a n t property for blind noise reduction using adaptive self-supervision learning (Figure 2). First, we reiterate the definition of J - i n v a r i a n t , originally introduced in N2S [15]. Next, we demonstrate that the proposed network satisfies the J - i n v a r i a n t , which allows self-supervised training without using a specific training scheme (e.g., random masking). Finally, we suggest the adaptive self-supervision loss to overcome the drawback of the conventional self-supervision loss.

3.1. Formulations

This section introduces the formal definition and proposition of J - i n v a r i a n t that is required to explain the proposed network (more details can be found in N2S).
Definition 1.
Let J be a partition of the dimensions { 1 , , m } . Let x be an observed noisy signal, and x J be a sub-sample of x restricted to J J . A function f : R m R m is J- i n v a r i a n t if the value of f ( x ) J does not depend on the value of x J ; f is J - i n v a r i a n t if it is J- i n v a r i a n t for each J J .
We employ self-supervision loss as follows to restore the noisy image using the J - i n v a r i a n t function f.
L ( f , x ) = | | f ( x ) x | | 2
To demonstrate that self-supervision loss can take the place of supervised loss, we borrow the same proposition from N2S under the J - i n v a r i a n t definition.
Proposition 1.
Let us assume that observed image x is an unbiased estimator of y. Let f be the J - i n v a r i a n t function. Then
E | | f ( x ) x | | 2 = E | | f ( x ) y | | 2 + E | | x y | | 2
Proof. 
Let us consider the self-supervision loss over f function.
E x | | f ( x ) x | | 2 = E x , y | | f ( x ) y ( x y ) | | 2 = E x , y | | f ( x ) y | | 2 + | | x y | | 2 2 f ( x ) y , x y
The inner product term f ( x ) y , x y can be considered as follows:
Σ i E y ( E x | y [ ( f ( x ) i y i ) ( x i y i ) ] )
Because f ( x ) i | y and x i | y are independent due to the invariant property of f, Equation (4) becomes Σ i E y ( E x | y [ f ( x ) i y i ] ) ( E x | y [ x i y i ] ) . Then, the third term of Equation (3) vanishes since E x | y [ x i y i ] is zero due to the zero-mean assumption of noise. □
From this, we can infer that the general self-supervised loss would be the sum of the general supervised loss and the variance of noise. Therefore, based on the similar scheme of N2S, we can conclude that an invariant function f can be a general denoiser if f is minimized using a self-supervision loss. In the following section, we introduce the proposed network, which is an J - i n v a r i a n t function using a donut-shaped kernel-based convolution layer and dilated convolutional layers.

3.2. Dilated Convolutional J - I n v a r i a n t Network

Assume that the function f is a CNN with a single donut-shaped kernel (center weight is always zero) (see Figure 3). Based on Definition 1, function f satisfies the J - i n v a r i a n t property because x i is the sum of the multiplication of the neighboring information with the donut kernel, except x J , for all J J where the size of the squared donut kernel K is always an odd number. We focus on this J - i n v a r i a n t function in a fully convolutional network (FCN). If only one general convolution layer is added, the invariant property is not satisfied even though the first layer may use the donut kernel. Furthermore, the receptive field of a single layer is too small to predict the correct pixel within the kernel.
Let f be a network as a function that consists of d-dilated convolution f ( k ) ([34]) for all k [ 1 , n ] where the size of the kernel is 3 × 3 . We infer the function f as f ( x ) = f ( n ) ( f ( n 1 ) ( f ( 1 ) ( f ( 0 ) ( x ) ) ) ) where f ( 0 ) and x are a convolution layer of the K × K donut-shaped kernel and an input noisy image, respectively, and y ( k ) is the output features for each k-th convolution layer. We then need to demonstrate that f ( x ) J does not depend on x J for all J J .
Proposition 2.
The proposed network f is J - i n v a r i a n t if d [ K / 2 ] .
Proof. 
Without loss of generality, we consider a one-dimensional case to prove this proposition. Let us choose one pixel x J where J J . Because of the donut convolution layer f ( 0 ) , x J information moves to the neighboring region { J [ K / 2 ] , , J 1 , J + 1 , , J + [ K / 2 ] } as shown in Figure 3. Let us suppose that the receptive field of x J in y ( k ) is R F ( y ( k ) , x J ) . Then,
R F ( y ( k ) , x J ) = j { d , 0 , d } { i + j | i R F ( y ( k 1 ) , x J ) }
where R F ( y ( 0 ) , x J ) = { J [ K / 2 ] , , J 1 , J + 1 , , J + [ K / 2 ] } for all k [ 1 , n ] . By this recurrence relationship, we can infer that [ K / 2 ] d < 0 and [ K / 2 ] + d > 0 lead to exclude an element J in R F ( y ( n ) , x J ) . In other words, the f ( x ) J never consists of the information of x J if d [ K / 2 ] . □
The combined structure of the donut convolution layer and dilated convolution layer, as shown in Figure 2, always guarantees the existence of the J - i n v a r i a n t property if d [ K / 2 ] and the size of the square kernel of donut convolution layer K is an odd number. In addition, as shown in Figure 2, there are two paths that both consist of 2-dilated or 3-dilated convolutional layers only. Because each path satisfies the J - i n v a r i a n t condition, the proposed network is J - i n v a r i a n t . According to Equation (5), the combination of two receptive fields of 2-dilation and 3-dilation paths can supplement each other the missing regions (i.e., blue pixels in Figure 3) of a 2-dilated convolutional architecture while increasing the field-of-view. To preserve the first prediction computed by a donut kernel, we added a skip connection after the dilated convolution operation. We discovered that the skip improved the convergence speed and image quality. In addition to its model architecture, another important benefit of the proposed method is that it no longer requires the masking scheme. A masked input x ˜ of the noisy image x with the dimension J J (chosen randomly) is defined as
x ˜ = 0 for j J x j for j J
Then, the general self-supervision loss with the masking scheme is defined as follows:
min θ i N J J | | ( f ( x ˜ i ; θ ) J x J i ) | | 2
Because random pixel discarding in the masking scheme introduces defects in image ( x ˜ ), N2V and N2S fill in these missing pixels by copying from random locations or through interpolation from neighboring pixels. Unlike such existing methods, the dilation convolution architecture of the proposed network can be trained using only the original x and the general self-supervision loss without a masking scheme shown below:
L ( f , x ) = | | f ( x ) x | | 2

3.3. Adaptive Self-Supervision Loss

In the unsupervised denoising problem, the zero-mean noise is considered a default noise model. However, the zero-mean condition is too strict to be used on blind denoising with self-supervision loss. For example, in the case of salt noise (i.e., random white dots), the general self-supervision loss may falsely treat the correct prediction as a noisy label due to large differences between the predicted and noisy pixel values, which causes brightness shifting toward white. This implies that self-supervision may fail to work on highly corrupted impulse noise. An additional constraint is required to avoid the convergence to the biased estimator. To address such limitations of standard self-supervision loss, we propose ADSS loss using the focusing parameter λ as follows:
L a d a p ( f , x ) = E j [ w j ( f ( x ) j x j ) 2 ]
w j = 1 1 + λ | f ( x ) j x j |
where x j is a pixel indicated by an index j. The ADSS loss adjusts the proportion of difference between x and f ( x ) adaptively. The main idea behind ADSS is that, if the prediction is significantly different from the input pixel value, it is highly likely that the input pixel is noise. Therefore, during the training process, backpropagation from such pixels should be suppressed (i.e., the correct predictions should not be shifted toward the noise pixel values) by adaptive control of the weight in the loss function. The ADSS loss is equivalent to the self-supervision loss when λ = 0 . Intuitively, λ controls the extent of the influence of discrimination. We expect the ADSS loss can avoid unnecessary learning from unpredictable noise.

4. Results

To assess the performance of the proposed method, we tested it on various noise models, such as those with additive white Gaussian noise (AWGN), speckle noise, and salt-and-pepper noise. In particular, because we focus on highly corrupted noisy images by unconventional noise in the blind aspect, the noise should be modeled by unknown distribution. To simulate this, we built a fusion noise model by mixing AWGN, speckle noise, and salt-and-pepper noise. We compared our method with several state-of-the-art blind denoising methods (N2V, N2S, and S2S). In addition, we also compared our proposed method with conventional denoising methods, such as BM3D [2], SAR-BM3D [23], and AT2F [35], known for the best performing filter-based denoising method specifically designed for each noise model. We implemented Noise2Clean (N2C) on using the same network structure as shown in Figure 2, with a regular 3 × 3 convolution kernel for supervised training using the clean–noisy pairs introduced in Section 4.2. Please note that N2C is a supervised learning method, which serves as the upper bound for the performance of the learning-based denoising method.
For all training (except for N2C), we used only noisy images corrupted by simulated noise. We chose the same dataset, BSD400, of gray scale images used in [10,14] as a training dataset. For more detail, we applied augmentation using rotation and mirroring for all learning-based methods. For testing the performance, we employed the BSD68 and Set14 datasets. In particular, S2S experiments on BSD68 were excluded because S2S is inner-prior-based denoising approach that causes the large computational cost. We used the BSD68 dataset for the ablation study as a validation set.
We used TensorFlow [36] (version 2.0.0) to implement the proposed architecture, as shown in Figure 2. For stable performance, we applied an exponential learning rate decay with an RAdam [37] optimizer. We used batch size 64 and 0.03 as the initial learning rate and λ = 10 for Equation (9). For a fair comparison, we used the default parameter settings from the authors’ code for other blind denoising methods. We picked the best hyperparameters for experimental comparison methods when the setting of a hyperparameter was required. Because the denoiser should satisfy rotation invariance, we rotated each test image by 90 degrees and made two mirrored versions. The average of the inverse of eight outputs was the final prediction.
To evaluate the image quality, we employ two full-reference image quality assessment (FR-IQA) metrics such as peak signal-to-noise ratio (PSNR) and structural similarity index metric (SSIM) defined as follows:
P S N R ( y ^ , y ) = 20 l o g 10 ( M A X y M S E ( y ^ , y ) ) ,
where M A X x is the maximum possible pixel value of the image, MSE is the mean squared error. y and y ^ are the ground-truth and restored image, respectively. SSIM is defined as follows:
S S I M ( y ^ , y ) = ( 2 μ y μ y ^ + c 1 ) ( 2 σ y , y ^ + c 2 ) ( μ y 2 + μ y ^ 2 + c 1 ) ( σ y 2 + σ y ^ 2 + c 2 ) ,
where μ y , μ y ^ are the average of each image, σ y , y ^ is the covariance of y and y ^ , and σ y , σ y ^ are the variance of each image. For the constants c 1 = ( k 1 L ) 2 and c 2 = ( k 2 L ) 2 , we set to k 1 = 0.01 and k 2 = 0.03 . L is the dynamic range of pixel values.

4.1. Denoising Results on Known Noise Models

4.1.1. Additive White Gaussian Noise (AWGN)

AWGN is a popular statistical noise model with a zero-mean characteristic as follows:
y = x + n , n N ( 0 , σ g 2 )
where N is a normal distribution with standard deviation σ g . For the baseline performance, we chose BM3D, which is known for being the best performing method for this noise model. For a fair comparison, we used the standard deviation σ g (i.e., noise level) of the given noise-corrupted images only for the case of BM3D (without the noise level prior, BM3D does not produce correct results). Figure 4 shows the quantitative performance comparison of denoisers over various noise levels, σ g from 20 to 70. The proposed method achieves similar or better performance than N2V and N2S, whereas S2S and BM3D outperformed the proposed method. We expect that the proposed method has no significant performance improvement compared with blind denoising methods on additive white Gaussian noise distribution that satisfy the zero-mean condition. Therefore, we conclude that under the zero-mean noise constraint, our method is comparable to most of the blind denoising methods except S2S.

4.1.2. Speckle Noise

Signal-dependent multiplicative speckle noise, often observed in synthetic aperture radar and ultrasound images, can be modeled as follows:
y = x + n * x , n U ( 0 , σ s 2 )
where U is the uniform distribution with a zero mean and a standard deviation of σ s . We chose SAR-BM3D [23], one of the conventional denoising methods specifically designed for speckle noise, as the baseline method to compare with our proposed method. We conducted the denoising experiment over various noise levels σ s from 5 to 50. Interestingly, blind denoising methods outperformed SAR-BM3D, as shown in the second column of Figure 4. Please note that the proposed method consistently outperformed the other blind denoising methods for all noise levels σ s we tested (see the middle graph in Figure 4). Furthermore, the performance gap between blind denoising methods and SAR-BM3D increases as the noise level increases, which implies that blind denoising methods are more robust to strong speckle noise than SAR-BM3D. The proposed approach achieved the best difference compared to other blind denoisers (by around 4.55 dB higher) on the foreman image in the Set14 dataset (the second row of Figure 5). Moreover, the overall intensity distribution in the predicted image of the proposed method is closer to that of the ground truth; those of other blind denoisers (N2V, N2S, and S2S) suffer from brightness shifting due to the non-zero-mean noise characteristic.

4.1.3. Salt-and-Pepper Noise

In this experiment, we employed salt-and-pepper noise, defined as follows:
y = f s p n ( x , d )
where f s p n is the projection function set to 0 or 1 with probability d. Conventional nonlinear denoising methods for salt-and-pepper noise, such as median filter or AT2F, work well on this noise model. We conducted the experiment using various noise levels from 5% to 50%. For the salt-and-pepper noise, our proposed method performed better than state-of-the-art methods because of its ability to overcome the problem of brightness shifting, as shown in Figure 5 (third row). Please note that other blind denoising methods (N2V, N2S, and S2S) performed poorly on this noise model. Furthermore, our proposed method outperformed AT2F when d 15 on Set14, as shown in Figure 5. Similar to speckle noise, blind denoising methods (i.e., N2V, N2S, and S2S) failed to restore the image contrast of the clean image but our method successfully preserved the contrast and brightness of the original image. Moreover, the proposed method shows better performance as the noise level increases. Note also that the results of AT2F look much blurrier than those of the proposed method.

4.2. Denoising Results on Fusion Noise (Unknown Noise Statistics)

In this section, we compare the performance of denoising methods when the prior knowledge of noise statistics is not available. For this, we generated the fusion noise, which is a mixture of different noise models. We combined three known noise models, AWGN, speckle noise, and salt-and-pepper noise, with σ g , σ s , and d to simulate this fusion noise, which is formally defined as follows:
y = f s p n ( ( x + n g ) + n s * ( x + n g ) , d )
n g N ( 0 , σ g 2 ) , n s U ( 0 , σ s 2 ) .
To compare the results on various noise levels, we selected σ g { 25 , 50 } , σ s { 5 , 25 } , and d { 5 , 25 } .
We compare our proposed method with three well-known blind denoisers (N2V, N2S, and S2S), along with N2C (supervised denoiser) as a baseline. For highly corrupted images such as fusion noise, the image prior knowledge related to gradient can improve the denoising performance to restore structured artifact that could be restored by total variation minimization. The structured artifact was also reported in N2V, and remains as a limitation of the proposed method in this paper. Hence, to resolve this limitation, we employed a TV regularization term as shown below:
L ( f , x ) = E j [ w j ( f ( x ) j x j ) 2 ] + α | | f ( x ) | | T V
We empirically found the value ( 1 × 10 7 ) of alpha that the scale of total variation can reach similar to the scale of adaptive loss. As shown in Figure 6, all other blind denoising methods inaccurately reconstructed the black color to brighter gray color. N2S and S2S also suffered from structural artifacts as well as incorrect brightness (Figure 6 last row). We observed that our model predicted the clean image more accurately while preserving the image contrast and details well as compared with N2S, N2V, and S2S. Furthermore, the TV added version, called N2K+TV, effectively removed noise while preserving sharp edges.
Table 1 and Table 2 summarize the results for various noise levels and denoising methods; our proposed method with total variation (N2K+TV) achieved the highest PSNR compared with the state-of-the-art blind denoising methods. It is clearly shown that the TV regularization effectively improves SSIM, especially for the higher noise levels. We also observed that the performance gap between our method and the others becomes larger as the noise level increases. In summary, we conclude that the proposed method with total variation regularization overcomes the problems caused by the fusion noise that affects most other denoising methods.

4.3. Ablation Study

In this section, we empirically show the difference in the performance of ADSS loss against the general self-supervision loss. In this experiment, we used the same network structure for all test cases; however, the network was trained using different loss functions to see how they affected the performance. The baseline model was trained using the general self-supervision L 2 loss Equation (1).
Table 3 shows the results of the previously introduced models when tested on the BSD68 dataset. It can be seen that the ADSS loss, which suppresses training from outliers, outperforms the general self-supervision loss at various levels of fusion noise except the case of σ g = 50, σ s = 5, and d = 5. In addition, ADSS+TV achieved higher PSNR and SSIM than the baseline and ADSS alone. As the general self-supervision loss considers all pixels to be training data, it is more sensitive to highly corrupted noisy pixels. The study result also confirms that the performance gap between baseline and ADSS is bigger for higher noise levels ( d = 25 ). As shown in the unknown noise statistics experiment, we observed that the TV loss helped to increase PSNR and SSIM in highly corrupted images. In this ablation study, we demonstrated that the ADSS loss outperformed the general self-supervision loss. We also observed that the TV regularization was highly effective at further improving the image quality. Additionally, TV can reduce the artifact from structured noise, which is the weakness of the proposed method.

4.4. Analysis for ADSS

The basic concept of the ADSS loss is to reduce the unnecessary training from noisy pixels which have a large mean squared error. We now show simple examples highly related to decrease of redundant training. The following examples provide more insights into the properties of the proposed ADSS loss.
First, we should assume that the large gap between x j and f ( x ) j for an arbitrary given j implies learning from x j disturb the denoising performance of the function f (i.e., f is close to an ideal denoiser). Then, we simply reduce the loss forcibly by clipping of the loss as
L c l i p ( f , x ) = E j [ c l i p ( ( f ( x ) j x j ) 2 , 0 , ε ) ] , ε > 0 ,
over the c l i p function that limits the loss exceeds a threshold ε . We conducted an additional experiment to analyze ADSS loss indirectly through the clipping method under the same conditions of salt-and-pepper noise experiments as Section 4.1.3. On the top row of Figure 7, trained denoisers by adjusted loss with ε { 0.2 , 0.4 , 0.6 } successfully recovered white background, unlike the results of state-of-the-art blind denoising methods reported in Section 4.1.3. Additionally, the loss clipping led to better PSNR compared with state-of-the-art blind denoising approaches, as shown in Figure 8. We discovered that the reduction of the loss from the pixels with a large gap between x j and f ( x ) j boosts the performance on salt-and-pepper noise. Unfortunately, the clipping method requires proper value of ε to obtain correct restoration result. Moreover, ε should be found for each pixel locally instead of global thresholding. On the bottom row of Figure 7, we discovered an over emphasized contrast on the pepper surface for significant loss clipping (where ε = 0.2 and 0.4 ). In summary, we empirically showed that the reduction of the loss can prevent unnecessary learning from unpredictable noise (e.g., salt-and-pepper noise and fusion noise) in the general self-supervision loss. Moreover, the ADSS loss can successfully adjust the size of self-supervision loss adaptively and automatically for each pixel.

5. Conclusions

We introduced a novel unsupervised denoising method based on the dilated convolutional J - i n v a r i a n t network, allowing for efficient kernel-based training without the masking scheme. The absence of preprocessing further pushes the performance in terms of training efficiency. We also proposed an adaptive self-supervision loss that is highly effective in preserving overall brightness and structures in the image, even with the extremely high noise level and even if the zero-mean assumption and prior knowledge of noise statistics are not present. Using simulations of known and unknown noise statistics, we showed that the proposed method leads to better denoising quality than other state-of-the-art methods of blind denoising.
We believe the proposed work will be useful in improving highly corrupted noisy images where noise statistics are not readily available. Extending the proposed architecture to general image enhancement problems, such as blind image super-resolution, is another interesting future work. As a limitation, we found no significant performance improvement on AWGN experiment and low-level noisy images such as DND [38] dataset. In the future, we plan to develop more improved ADSS loss, such as the exponential form to outperform the state-of-the-art blind denoising methods on AWGN and real noisy dataset. Furthermore, we plan to explore applications of our method, especially in the biomedical imaging domain.

Author Contributions

Conceptualization, K.L. and W.-K.J.; methodology, K.L.; software, K.L.; validation, K.L. and W.-K.J.; formal analysis, K.L. and W.-K.J.; data curation, K.L.; investigation, K.L.; resources, W.-K.J.; writing—original draft preparation, K.L.; writing—review and editing, K.L. and W.-K.J.; visualization, K.L.; supervision, W.-K.J.; project administration, W.-K.J. All authors have read and agreed to the published version of the manuscript.

Funding

This work was partially supported by the Bio & Medical Technology Development Program of the National Research Foundation of Korea (NRF) funded by the Ministry of Science and ICT (MSIT) (NRF-2019M3E5D2A01063819), the Basic Science Research Program through the NRF funded by the Ministry of Education (NRF-2021R1A6A1A13044830), a grant from the Korea Health Technology R&D Project through the Korea Health Industry Development Institute (KHIDI) funded by the Ministry of Health & Welfare (HI18C0316), the ICT Creative Consilience program (IITP-2022-2020-0-01819) of the Institute for Information & communications Technology Planning & Evaluation (IITP) funded by MSIT, the Korea Institute of Science and Technology (KIST) Institutional Program, Republic of Korea (2E31511), and a Korea University Grant.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data are from the public datasets. BSD400 and BSD68 are available in https://www2.eecs.berkeley.edu/Research/Projects/CS/vision/bsds/ (accessed on 26 May 2022). Set14 is available in https://github.com/jbhuang0604/SelfExSR (accessed on 26 May 2022).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Buades, A.; Coll, B.; Morel, J.M. A non-local algorithm for image denoising. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), San Diego, CA, USA, 20–25 June 2005; Volume 2, pp. 60–65. [Google Scholar]
  2. Dabov, K.; Foi, A.; Katkovnik, V.; Egiazarian, K. Image denoising by sparse 3-D transform-domain collaborative filtering. IEEE Trans. Image Process. 2007, 16, 2080–2095. [Google Scholar] [CrossRef] [PubMed]
  3. Mairal, J.; Bach, F.; Ponce, J.; Sapiro, G.; Zisserman, A. Non-local sparse models for image restoration. In Proceedings of the IEEE 12th International Conference on Computer Vision, Kyoto, Japan, 29 September–2 October 2009; pp. 2272–2279. [Google Scholar]
  4. Dong, W.; Zhang, L.; Shi, G.; Li, X. Nonlocally centralized sparse representation for image restoration. IEEE Trans. Image Process. 2013, 22, 1620–1630. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Gu, S.; Zhang, L.; Zuo, W.; Feng, X. Weighted nuclear norm minimization with application to image denoising. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 2862–2869. [Google Scholar]
  6. Minh Quan, T.; Grant Colburn Hildebrand, D.; Lee, K.; Thomas, L.A.; Kuan, A.T.; Allen Lee, W.C.; Jeong, W.K. Removing Imaging Artifacts in Electron Microscopy using an Asymmetrically Cyclic Adversarial Network without Paired Training Data. In Proceedings of the IEEE/CVF International Conference on Computer Vision Workshop, Seoul, Korea, 27–28 October 2019. [Google Scholar]
  7. Varga, D. No-reference video quality assessment using multi-pooled, saliency weighted deep features and decision fusion. Sensors 2022, 22, 2209. [Google Scholar] [CrossRef] [PubMed]
  8. Yamashita, R.; Nishio, M.; Do, R.K.G.; Togashi, K. Convolutional neural networks: An overview and application in radiology. Insights Imaging 2018, 9, 611–629. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  9. Kattenborn, T.; Leitloff, J.; Schiefer, F.; Hinz, S. Review on Convolutional Neural Networks (CNN) in vegetation remote sensing. ISPRS J. Photogramm. Remote Sens. 2021, 173, 24–49. [Google Scholar] [CrossRef]
  10. Zhang, K.; Zuo, W.; Chen, Y.; Meng, D.; Zhang, L. Beyond a gaussian denoiser: Residual learning of deep cnn for image denoising. IEEE Trans. Image Process. 2017, 26, 3142–3155. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  11. Lefkimmiatis, S. Universal denoising networks: A novel CNN architecture for image denoising. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3204–3213. [Google Scholar]
  12. Lehtinen, J.; Munkberg, J.; Hasselgren, J.; Laine, S.; Karras, T.; Aittala, M.; Aila, T. Noise2Noise: Learning Image Restoration without Clean Data. In Proceedings of the International Conference on International Conference on Machine Learning, Stockholm, Sweden, 10–15 July 2018; Dy, J., Krause, A., Eds.; Volume 80, pp. 2965–2974. [Google Scholar]
  13. Ulyanov, D.; Vedaldi, A.; Lempitsky, V. Deep image prior. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT, USA, 18–22 June 2018; pp. 9446–9454. [Google Scholar]
  14. Krull, A.; Buchholz, T.O.; Jug, F. Noise2Void-learning denoising from single noisy images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Long Beach, CA, USA, 15–20 June 2019; pp. 2129–2137. [Google Scholar]
  15. Batson, J.; Royer, L. Noise2Self: Blind denoising by self-supervision. In Proceedings of the International Conference on International Conference on Machine Learning, Long Beach, CA, USA, 10–15 June 2019; pp. 524–533. [Google Scholar]
  16. Quan, Y.; Chen, M.; Pang, T.; Ji, H. Self2Self With Dropout: Learning Self-Supervised Denoising From Single Image. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 13–19 June 2020; pp. 1890–1898. [Google Scholar]
  17. Vogel, C.R.; Oman, M.E. Iterative methods for total variation denoising. SIAM J. Sci. Comput. 1996, 17, 227–238. [Google Scholar] [CrossRef]
  18. Vese, L.A.; Osher, S.J. Modeling textures with total variation minimization and oscillating patterns in image processing. J. Sci. Comput. 2003, 19, 553–572. [Google Scholar] [CrossRef]
  19. Vese, L.A.; Osher, S.J. Image denoising and decomposition with total variation minimization and oscillatory functions. J. Math. Imaging Vis. 2004, 20, 7–18. [Google Scholar] [CrossRef] [Green Version]
  20. Getreuer, P. Rudin-Osher-Fatemi total variation denoising using split Bregman. Image Process. Line 2012, 2, 74–95. [Google Scholar] [CrossRef]
  21. Tomasi, C.; Manduchi, R. Bilateral filtering for gray and color images. In Proceedings of the International Conference on Computer Vision IEEE, Bombay, India, 7 January 1998; pp. 839–846. [Google Scholar]
  22. Djurović, I. BM3D filter in salt-and-pepper noise removal. EURASIP J. Image Video Process. 2016, 2016, 13. [Google Scholar] [CrossRef] [Green Version]
  23. Parrilli, S.; Poderico, M.; Angelino, C.V.; Verdoliva, L. A nonlocal SAR image denoising algorithm based on LLMMSE wavelet shrinkage. IEEE Trans. Geosci. Remote Sens. 2012, 50, 606–616. [Google Scholar] [CrossRef]
  24. Wohlberg, B. SPORCO: A Python package for standard and convolutional sparse representations. In Proceedings of the Python in Science Conference, Austin, TX, USA, 10–16 July 2017; pp. 1–8. [Google Scholar]
  25. Bao, C.; Cai, J.F.; Ji, H. Fast sparsity-based orthogonal dictionary learning for image restoration. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Sydney, Australia, 1–8 December 2013; pp. 3384–3391. [Google Scholar]
  26. Bao, C.; Ji, H.; Quan, Y.; Shen, Z. Dictionary learning for sparse coding: Algorithms and convergence analysis. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 38, 1356–1369. [Google Scholar] [CrossRef] [PubMed]
  27. Elad, M.; Aharon, M. Image denoising via sparse and redundant representations over learned dictionaries. IEEE Trans. Image Process. 2006, 15, 3736–3745. [Google Scholar] [CrossRef] [PubMed]
  28. Papyan, V.; Romano, Y.; Sulam, J.; Elad, M. Convolutional dictionary learning via local processing. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 5296–5304. [Google Scholar]
  29. Mao, X.; Shen, C.; Yang, Y.B. Image Restoration Using Very Deep Convolutional Encoder-Decoder Networks with Symmetric Skip Connections. In Proceedings of the Advances in Neural Information Processing Systems 29 (NIPS 2016), Barcelona, Spain, 5–10 December 2016; pp. 2802–2810. [Google Scholar]
  30. Lefkimmiatis, S. Non-local color image denoising with convolutional neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 3587–3596. [Google Scholar]
  31. Laine, S.; Karras, T.; Lehtinen, J.; Aila, T. High-quality self-supervised deep image denoising. In Proceedings of the Advances in Neural Information Processing Systems 32 (NeurIPS 2019), Vancouver, BC, Canada, 8–14 December 2019; pp. 6968–6978. [Google Scholar]
  32. Xu, J.; Huang, Y.; Cheng, M.M.; Liu, L.; Zhu, F.; Xu, Z.; Shao, L. Noisy-As-Clean: Learning Self-supervised Denoising from Corrupted Image. IEEE Trans. Image Process. 2020, 29, 9316–9329. [Google Scholar] [CrossRef] [PubMed]
  33. Moran, N.; Schmidt, D.; Zhong, Y.; Coady, P. Noisier2Noise: Learning to Denoise from Unpaired Noisy Data. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 12064–12072. [Google Scholar]
  34. Yu, F.; Koltun, V. Multi-scale context aggregation by dilated convolutions. arXiv 2015, arXiv:1511.07122. [Google Scholar]
  35. Singh, V.; Dev, R.; Dhar, N.K.; Agrawal, P.; Verma, N.K. Adaptive type-2 fuzzy approach for filtering salt and pepper noise in grayscale images. IEEE Trans. Fuzzy Syst. 2018, 26, 3170–3176. [Google Scholar] [CrossRef]
  36. Abadi, M.; Agarwal, A.; Barham, P.; Brevdo, E.; Chen, Z.; Citro, C.; Corrado, G.S.; Davis, A.; Dean, J.; Devin, M.; et al. TensorFlow: Large-Scale Machine Learning on Heterogeneous Systems. arXiv 2016, arXiv:1603.04467. [Google Scholar]
  37. Liu, L.; Jiang, H.; He, P.; Chen, W.; Liu, X.; Gao, J.; Han, J. On the variance of the adaptive learning rate and beyond. arXiv 2019, arXiv:1908.03265. [Google Scholar]
  38. Plotz, T.; Roth, S. Benchmarking denoising algorithms with real photographs. In Proceedings of the IEEE Transactions on Pattern Analysis and Machine Intelligence, Honolulu, HI, USA, 21–26 July 2017; pp. 1586–1595. [Google Scholar]
Figure 1. Denoising results on the image highly corrupted by non-zero mean salt-and-pepper noise. Note that the results of N2S, N2V, and S2S look much darker than the ground truth. Our proposed method (N2K) successfully removes noise without brightness shifting.
Figure 1. Denoising results on the image highly corrupted by non-zero mean salt-and-pepper noise. Note that the results of N2S, N2V, and S2S look much darker than the ground truth. Our proposed method (N2K) successfully removes noise without brightness shifting.
Sensors 22 04255 g001
Figure 2. Overview of the network structure of Noise2Kernel.
Figure 2. Overview of the network structure of Noise2Kernel.
Sensors 22 04255 g002
Figure 3. An example of dependency between the pixels in the input (x) and the output (y) images with one 3 × 3 donut convolution filter and d-dilated convolution. Each row represents the dependency visualization with two 2-dilated convolution and two 3-dilated convolution layers, respectively. The green pixels indicate the pixel locations that have dependency with x i , j (the center pixel in x). The red pixels represent the trainable variables of the convolution kernels. The blue pixels indicate the area independent of x i , j . This figure shows the intermediate convolution processes of y = f ( x ) = f ( 2 ) ( f ( 1 ) ( f ( 0 ) ( x ) ) ) from the input image to the output prediction.
Figure 3. An example of dependency between the pixels in the input (x) and the output (y) images with one 3 × 3 donut convolution filter and d-dilated convolution. Each row represents the dependency visualization with two 2-dilated convolution and two 3-dilated convolution layers, respectively. The green pixels indicate the pixel locations that have dependency with x i , j (the center pixel in x). The red pixels represent the trainable variables of the convolution kernels. The blue pixels indicate the area independent of x i , j . This figure shows the intermediate convolution processes of y = f ( x ) = f ( 2 ) ( f ( 1 ) ( f ( 0 ) ( x ) ) ) from the input image to the output prediction.
Sensors 22 04255 g003
Figure 4. Quantitative performance comparison of various denoising methods on known noise models in the BSD68 and Set14 dataset. Left to right: AWGN, speckle noise, and salt-and-pepper, respectively. Top to bottom: BSD68 and Set14.
Figure 4. Quantitative performance comparison of various denoising methods on known noise models in the BSD68 and Set14 dataset. Left to right: AWGN, speckle noise, and salt-and-pepper, respectively. Top to bottom: BSD68 and Set14.
Sensors 22 04255 g004
Figure 5. Qualitative performance comparison of various denoising methods on three noise types. Top to bottom row: AWGN ( σ g = 50), speckle noise ( σ s = 50), and salt-and-pepper noise (d = 50), respectively. The best PSNR in each case is highlighted in bold. Each row indicates the results of AWGN, speckle noise, and salt-and-pepper noise, respectively.
Figure 5. Qualitative performance comparison of various denoising methods on three noise types. Top to bottom row: AWGN ( σ g = 50), speckle noise ( σ s = 50), and salt-and-pepper noise (d = 50), respectively. The best PSNR in each case is highlighted in bold. Each row indicates the results of AWGN, speckle noise, and salt-and-pepper noise, respectively.
Sensors 22 04255 g005
Figure 6. Qualitative performance comparison of various denoising methods on fusion noise. Top to bottom ( σ g , σ s , d): (25,5,5), (25,5,25), (25,25,5), (25,25,25), (50, 5, 5), (50, 5, 25), (50, 25, 5), (50, 25, 25), respectively. The best PSNR is highlighted in bold, except N2C.
Figure 6. Qualitative performance comparison of various denoising methods on fusion noise. Top to bottom ( σ g , σ s , d): (25,5,5), (25,5,25), (25,25,5), (25,25,25), (50, 5, 5), (50, 5, 25), (50, 25, 5), (50, 25, 25), respectively. The best PSNR is highlighted in bold, except N2C.
Sensors 22 04255 g006
Figure 7. Effect of the loss clipping for various threshold values on salt-and-pepper noise (d = 50).
Figure 7. Effect of the loss clipping for various threshold values on salt-and-pepper noise (d = 50).
Sensors 22 04255 g007
Figure 8. Average PSNR graph of loss clipping on Set 14 in terms of ε .
Figure 8. Average PSNR graph of loss clipping on Set 14 in terms of ε .
Sensors 22 04255 g008
Table 1. Performance of baselines, the proposed method and the proposed method+TV on the BSD68 test set. Boldface denotes the best among all except N2C.
Table 1. Performance of baselines, the proposed method and the proposed method+TV on the BSD68 test set. Boldface denotes the best among all except N2C.
Noise Level σ g = 25, σ s = 5, d = 5 σ g = 25, σ s = 5, d = 25 σ g = 25, σ s = 25, d = 5 σ g = 25, σ s = 25, d = 25
Method\MetricPSNRSSIMPSNRSSIMPSNRSSIMPSNRSSIM
N2C26.990.758826.240.730324.680.667323.960.6353
N2V24.610.681720.960.590821.880.594019.290.5111
N2S24.420.678921.160.587921.490.572719.030.4896
N2K (ours)25.280.689224.520.643522.420.558021.460.4869
N2K+TV (ours)25.130.685324.420.651322.610.604321.860.5673
Noise Level σ g = 50, σ s = 5, d = 5 σ g = 50, σ s = 5, d = 25 σ g = 50, σ s = 25, d = 5 σ g = 50, σ s = 25, d = 25
Method\MetricPSNRSSIMPSNRSSIMPSNRSSIMPSNRSSIM
N2C25.210.678224.380.639123.850.622522.960.5810
N2V22.640.593019.830.533720.570.544418.480.4794
N2S22.000.574619.710.499919.950.514118.410.4404
N2K (ours)23.400.603822.550.547120.490.506319.730.4321
N2K+TV (ours)23.400.614922.670.578620.590.563519.820.5195
Table 2. Performance of baselines, the proposed method and the proposed method+TV on the Set14 test set. Boldface denotes the best among all except N2C.
Table 2. Performance of baselines, the proposed method and the proposed method+TV on the Set14 test set. Boldface denotes the best among all except N2C.
Noise Level σ g = 25, σ s = 5, d = 5 σ g = 25, σ s = 5, d = 25 σ g = 25, σ s = 25, d = 5 σ g = 25, σ s = 25, d = 25
Method\MetricPSNRSSIMPSNRSSIMPSNRSSIMPSNRSSIM
N2C28.060.774927.230.747625.610.691824.790.6615
N2V25.510.707420.930.608922.590.619919.320.5335
N2S24.060.668320.400.580521.340.579718.680.4971
S2S25.720.725620.880.595122.580.625219.270.5149
N2K (ours)26.420.716925.460.667423.250.578222.190.4992
N2K+TV (ours)26.260.716325.330.679123.520.637222.670.5966
Noise Level σ g = 50, σ s = 5, d = 5 σ g = 50, σ s = 5, d = 25 σ g = 50, σ s = 25, d = 5 σ g = 50, σ s = 25, d = 25
Method\MetricPSNRSSIMPSNRSSIMPSNRSSIMPSNRSSIM
N2C26.160.702925.190.663124.660.649723.620.6082
N2V23.010.619219.670.557220.880.569918.310.5005
N2S21.650.571419.070.499419.260.510317.760.4438
S2S23.310.644119.640.536121.040.572518.350.4748
N2K (ours)24.240.630523.220.570821.200.524820.380.4438
N2K+TV (ours)24.240.647123.350.607621.310.594420.490.5452
Table 3. Comparison of ADSS and general self-supervision loss. Average PSNR and SSIM for fusion noise on BSD68 validation set. The baseline uses only the structure of the proposed method with general self-supervision L 2 loss. Boldface denotes the best performance among Baseline, ADSS, and ADSS +TV.
Table 3. Comparison of ADSS and general self-supervision loss. Average PSNR and SSIM for fusion noise on BSD68 validation set. The baseline uses only the structure of the proposed method with general self-supervision L 2 loss. Boldface denotes the best performance among Baseline, ADSS, and ADSS +TV.
ModelBaselineADSSADSS + TV
Noise Level\MetricPSNRSSIMPSNRSSIMPSNRSSIM
σ g = 25, σ s = 5, d = 524.540.676125.280.689225.130.6853
σ g = 25, σ s = 5, d = 2520.930.557724.520.643524.420.6513
σ g = 25, σ s = 25, d = 521.660.567922.420.558021.610.6043
σ g = 25, σ s = 25, d = 2519.220.485021.460.486921.860.5673
σ g = 50, σ s = 5, d = 522.540.587223.400.603823.400.6149
σ g = 50, σ s = 5, d = 2519.710.516222.550.547122.670.5786
σ g = 50, σ s = 25, d = 520.590.539020.490.506320.590.5635
σ g = 50, σ s = 25, d = 2519.220.485021.460.486921.860.5673
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lee, K.; Jeong, W.-K. Noise2Kernel: Adaptive Self-Supervised Blind Denoising Using a Dilated Convolutional Kernel Architecture. Sensors 2022, 22, 4255. https://doi.org/10.3390/s22114255

AMA Style

Lee K, Jeong W-K. Noise2Kernel: Adaptive Self-Supervised Blind Denoising Using a Dilated Convolutional Kernel Architecture. Sensors. 2022; 22(11):4255. https://doi.org/10.3390/s22114255

Chicago/Turabian Style

Lee, Kanggeun, and Won-Ki Jeong. 2022. "Noise2Kernel: Adaptive Self-Supervised Blind Denoising Using a Dilated Convolutional Kernel Architecture" Sensors 22, no. 11: 4255. https://doi.org/10.3390/s22114255

APA Style

Lee, K., & Jeong, W. -K. (2022). Noise2Kernel: Adaptive Self-Supervised Blind Denoising Using a Dilated Convolutional Kernel Architecture. Sensors, 22(11), 4255. https://doi.org/10.3390/s22114255

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop