Next Article in Journal
Efficient Index Modulation-Based MIMO OFDM Data Transmission and Detection for V2V Highly Dispersive Channels
Next Article in Special Issue
An Analysis of Climate Change Based on Machine Learning and an Endoreversible Model
Previous Article in Journal
Study of Dynamic Solutions for Human–Machine System with Human Error and Common-Cause Failure
Previous Article in Special Issue
Multi-Country and Multi-Horizon GDP Forecasting Using Temporal Fusion Transformers
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

NSNet: An N-Shaped Convolutional Neural Network with Multi-Scale Information for Image Denoising

1
School of Economics and Management, Changsha University, Changsha 410022, China
2
School of Automation, Central South University, Changsha 410083, China
*
Author to whom correspondence should be addressed.
Mathematics 2023, 11(12), 2772; https://doi.org/10.3390/math11122772
Submission received: 4 June 2023 / Revised: 15 June 2023 / Accepted: 18 June 2023 / Published: 19 June 2023
(This article belongs to the Special Issue Data Mining and Machine Learning with Applications)

Abstract

:
Deep learning models with convolutional operators have received widespread attention for their good image denoising performance. However, since the convolutional operation prefers to extract local features, the extracted features may lose some global information, such as texture, structure, and color characteristics, when the object in the image is large. To address this issue, this paper proposes an N-shaped convolutional neural network with the ability to extract multi-scale features to capture more useful information and alleviate the problem of global information loss. The proposed network has two main parts: a multi-scale input layer and a multi-scale feature extraction layer. The former uses a two-dimensional Haar wavelet to create an image pyramid, which contains the corrupted image’s high- and low-frequency components at different scales. The latter uses a U-shaped convolutional network to extract features at different scales from this image pyramid. The method sets the mean-squared error as the loss function and uses the residual learning strategy to learn the image noise directly. Compared with some existing image denoising methods, the proposed method shows good performance in gray and color image denoising, especially in textures and contours.

1. Introduction

Image denoising is one of the basic tasks of computer vision and is of wide interest to academia and industry, as it can effectively improve image quality. The purpose of image denoising is to remove noise from a corrupted image and restore its original content as much as possible. In many computer vision tasks, image denoising is often used as a preprocessing method to improve the practical performance of advanced computer vision tasks [1]. Over the past few decades, many outstanding image denoising methods, as shown in Figure 1, have been proposed, including filtering-based [2,3], sparse-representation-based [4,5,6,7,8], external-prior-based [9,10,11,12], low-rank-representation-based [13,14], and deep-learning-based methods [15,16,17,18].
Filtering-based methods were the first techniques to be applied to image denoising and rely on the self-similarity of images. Well-known approaches include Gaussian filtering, mean filtering, and median filtering. These three methods assume that the pixels in an image do not exist in isolation and have connections to other pixels. However, Buades et al. [2] found that similar pixels are not limited to local areas, and making full use of the redundant information in an image can improve the image denoising performance. Hence, they proposed a nonlocal mean filtering method (NLM) based on existing smoothing filtering methods. Although NLM can achieve good denoising performance, it needs to find a sufficient number of similar blocks when computing each pixel, which gives it high computational complexity. To solve this problem, Kostadin et al. [3] proposed a block-matching and 3D filtering (BM3D) method, which has a good denoising performance and fast computational speed.
Sparse-representation-based methods are based on image sparsity and achieve image denoising by training an over-complete dictionary. A more representative case is the K-singular value decomposition (KSVD) method using sparse representation [4]. Inspired by KSVD, Mairal et al. [5] combined image self-similarity with sparse coding to decompose similar patches using similar sparse patterns, thus forming a Learned Simultaneous Sparse Coding (LSSC) method. Although sparse representation models have shown good results in image denoising, the sparse representation of traditional models may not be accurate enough due to the degradation of the observed images. To further improve the performance of image denoising based on sparse representation, Dong et al. [6] proposed a nonlocally centralized sparse representation (NCSR), which transformed the denoising problem into a problem of suppressing sparse coding noise. In addition, because the sparse coding of images using a single transform can limit performance, Wen et al. [7] proposed a structured over-complete sparsifying transform model with block cosparsity (OCTOBOS). These methods [4,5,6,7] have exhibited good results in denoising additive Gaussian white noise (AWGN); however, it is difficult to obtain good performance in real image denoising. To achieve better denoising of real images, Xu et al. [8] proposed a trilateral weighted sparse coding scheme (TWSC).
External-prior-based methods realize image denoising by using the statistical properties of natural images. A representative method is the denoising method based on expected patch log likelihood (EPLL) proposed by Zoran and Weiss [9]. This method applies the Gaussian mixture model to learn prior knowledge from a large number of natural image blocks and applies it to the denoising of other natural images. Similar to EPLL, Xu et al. [10] proposed a patch group prior-based denoising method (PGPD) to learn the self-similar features of natural images from groups of similar patches using the Gaussian mixture distribution. Inspired by EPLL, Chen et al. [11] proposed an external patch prior-guided internal clustering approach by combining an image external prior and an internal self-similarity prior, which is named PCLR. To improve the texture restoration capability of the image denoising method, Zou et al. [12] proposed a gradient histogram preservation method (GHP) based on texture enhancement. GHP improves texture recovery by preserving the gradient distribution of the corrupted image.
Low-rank representation-based methods exploit the low-rank properties of natural images and achieve denoising by extracting their low-rank components. A typical case is the Weighted Nuclear Norm Minimization (WNNM) proposed by Gu et al. [13]. Low-rank matrix factorization is also a method used to extract low-rank components from a dataset and is often applied in cases where the image size is large and its rank is much smaller than the length and width of the dataset. The most well-known method is the low-rank matrix factorization based on variational Bayesian (VBMFL), which was proposed by Zhao et al. [14]. This method improves the robustness of the model to outliers by using a Laplace distribution to establish a noise model.
Deep-learning-based denoising methods are currently the most popular. They usually learn the direct mapping from the corrupted image to the clean image or the noise. Since deep-learning-based denoising methods do not rely on image priori (e.g., self-similarity, sparsity, gradient, statistical properties, and low-rank properties), they do not have to spend much time finding and processing similar blocks in the images. Thus, they not only achieve a good denoising performance but also have a fast inference speed. Schmidt et al. [15] proposed a method based on a cascade of shrinkage fields (CSF) to improve the denoising performance while considering computational efficiency. Chen et al. [16] extended conventional nonlinear reaction–diffusion models with several parametrized linear filters as well as several parametrized influence functions and proposed a trainable nonlinear reaction–diffusion method (TNRD). Although CSF and TNRD show good denoising performance, they can only provide the best denoising results at known noise levels. To solve the problem of blind image denoising, Zhang et al. [17] proposed a deep learning method using a denoising convolutional neural network (DnCNN), which was the first application of residual learning to general image denoising. The application of residual learning to image denoising has greatly improved the denoising performance of networks and inspired many outstanding denoising methods based on deep learning [17,18,19,20,21,22]. In addition, Zhang et al. [18] further improved the DnCNN and proposed a fast and flexible denoising convolutional neural network (FFDNet), which achieves a good trade-off between the inference speed and denoising performance by downsampling and manually inputting a noise estimation map. Binh et al. [23] combined DnCNN with ResNet and proposed a convolutional denoising neural network called FlashLight CNN. A complex-valued deep convolutional neural network called CDNet was proposed by Quan et al. [24], and it effectively improved the denoising performance of the network. Guan et al. [25] proposed an image denoising method for remote sensing images called MRFENet. It demonstrated good denoising performance and preserved the edge details of the images. Zhang et al. [26] utilized dilated convolutions to capture more contextual information and then proposed a hybrid denoising neural network called HDCNN to enhance the denoising performance of CNN networks in complex application scenarios. Tian et al. [27] combined dynamic convolution, wavelet transform, and discriminative learning to propose a convolutional neural network based on the wavelet transform called Multi-stage Image Denoising CNN with the Wavelet Transform (MWDCNN). To reduce the parameter size and training burden of deep denoising networks, Tang et al. [28] employed a cascaded residual network and proposed a lightweight, multi-scale, efficient convolutional neural network.
The results of most denoising methods are obtained directly from the fusion of high-level features, while low-level features containing texture and contour information are ignored, resulting in the loss of some important information. Furthermore, since the convolutional operation prefers to extract local features, it is difficult to extract global information such as textures and contours when the objects in the image are relatively large. To solve these problems, an N-shaped convolutional neural network, named NSNet, using multi-scale features is proposed in this paper. In this model, a 2D Haar wavelet is used to construct an image pyramid that contains high- and low-frequency components of the corrupted image at different scales. The multi-scale features are extracted from the image pyramid by a U-shaped convolutional network [29], and the low- and high-level features are fused by skip connections in the U-shaped network. The 2D Haar wavelet is widely used in image denoising, and many scholars have achieved excellent denoising performance with it [30,31,32,33]. To verify the denoising performance of NSNet, the denoising of gray and color images was carried out at different noise levels and compared with existing denoising methods. The contributions of this work are summarized as follows:
(1)
An N-shaped convolutional neural network for extracting multi-scale information is proposed. The network exploits multi-scale information to compensate for the drawbacks of convolutional operations in extracting global features, which effectively improves the network’s ability to recover textures and contours.
(2)
A scheme for constructing image pyramids using a 2D Haar wavelet is proposed. The image pyramid is obtained by using a multi-scale 2D Haar wavelet, and each layer of the pyramid contains one low-frequency component and three high-frequency components. In image denoising, the high-frequency components can be used as an estimate of the noise level to facilitate denoising.
(3)
NSNet shows good denoising performance for AWGN at a noise level range of (0, 55) and good recovery of textures and contours. It provides a solution for applications that need not only denoising but also texture and contour recovery.
The rest of this paper is organized as follows. Section 2 presents the techniques involved in the proposed model. Section 3 describes the proposed NSNet and the construction of the image pyramid in detail. Section 4 presents the results of experiments, and Section 5 concludes the paper.

2. Theoretical Aspects

2.1. The 2D Haar Wavelet

The process of decomposing an image using the 2D Haar wavelet is shown in Figure 2. The blocks with five-pointed stars in the figure are the finite impulse response filter. h φ = { 1 2 , 1 2 } denotes a low-pass filter, and h ψ = { 1 2 , 1 2 } denotes a high-pass filter. The down arrow ( ) indicates downsampling, which means adding two adjacent pixels in the column or row direction.
The image is first processed with the two filters separately and is compressed in the column direction to obtain a low-frequency component L and a high-frequency component H . The two components are then processed with low- and high-pass filters in turn and compressed in the row direction to obtain the low-frequency component L L , the high-frequency component L H in the vertical direction, the high-frequency component H L in the horizontal direction, and the high-frequency component H H in the diagonal direction. According to the principle of 2D Haar wavelets, assuming that the corrupted image is X R m × n , the formulas for the components of the image can be simplified as
A i j = X 2 i , 2 j , B i j = X 2 i + 1,2 j
C i j = X 2 i , 2 j + 1 , D i j = X 2 i + 1,2 j + 1
L L i j = 1 2 A i j + B i j + C i j + D i j , H L i j = 1 2 B i j + D i j A i j C i j
L H i j = 1 2 C i j + D i j A i j B i j , H H i j = 1 2 A i j + D i j B i j C i j
where { i | 0 i m / 2 } and { j | 0 j n } . The two-scale 2D Haar wavelet applied to the image “Monarch” is shown in Figure 3. Through the 2D Haar wavelet, the image is decomposed into three high-frequency components ( L H , H L , H H ) and a low-frequency component ( L L ).

2.2. U-Shaped Convolutional Network

The U-shaped convolutional network was first proposed by Ronneberger et al. [29]. The network has been widely used in various fields for its powerful encoding and decoding capabilities [34,35]. The U-shaped convolutional network includes four downsampling operators and four upsampling operators and uses skip connections at the same stage, which not only gives the network the ability to extract multi-scale features but also ensures that the output integrates more low-level features [21]. The structure of the U-shaped convolutional network with batch normalization (BN) is shown in Figure 4.
When the neural network becomes very deep, an internal covariate offset may occur, which can lead to two problems: (1) it affects the learning efficiency and makes the learning process unstable, and (2) it makes the input data of the later layers too large or small, thus falling into the saturation area of the activation function and causing the learning process to stop prematurely. To solve the problem of the internal covariate offset, a general approach is to add BN, as proposed by Ioffe and Szegedy [36], to the U-shaped convolutional network.
The BN layer is usually placed between the convolutional operation and the Rectified Linear Unit (ReLU), and the parameters of the BN layer are adjusted by training. Supposing that there is a mini-batch B = { X 1 m } of size m , and the parameters to be learned are γ and β , the BN process can be expressed as
γ = V a r [ x ] , β = E x ,
μ B = 1 m i = 1 m x i , δ B 2 = 1 m i = 1 m ( x i μ B ) 2 ,
x ^ i = x i μ B δ B 2 + ε , y i = γ x ^ i + β ,
where ε is a constant added to the mini-batch variance for numerical stability.

2.3. Residual Learning

When the network becomes very deep, some convolutional layers may appear to have identity mapping, resulting in degradation problems and vanishing and exploding gradients. To solve this problem, He et al. [37] proposed a residual network with the residual block shown in Figure 5. It connects the input and output directly through a shortcut connection, allowing F ( x ) to learn small changes. This not only allows the convolutional layer to maintain identity mapping but also avoids vanishing and exploding gradients. The relationship between the input x and output x * of the residual block is
x * = F ( x ) + x .
In image denoising, when the noise level is low, the mapping from the noisy image to the clean image is close to an identity mapping, which is not conducive to the training of the network. To solve this problem, Zhang et al. [17] first applied residual learning to image denoising. Assuming that the input image is I n p u t and the output image is O u t p u t , their relationship is
F ( x ) = O u t p u t I n p u t .
It can be seen from (9) that residual learning for image denoising uses noise as the training target, which is a valuable technique for improving the denoising performance of the model.

3. The Proposed NSNet Model

In this section, the proposed NSNet model is introduced in detail; its architecture is shown in Figure 6. It mainly consists of a multi-scale input layer and a multi-scale feature extraction layer. The multi-scale input layer uses a 2D Haar wavelet to create an image pyramid, which decomposes the corrupted image into high- and low-frequency components at different scales. The multi-scale feature extraction layer uses a U-shaped convolutional network to extract features at different scales from the image pyramid. Additionally, NSNet sets the mean-squared error as the loss function and uses the residual learning strategy to learn the noise directly.
The 2D Haar wavelet can decompose the image into four sub-images, each with a size half that of the original image. By using a 2D Haar wavelet to decompose the image, we can obtain
L L 1 , L H 1 , H L 1 , H H 1 = d w t y ,
where d w t ( · ) represents the 2D Haar wavelet, y is the corrupted image, L L 1 is the low-frequency component, and L H 1 , H L 1 , and H H 1 are the high-frequency components. Then, the 2D Haar wavelet is applied once again to the low-frequency component L L 1 to obtain
L L 2 , L H 2 , H L 2 , H H 2 = d w t L L 1 ,
Finally, to obtain the image pyramid shown in Figure 7, the same operation is repeated twice, resulting in
L L 3 , L H 3 , H L 3 , H H 3 = d w t L L 2 ,
L L 4 , L H 4 , H L 4 , H H 4 = d w t L L 3 .
The image pyramid contains images at five different scales, each of which corresponds to a different stage of the U-shaped convolutional network. In addition to the original scale, each scale contains a low-frequency component L L and three high-frequency components L H , H L , and H H . As shown in Figure 7, the low-frequency component is close to the corrupted input image, while the high-frequency components contain a lot of noise and some textures, which can be considered an estimate of the noise level.
The image degradation model is established as y = x + v , where y denotes the corrupted image, x denotes the clean image, and v denotes the noise. The proposed model inputs the corrupted image y into the network to predict the noise v F ( y ) and finally obtains the clean image x = y v through simple subtraction. The mean-squared error is used as the loss function:
L ( θ ) = 1 2 N i = 1 N | F y i ; θ y i x i | F 2 ,
where θ represents the parameter set of the model, N is the total number of images, and x i and y i represent the ith clean image and noisy image, respectively.
For convenience, the proposed model trained at a known noise level is named NSNet-S, and the model trained at an unknown noise level is named NSNet-B. The pseudo-code of the proposed method is shown in Algorithm 1.
Algorithm 1 The algorithm of NSNet
Input: All training images D from the observed dataset, denoising mode (B or S), noise level n o i s e L , range of noise level n o i s e R , maximum epoch M e p o c h .
Output: The trained network f .
1:Initialing model parameters θ and learning rate η ;
2:Sampling m patches ( x 1 , , x m ) from D;
3:for epoch = 1 to M e p o c h  do
4:if epoch > 30 then
5:     η η / 10 ;
6: end if
7:  Set g ^ = 0 ;
8:for i = 1 to m  do
9:   if mode == “B” then
10:      Setting n o i s e L as an integer at the range n o i s e R randomly;
11:    end if
12:   Adding Gaussian noise with the noise level of n o i s e L to x i :
y i = x i + n o i s e i ;
13:   Performing multi-scale wavelet transform on y i to obtain y i 1 , y i 2 , y i 3 , y i 4 :
y i 1 = { L L 1 , L H 1 , H L 1 , H H 1 } = d w t y i , y i 2 = L L 2 , L H 2 , H L 2 , H H 2 = d w t L L 1 , y i 3 = L L 3 , L H 3 , H L 3 , H H 3 = d w t L L 2 , y i 4 = { L L 4 , L H 4 , H L 4 , H H 4 } = d w t L L 3 ;
14:   Predicting noise using the network f with parameter θ :
o u t i f y i , y i 1 , y i 2 , y i 3 , y i 4 ; θ ;
15:   Calculating the loss according to Equation (14);
16:   Computing the gradient: g ^ g ^ + 1 m θ L θ ;
17:end for
18: Updating θ : θ θ η × g ^ ;
19:end for

4. Experimental Results

Gray image denoising and color image denoising were carried out to compare the denoising performance of the proposed NSNet with those of existing models, including NLM [2], BM3D [3], KSVD [4], NCSR [6], OCTOBOS [7], TWSC [8], GHP [9], EPLL [10], PGPD [11], PCLR [12], WNNM [13], CSF [15], TNRD [16], and FFDNet [18]. Moreover, two different types of DnCNNs [17] were also selected as the compared models. They are DnCNN-S and DnCNN-B, which are trained at known and unknown noise levels, respectively.

4.1. Evaluation Metrics

The results of all denoising methods were analyzed quantitatively in terms of the peak signal-to-noise ratio (PSNR) and structural similarity (SSIM).
(1) Supposing that the recovered image is I R m × n and the corrupted image is K R m × n , the PSNR is calculated as
M S E = 1 m n i = 0 m 1 j = 0 n 1 [ I i , j K ( i , j ) ] 2 ,
P S N R = 10 · log 10 M A X I 2 M S E ,
where M S E is the mean-squared error, and M A X I denotes the maximum value of the pixels in the image. In general, M A X I = 255 if each pixel is represented in 8-bit binary form or 1 if it is represented in 1-bit binary.
(2) The SSIM is calculated as
S S I M ( I , K ) = ( 2 μ I μ K + c 1 ) ( σ I K + c 2 ) ( μ I 2 + μ K 2 + c 1 ) ( σ I 2 + σ K 2 + c 2 )
where μ I and μ K denote the means of I and K , respectively, and σ I and σ K denote their standard deviations, while σ I K denotes the covariance of I and K , and c 1 and c 2 are constants.

4.2. Experimental Setting

For the ablation experiment, NSNet, NSNet without BN, Unet, and Unet without BN were compared. All compared methods were trained using 400 images of size 180 × 180 pixels, as mentioned in [17]. The test sets were Set12, which is widely used in the evaluation of denoising methods, and BSD68 [38]. In training the model, the size of the patch was set to 48 × 48, and 128 × 618 patches were cropped from the 400 images. Four denoising methods were trained at noise levels of 15, 25, 35, 45, and 50. For a noise level of α , the noise was generated by a Gaussian distribution with a mean of zero and a variance of α .
For the denoising of gray images, the 400 images were still used as the training set, and 128 × 2934 patches of size 48 × 48 were cropped from them. Since most image denoising methods can only obtain the best denoising performance at a known noise level, to achieve a fair comparison, the proposed method was trained at an unknown noise level and at noise levels of 15, 25, and 50. The test sets were Set12 and BSD68, neither of which participated in model training.
For color image denoising, 432 images were selected from the color image dataset CBSD500 [39] as training samples, and the remaining 68 images (CBSD68) were used as the test set. The test set also included Kodak24 [40] and McMaster [41]. In this experiment, 96 × 3900 patches were cropped from the 432 images to train the color image denoising model. The other settings were largely consistent with the settings used for gray image denoising. The specific settings of NSNet are shown in Table 1.
When training NSNet-S, each clean image input to the model was corrupted by the same level of noise. When training NSNet-B, each clean image input to the model was corrupted by noise at a level drawn randomly from the range (0, 55). The Adam optimizer was used to tune the model with an initial learning rate of 0.001. The maximum training epoch was 50. After 30 epochs, the learning rate was adjusted to 0.0001. The size of each mini-batch was set to 128. The denoising network was trained in PyTorch, and all experiments were carried out in the pycharm environment running on a PC with an AMD Ryzen 9 5900HX with Radeon Graphics 3.30 GHz CPU and an NVIDIA GeForce RTX 3070 Laptop GPU.

4.3. Ablation Experiment

This section describes the ablation experiment that was carried out to demonstrate the effectiveness of the main components of the proposed model. The experiment tested the denoising performance of NSNet, Unet, NSNet without BN, and Unet without BN at noise levels of 15, 25, 35, 45, and 50. The denoising results on the Set12 dataset and the gray version of BSD68 are shown in Table 2, in which values with # and * represent the best and second-best denoising performance, respectively.
The denoising performance of NSNet is better than that of Unet at all noise levels, which shows that multi-scale input can improve the denoising performance. The results for NSNet and NSNet without BN show that the denoising performance of NSNet is greatly improved after adding the BN layer. As mentioned in [29], the BN layer can improve the denoising performance of the neural network. At all noise levels, NSNet without BN achieves better denoising than Unet without BN, and its denoising performance is close to that of Unet at low noise levels, which indicates that the multi-scale input greatly improved the denoising performance of the model at low noise levels. With increases in the noise level, the structure of the compressed image is increasingly corrupted; thus, it cannot provide accurate information for the network. In this case, the performance of NSNet is similar to that of Unet. For example, the denoising performance of NSNet is similar to that of Unet at a noise level of 50.

4.4. Gray Image Denoising

In this experiment, AWGN was added to Set12 and the gray version of BSD68, and the noise levels were set to 15, 25, and 50. The denoising results of all methods on Set12 are shown in Table 3, Table 4 and Table 5. Due to insufficient parameters, CSF cannot be tested at a noise level of 50.
Table 3 shows the denoising results of all methods at a noise level of 15. NSNet-S has a better denoising performance than other methods and obtained the first-ranked denoising performance on ten images. NSNet-S ranked a very close second in terms of denoising the image “House”. In this case, the denoising performance of NSNet-S is 1.86 dB higher than that of the worst method, NLM, on average, and 0.1 dB higher than that of DnCNN-S, on average, in terms of PSNR. The average results of all methods show that there is little difference in the denoising performance of most methods at a noise level of 15. At a low noise level, the self-similarity, sparsity, and low-rank properties of the image are still relatively complete, and the traditional denoising methods (e.g., BM3D, NCSR, TWSC, PCLR, WNNM) also achieve a good denoising performance.
Table 4 shows the denoising results obtained by all methods at a noise level of 25. With increases in the noise level, the best and second-best denoising performance was obtained with the deep-learning-based methods, which shows the superiority of deep learning in image denoising. As the noise level increases, the self-similarity and other features of the image are increasingly corrupted, which leads to the fact that the traditional image-prior-based denoising methods are no longer advantageous. The deep-learning-based denoising methods learn the potential noise directly from the corrupted image and rely less on the prior knowledge of the image. This allows them to achieve a good denoising performance, even at high noise levels. In this case, NSNet-S is ranked first in terms of denoising performance on ten images and second on one image. NSNet-B is ranked second in terms of denoising performance on three images. The denoising performance of NSNet-S is 1.99 dB higher than that of the worst method, NLM, and 0.17 dB higher than that of the second-best method, DnCNN-S, in terms of PSNR. Compared with the traditional denoising methods BM3D, NCSR, TWSC, and WNNM, NSNet-B is close to WNNM and outperforms BM3D, NCSR, and TWSC at a noise level of 25.
Table 5 shows the denoising results obtained by all methods at a noise level of 50. In this case, NSNet-S is ranked first in terms of denoising performance on eight images and second on two images. NSNet-B is ranked first in terms of denoising performance on two images and second on six images. Compared with other methods, NSNet-B provides a better denoising performance at a high noise level. The reason for this may be that an image with high noise will cause a greater deviation than one with low noise, and the network will pay more attention to the restoration of images with high noise when training NSNet-B. In this case, NSNet-S is 2.86 dB higher than the worst method, NLM, and 0.17 dB higher than the second-best method, FFDNet. NSNet-B also has an outstanding performance; its denoising performance surpasses that of DnCNN-B and WNNM and differs from that of FFDNet by only 0.03 dB.
In the above three experiments, NSNet shows a good denoising performance in most cases, although that of NSNet on the image “Barbara” is not as good as that of traditional methods (e.g., TWSC, WNNM). “Barbara” has similar rich textures, and the method based on image self-similarity can effectively use them to achieve better denoising. Both texture and noise are high-frequency information; therefore, image denoising methods that use residual learning tend to treat texture as noise, which makes the denoising performance of the proposed model poor. In addition, the same experiments were conducted on the dataset BSD68 to better demonstrate the denoising performance of NSNet. The average PSNRs and SSIMs of all methods on Set12 and BSD68 are shown in Table 6. Compared with other methods, the average denoising performance of NSNet-S on the two datasets is the best, and NSNet-B has a better denoising performance at a high noise level.
Figure 8 and Figure 9 show the denoising performance of all compared methods on gray images at noise levels of 25 and 50, respectively. In order to facilitate the comparison of the denoising performance, the results were transformed into pseudo-color images. The red box in Figure 8 shows the restoration effect of all compared methods on the “grass”. It can be seen that NSNet’s recovery of the “grass” is closer to the clean image than the other compared methods and results in sharper and clearer edges and textures. The red box in Figure 9 further demonstrates the advantages of NSNet in edge and texture restoration. As shown in Figure 9, compared to other methods, NSNet can not only make the edges and textures sharper but also restore more image details.

4.5. Color Image Denoising

The previous section compared the denoising performance of the methods on gray images, where BM3D, DnCNN, and FFDNet had the better denoising performance and computational speed. In the experiment described in this subsection, these methods were selected for a further comparative test of the denoising ability of the proposed model with color images. The datasets used here are the color versions of CBSD68, Kodak24, and McMaster.
The denoising performance of four methods (BM3D, DnCNN, FFDNet, and NSNet) at noise levels of 35, 45, and 55 is shown in Table 7. The proposed NSNet is the best model in the denoising experiments with CBSD68 and Kodak24, with McMaster ranked second.
In the denoising experiment with the dataset CBSD68, NSNet performed 0.81 dB better than BM3D, 0.09 dB better than DnCNN, and 0.1 dB better than FFDNet at a noise level of 35. It was 0.74 dB better than BM3D, 0.15 dB better than DnCNN, and 0.14 dB better than FFDNet at a noise level of 45, and 0.73 dB better than BM3D, 0.2 dB better than DnCNN, and 0.14 dB better than FFDNet at a noise level of 55. In the denoising experiment with the dataset Kodak24, NSNet was 0.78 dB better than BM3D, 0.22 dB better than DnCNN, and 0.11 dB better than FFDNet at a noise level of 35. It was 0.68 dB better than BM3D, 0.28 dB better than DnCNN, and 0.15dB better than FFDNet at a noise level of 45, and 0.66 dB better than BM3D, 0.34 dB better than DnCNN, and 0.15 dB better than FFDNet at a noise level of 55. In the denoising experiment using the McMaster dataset, although the denoising performance of NSNet was not as good as that of FFDNet, the comparison of all methods in terms of SSIM shows that NSNet is able to recover the structure of the color image better.
Figure 10 shows the denoising results of four methods (BM3D, DnCNN, FFDNet, and NSNet) on one image of the public dataset CBSD68 at a noise level of 45. To better demonstrate the denoising performance of the proposed method, two representative parts are highlighted. These show that NSNet repairs the texture better than other methods. Figure 11 shows the denoising results of four methods at a noise level of 40. NSNet has a more significant repair effect on “stone” in the image, and its recovery of textures is better than those of other methods.

5. Conclusions

In this study, an N-shaped convolutional network that can extract multi-scale information is proposed. NSNet is able to extract multi-scale information from corrupted images and uses it to compensate for the drawbacks of convolutional operations in extracting global features, thus enhancing NSNet’s ability to capture global information and reconstruct textures and contours. Ablation experiments demonstrate that extracting multi-scale information is beneficial to improving the denoising performance of the model at noise levels in the range of (0, 50). Gray and color image denoising experiments demonstrate that NSNet outperforms many existing image denoising methods at noise levels of 15, 25, and 50, especially at high noise levels. In addition, NSNet has a good blind denoising performance, where its performance at high noise levels is close to that at known noise levels.

Author Contributions

Conceptualization, writing—review and editing, visualization, supervision, and funding acquisition, Y.L. Methodology, software, validation, investigation, and writing—original draft preparation, Y.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Changsha Municipal Natural Science Foundation under Grant kq2208432.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Chatterjee, P.; Milanfar, P. Is denoising dead? IEEE Trans. Image Process. 2009, 19, 895–911. [Google Scholar] [CrossRef] [PubMed]
  2. Buades, A.; Coll, B.; Morel, J.-M. A non-local algorithm for image denoising. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Diego, CA, USA, 20–25 June 2005; Volume 2, pp. 60–65. [Google Scholar]
  3. Dabov, K.; Foi, A.; Katkovnik, V.; Egiazarian, K. Image denoising by sparse 3-D transform-domain collaborative filtering. IEEE Trans. Image Process. 2007, 16, 2080–2095. [Google Scholar] [CrossRef] [PubMed]
  4. Elad, M.; Aharon, M. Image denoising via sparse and redundant representations over learned dictionaries. IEEE Trans. Image Process. 2006, 15, 3736–3745. [Google Scholar] [CrossRef] [PubMed]
  5. Mairal, J.; Bach, F.; Ponce, J.; Sapiro, G.; Zisserman, A. Non-local sparse models for image restoration. In Proceedings of the 2009 IEEE 12th International Conference on Computer Vision, Kyoto, Japan, 29 September–2 October 2009; pp. 2272–2279. [Google Scholar]
  6. Dong, W.; Zhang, L.; Shi, G.; Li, X. Nonlocally centralized sparse representation for image restoration. IEEE Trans. Image Process. 2013, 22, 1620–1630. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Wen, B.; Ravishankar, S.; Bresler, Y. Structured overcomplete sparsifying transform learning with convergence guarantees and applications. Int. J. Comput. Vis. 2015, 114, 137–167. [Google Scholar] [CrossRef]
  8. Xu, J.; Zhang, L.; Zhang, D. A trilateral weighted sparse coding scheme for real-world image denoising. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018; pp. 20–36. [Google Scholar]
  9. Zoran, D.; Weiss, Y. From learning models of natural image patches to whole image restoration. In Proceedings of the 2011 International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011; pp. 479–486. [Google Scholar]
  10. Xu, J.; Zhang, L.; Zuo, W.; Zhang, D.; Feng, X. Patch group based nonlocal self-similarity prior learning for image denoising. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 244–252. [Google Scholar]
  11. Chen, F.; Zhang, L.; Yu, H. External patch prior guided internal clustering for image denoising. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 603–611. [Google Scholar]
  12. Zuo, W.; Zhang, L.; Song, C.; Zhang, D. Texture enhanced image denoising via gradient histogram preservation. In Proceedings of the 2013 IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013; pp. 1203–1210. [Google Scholar]
  13. Gu, S.; Zhang, L.; Zuo, W.; Feng, X. Weighted nuclear norm minimization with application to image denoising. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 2862–2869. [Google Scholar]
  14. Zhao, Q.; Meng, D.; Xu, Z.; Zuo, W.; Yan, Y. L1-norm low-rank matrix factorization by variational bayesian method. IEEE Trans. Neural Netw. Learn. Syst. 2015, 26, 825–839. [Google Scholar] [CrossRef] [PubMed]
  15. Schmidt, U.; Roth, S. Shrinkage fields for effective image restoration. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 2774–2781. [Google Scholar]
  16. Chen, Y.; Yu, W.; Pock, T. On learning optimized reaction diffusion processes for effective image restoration. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 5261–5269. [Google Scholar]
  17. Zhang, K.; Zuo, W.; Chen, Y.; Meng, D.; Zhang, L. Beyond a gaussian denoiser: Residual learning of deep cnn for image denoising. IEEE Trans. Image Process. 2017, 26, 3142–3155. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  18. Zhang, K.; Zuo, W.; Zhang, L. Ffdnet: Toward a fast and flexible solution for cnn-based image denoising. IEEE Trans. Image Process. 2018, 27, 4608–4622. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  19. Guo, S.; Yan, Z.; Zhang, K.; Zuo, W.; Zhang, L. Toward convolutional blind denoising of real photographs. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 1712–1722. [Google Scholar]
  20. Ma, R.; Zhang, B.; Hu, H. Gaussian pyramid of conditional generative adversarial network for real-world noisy image denoising. Neural Process. Lett. 2020, 51, 2669–2684. [Google Scholar] [CrossRef]
  21. Anwar, S.; Barnes, N. Real image denoising with feature attention. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 3155–3164. [Google Scholar]
  22. Li, D.; Chen, H.; Jin, G.; Jin, Y.; Chen, E. A multiscale dilated residual network for image denoising. Multim. Tools Appl. 2020, 79, 34443–34458. [Google Scholar] [CrossRef]
  23. Binh, P.H.T.; Cruz, C.; Egiazarian, K. Flashlight CNN image denoising. In Proceedings of the IEEE European Signal Processing Conference, Amsterdam, The Netherlands, 18–21 January 2021; pp. 670–674. [Google Scholar]
  24. Quan, Y.; Chen, Y.; Shao, Y.; Teng, H.; Xu, Y.; Ji, H. Image denoising using complex-valued deep CNN. Pattern Recognit. 2021, 111, 107–113. [Google Scholar] [CrossRef]
  25. Guan, X.; Hu, W.; Fu, H. Remote sensing image denoising algorithm with multi-receptive field feature fusion and enhancement. Acta Photonica Sin. 2022, 51, 365–377. [Google Scholar]
  26. Zheng, M.; Zhi, K.; Zeng, J.; Tian, C.; You, L. A hybrid CNN for image denoising. J. Artif. Intell. Technol. 2022, 2, 93–99. [Google Scholar] [CrossRef]
  27. Tian, C.; Zheng, M.; Zuo, W.; Zhang, B.; Zhang, Y.; Zhang, D. Multi-stage image denoising with the wavelet transform. Pattern Recognit. 2023, 134, 109050. [Google Scholar] [CrossRef]
  28. Tang, Y.; Wang, X.; Zhu, J.; Gao, Y.; Jiang, A. LMENet: A lightweight multiscale efficient convolutional neural network for image denoising. In Proceedings of the IEEE Region 10 Conference 2022, Hong Kong, China, 1–4 November 2022; pp. 1–6. [Google Scholar]
  29. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-assisted Intervention 2015, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  30. Gungor, M.A.; Gencol, K. Developing a compression procedure based on the wavelet denoising and jpeg2000 compression. Optik 2020, 218, 164933. [Google Scholar] [CrossRef]
  31. Liu, P.; Zhang, H.; Zhang, K.; Lin, L.; Zuo, W. Multi-level wavelet-CNN for image restoration. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT, USA, 18–22 June 2018; pp. 773–782. [Google Scholar]
  32. Gungor, M.A. A comparative study on wavelet denoising for high noisy ct images of COVID-19 disease. Optik 2021, 235, 166652. [Google Scholar] [CrossRef] [PubMed]
  33. Qian, Y.; Huang, Z.; Fang, H.; Zuo, Z. Wglfnets: Wavelet-based global-local filtering networks for image denoising with structure preservation. Optik 2022, 261, 169089. [Google Scholar] [CrossRef]
  34. Fu, H.; Cheng, J.; Xu, Y.; Wong, D.W.K.; Liu, J.; Cao, X. Joint optic disc and cup segmentation based on multi-label deep network and polar transformation. IEEE Trans. Med. Imaging 2018, 37, 1597–1605. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  35. Liu, Y.; Cheng, M.-M.; Hu, X.; Wang, K.; Bai, X. Richer convolutional features for edge detection. In Proceedings of the 2017 Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 5872–5881. [Google Scholar]
  36. Ioffe, S.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In Proceedings of the 32nd International Conference on International Conference on Machine Learning, Lille, France, 6–11 July 2015; pp. 448–456. [Google Scholar]
  37. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  38. Roth, S.; Black, M.J. Fields of experts: A framework for learning image priors. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Diego, CA, USA, 20–25 June 2005; Volume 2, pp. 860–867. [Google Scholar]
  39. Martin, D.; Fowlkes, C.; Tal, D.; Malik, J. A database of human segmented natural images and its application to evaluating segmentation algorithms and measuring ecological statistics. In Proceedings of the Eighth IEEE International Conference on Computer Vision, Vancouver, BC, Canada, 7–14 July 2001; Volume 2, pp. 416–4232. [Google Scholar]
  40. Franzen, R. Kodak Lossless True Color Image Suite. 1999. Available online: http://r0k.us/graphics/kodak (accessed on 1 January 2022).
  41. Zhang, L.; Wu, X.; Buades, A.; Li, X. Color demosaicking by local directional interpolation and nonlocal adaptive thresholding. J. Electron. Imaging 2011, 20, 23016. [Google Scholar]
Figure 1. Classification of image denoising methods.
Figure 1. Classification of image denoising methods.
Mathematics 11 02772 g001
Figure 2. The 2D Haar wavelet transform.
Figure 2. The 2D Haar wavelet transform.
Mathematics 11 02772 g002
Figure 3. Two-scale 2D Haar wavelet processing results on the image “Monarch”.
Figure 3. Two-scale 2D Haar wavelet processing results on the image “Monarch”.
Mathematics 11 02772 g003
Figure 4. Network structure of the U-shaped network with BN.
Figure 4. Network structure of the U-shaped network with BN.
Mathematics 11 02772 g004
Figure 5. Basic structure of a residual block.
Figure 5. Basic structure of a residual block.
Mathematics 11 02772 g005
Figure 6. Architecture of NSNet.
Figure 6. Architecture of NSNet.
Mathematics 11 02772 g006
Figure 7. Image pyramid constructed by a 2D Haar wavelet.
Figure 7. Image pyramid constructed by a 2D Haar wavelet.
Mathematics 11 02772 g007
Figure 8. Results of some denoising methods at a noise level of 25.
Figure 8. Results of some denoising methods at a noise level of 25.
Mathematics 11 02772 g008
Figure 9. Results of some denoising methods at a noise level of 50.
Figure 9. Results of some denoising methods at a noise level of 50.
Mathematics 11 02772 g009
Figure 10. Comparison of denoising results on one color image of the public dataset CBSD68 [39] at a noise level of 45.
Figure 10. Comparison of denoising results on one color image of the public dataset CBSD68 [39] at a noise level of 45.
Mathematics 11 02772 g010
Figure 11. Comparison of denoising results on one color image of the public dataset CBSD68 [39] at a noise level of 40.
Figure 11. Comparison of denoising results on one color image of the public dataset CBSD68 [39] at a noise level of 40.
Mathematics 11 02772 g011
Table 1. Specific settings of NSNet used in all experiments.
Table 1. Specific settings of NSNet used in all experiments.
ExperimentModelNoise LevelPatch SizeNumber of Patches
Ablation experimentNSNet-S15, 25, 35, 45, 5048 × 48128 × 618
Gray image denoisingNSNet-S15, 25, 5048 × 48128 × 2934
Gray image denoisingNSNet-B(0, 55)48 × 48128 × 2934
Color image denoisingNSNet-B(0, 55)48 × 4896 × 3900
Table 2. Average PSNR/SSIM of four denoising methods in the ablation experiment.
Table 2. Average PSNR/SSIM of four denoising methods in the ablation experiment.
ModelNosie LevelMethod
NSNetUnetNSNet(-BN)Unet(-BN)
Set12 σ = 15 32.90 #/0.9040 #32.75/0.901732.76 */0.9025 *31.81/0.8835
σ = 25 30.50 #/0.8643 #30.48 */0.8640 *30.33/0.861030.28/0.8602
σ = 35 28.95 #/0.8319 #28.89 */0.8300 *28.65/0.824328.77/0.8270
σ = 45 27.81 #/0.8041 #27.80 */0.8029 *27.66/0.799326.52/0.7945
σ = 50 27.32 #/0.7910 #27.32 */0.7901 *26.93/0.773726.90/0.7724
BSD68 σ = 15 31.79 #/0.8927 #31.71 */0.8912 *31.70/0.8916 *31.04/0.8748
σ = 25 29.31 #/0.8322 #29.29 */0.8321 *29.20/0.829429.18/0.8288
σ = 35 27.82 #/0.7849 #27.76 */0.7810 *27.65/0.777927.71/0.7788
σ = 45 26.78 #/0.7446 #26.77 */0.7442 *26.70/0.741326.63/0.7364
σ = 50 26.35 #/0.7267 #26.34 */0.7260 *26.10/0.710126.11/0.7106
Note: NSNet(-BN) means NSNet without BN, and Unet(-BN) means Unet without BN. # The best denoising performance. * The second-best denoising performance.
Table 3. PSNRs of all methods on Set12 at a noise level of 15.
Table 3. PSNRs of all methods on Set12 at a noise level of 15.
ModelC.manHousePappersStarfishMonar.Airpl.ParrotLenaBarbaraBoatManCoupleAverage
NLM30.0533.2331.5930.3030.4729.4230.0733.1631.3331.2531.2730.9731.09
BM3D31.9234.9532.7031.1531.8631.0831.3834.2733.1132.1431.9332.1132.38
KSVD31.4634.2432.2030.7031.4130.7530.9533.71 32.4231.7631.5331.4731.89
NCSR32.0135.0432.6631.5032.2531.19 31.3734.1233.0632.0831.9832.0032.44
OCTOBOS31.9134.3232.4931.0431.7230.9831.2933.9132.5931.8731.7431.7732.14
TWSC32.0135.11 #32.8331.6432.4731.14 31.5234.3933.64 #32.24 32.0932.1532.60
EPLL31.8234.1432.5831.0832.0331.1631.4033.8731.3431.9131.9731.9132.10
GHP31.4834.0732.4031.0931.6330.7731.1633.5432.0131.7231.6231.5431.92
PCLR32.2335.0733.0031.7532.6331.4531.6234.2733.1232.2532.1632.1432.64
PGPD31.8334.7932.6131.2532.1531.1931.3234.0432.7432.0331.9932.0732.33
WNNM32.1835.1532.9731.8332.7231.4031.6134.3833.61 *32.2832.1232.1832.70
CSF31.9534.4032.8331.5632.3431.3431.3634.07 31.9332.0132.0931.9932.32
TNRD32.1934.5533.0331.7632.5731.4731.6334.2532.1432.1532.2432.1132.51
DnCNN-S32.59 *34.9933.24 *32.13 *33.25 *31.67 *31.88 *34.5832.6132.42 *32.43 * 32.43 *32.85 *
DnCNN-B32.1434.9633.0931.9233.0831.5431.6434.5232.0332.36 32.3732.3832.67
FFDnet32.3735.0533.0131.9532.9231.5531.7934.60 *32.4832.3632.3732.43 *32.74
NSNet-S32.67 #35.09 *33.33 #32.29 #33.32 #31.79 #31.97 #34.69 #32.8032.50 #32.48 #32.52 #32.95 #
NSNet-B32.0234.6332.7432.0832.8731.3131.6034.4230.5932.2932.1832.2032.41
Note: # The best denoising performance. * The second-best denoising performance.
Table 4. PSNRs of all methods on Set12 at a noise level of 25.
Table 4. PSNRs of all methods on Set12 at a noise level of 25.
ModelC.manHousePappersStarfishMonar.Airpl.ParrotLenaBarbaraBoatManCoupleAverage
NLM29.9730.3829.0227.8228.0727.3327.9830.3828.5928.7428.8028.3228.61
BM3D29.4532.8630.1628.5629.2528.4328.9332.0830.7229.9129.6229.7229.98
KSVD28.9032.1029.6528.1928.8128.1628.4631.3629.5829.3229.1128.8829.38
NCSR29.4332.8930.0528.7729.4328.4528.8631.9230.6229.7729.5829.4929.94
OCTOBOS29.2532.0829.7828.2428.7828.2828.6731.5629.8829.5129.2629.2329.54
TWSC29.5433.0530.3228.9829.7128.5529.0832.2231.26 #29.9929.7129.7930.18
EPLL29.2432.0430.0728.4329.3028.5628.9131.6228.5529.6929.6329.4829.63
GHP29.2832.5030.0428.6629.0228.2828.8731.6930.2929.7129.4929.3729.77
PCLR29.6732.9830.4628.8729.7528.7729.1332.1730.6530.0029.7729.7330.16
PGPD29.2632.7930.0728.4929.2928.5428.8031.9330.2829.8229.6629.6829.88
WNNM29.6433.2330.4029.0329.8528.7029.1332.2531.24 *30.0329.7729.8230.26
CSF29.4732.4030.2828.8029.6228.7228.8931.8029.0329.7729.7229.5329.84
TNRD29.7132.5430.5529.0229.8628.8929.1832.0129.4129.9229.8829.7130.06
DnCNN-S30.21 # 33.1030.82 *29.3630.41 * 29.08 * 29.44 * 32.4130.0130.2030.0830.0830.43 *
DnCNN-B30.0333.0430.7329.2430.3729.0629.3532.4029.6730.1930.0630.0530.35
FFDnet30.0533.26 *30.7229.2830.2929.0129.4232.57 * 29.9830.23 * 30.07 * 30.15 *30.42
NSNet-S30.12 *33.30 #31.05 #29.90 #30.48 #29.24 #29.45 #32.64 # 30.3930.27 # 30.14 # 30.22 #30.60 #
NSNet-B30.0533.0530.6229.62 *30.3328.9729.3932.57 * 27.5430.23 * 29.9930.0530.20
Note: # The best denoising performance. * The second-best denoising performance.
Table 5. PSNRs of all methods on Set12 at a noise level of 50.
Table 5. PSNRs of all methods on Set12 at a noise level of 50.
ModelC.manHousePappersStarfishMonar.Airpl.ParrotLenaBarbaraBoatManCoupleAverage
NLM24.26 25.69 24.79 23.88 24.16 23.60 24.35 25.97 24.31 24.80 25.06 24.41 24.61
BM3D26.13 29.69 26.68 25.04 25.82 25.10 25.90 29.05 27.23 26.78 26.81 26.46 26.73
KSVD25.68 27.97 26.09 24.53 25.30 24.61 25.38 27.86 25.47 25.95 26.10 25.30 25.85
NCSR26.15 29.62 26.53 25.09 25.77 24.93 25.71 28.90 26.99 26.67 26.67 26.19 26.60
OCTOBOS25.62 28.59 26.15 24.57 25.04 24.86 25.39 28.37 26.17 26.30 26.27 25.82 26.10
TWSC26.46 30.17 26.88 25.41 26.27 25.38 26.11 29.08 27.54 *26.88 26.82 26.48 26.96
EPLL26.03 28.77 26.63 25.04 25.78 25.24 25.84 28.42 24.82 26.65 26.72 26.24 26.35
GHP25.91 28.51 26.38 24.39 25.53 24.76 25.71 27.43 25.44 25.99 25.92 25.46 25.95
PCLR26.56 29.77 27.03 25.32 26.24 25.50 26.15 29.11 27.11 26.99 26.94 26.55 26.94
PGPD26.40 29.73 26.69 25.10 25.89 25.34 25.84 29.00 26.84 26.82 26.84 26.47 26.75
WNNM26.42 30.33 26.91 25.43 26.32 25.42 26.09 29.25 27.79 # 26.97 26.94 26.64 27.04
CSF-------------
TNRD26.61 29.49 27.08 25.42 26.32 25.59 26.16 28.94 25.70 26.94 26.99 26.50 26.81
DnCNN-S27.26 29.96 27.35 25.64 26.83 25.83 26.42 29.34 26.15 27.19 27.19 26.86 27.17
DnCNN-B27.26 29.91 27.35 25.60 26.84 25.82 26.48 29.34 26.32 27.18 27.17 26.87 27.18
FFDnet27.24 30.36 * 27.41 25.68 26.92 25.79 26.57 # 29.63 26.41 27.30 * 27.26 *27.04 *27.30 *
NSNet-S27.38 # 30.43 # 27.54 # 26.24 *26.93 *25.96 # 26.50 29.74 # 27.04 27.38 # 27.31 # 27.14 # 27.47 #
NSNet-B27.35 *30.24 27.47 *26.17 # 26.94 # 25.89 *26.54 *29.70 * 25.46 27.30 * 27.19 27.01 27.27
Note: # The best denoising performance. * The second-best denoising performance.
Table 6. Average PSNRs/SSIMs of all methods on datasets Set12 and BSD68.
Table 6. Average PSNRs/SSIMs of all methods on datasets Set12 and BSD68.
ModelSet 12BSD68
σ = 15 σ = 25 σ = 50 σ = 15 σ = 25 σ = 50
NLM31.09/0.859428.62/0.771124.61/0.569529.82/0.832227.56/0.729624.01/0.5212
BM3D32.38/0.895729.98/0.851026.73/0.768131.08/0.872228.57/0.801725.62/0.6869
KSVD31.89/0.884729.38/0.830825.85/0.726030.86/0.867728.29/0.788925.18/0.6548
NCSR32.44/0.895829.94/0.850126.60/0.767331.19/0.877028.61/0.804525.59/0.6864
OCTOBOS32.14/0.888929.54/0.837826.10/0.743331.08/0.874428.46/0.798925.33/0.6705
TWSC32.60/0.898930.18/0.854926.96/0.773131.28/0.878228.76/0.807725.77/0.6903
EPLL32.10/0.893629.63/0.844426.35/0.747531.19/0.882528.68/0.812325.68/0.6877
GHP31.92/0.869329.77/0.841525.95/0.756230.83/0.851328.49/0.803924.94/0.6809
PCLR32.64/0.897930.16/0.854226.94/0.776331.38/0.879928.84/0.810625.88/0.6970
PGPD32.33/0.890029.88/0.844726.75/0.760231.14/0.870528.64/0.801925.76/0.6877
WNNM32.70/0.898230.26/0.855727.04/0.777531.32/0.876628.80/0.802924.43/0.6838
CSF32.32/0.892329.84/0.8450-/-31.24/0.874628.73/0.8055-/-
TNRD32.51/0.896730.06/0.852026.81/0.766631.42/0.882528.91/0.815525.96/0.7024
DnCNN-S32.85 */0.9025 *30.43 */0.861627.17/0.782831.74 */0.8907 *29.23 */0.827926.24/0.7189
DnCNN-B32.67/0.900030.35/0.859927.18/0.781631.62/0.886829.16/0.824426.23/0.7164
FFDnet32.74/0.902430.42/0.8631 *27.30 */0.7900 *31.64/0.890229.19/0.882826.29/0.7239
NSNet-S32.95 #/0.9054 #30.59 #/0.8662 #27.47 #/0.7956 #31.81 #/0.8936 #29.33 #/0.8339 #26.42 #/0.7316 #
NSNet-B32.41/0.894330.20/0.859527.27/0.789531.46/0.887029.18/0.8304 *26.32 */0.7242 *
Note: # The best denoising performance. * The second-best denoising performance.
Table 7. Average PSNRs/SSIMs of all methods on color images.
Table 7. Average PSNRs/SSIMs of all methods on color images.
DatasetNoise LevelMethod
BM3DDnCNNFFDNetNSNet
CBSD68 σ = 35 28.88/0.816029.60 */0.8422 *29.59/0.840829.69 #/0.8481 #
σ = 45 27.84/0.779328.43/0.8060 *28.44 */0.804828.58 #/0.8150 #
σ = 55 26.97/0.746827.50/0.773627.56 */0.7738 *27.70 #/0.7854 #
Kodak24 σ = 35 29.89/0.820830.45/0.838730.56 */0.8405 *30.67 #/0.8478 #
σ = 45 28.91/0.790629.31/0.805729.44 */0.8083 *29.59 #/0.8187 #
σ = 55 28.06/0.762928.38/0.775328.57 */0.7811 *28.72 #/0.7929 #
McMaster σ = 35 29.93/0.823730.15/0.839230.83 #/0.8550 #30.60 */0.8527 *
σ = 45 29.00/0.798829.08/0.811629.68 #/0.8275 *29.55 */0.8281 #
σ = 55 28.13/0.771228.17/0.783228.76 #/0.8032 *28.73 */0.8049 #
Note: # The best denoising performance. * The second-best denoising performance.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, Y.; Chen, Y. NSNet: An N-Shaped Convolutional Neural Network with Multi-Scale Information for Image Denoising. Mathematics 2023, 11, 2772. https://doi.org/10.3390/math11122772

AMA Style

Li Y, Chen Y. NSNet: An N-Shaped Convolutional Neural Network with Multi-Scale Information for Image Denoising. Mathematics. 2023; 11(12):2772. https://doi.org/10.3390/math11122772

Chicago/Turabian Style

Li, Yifen, and Yuanyang Chen. 2023. "NSNet: An N-Shaped Convolutional Neural Network with Multi-Scale Information for Image Denoising" Mathematics 11, no. 12: 2772. https://doi.org/10.3390/math11122772

APA Style

Li, Y., & Chen, Y. (2023). NSNet: An N-Shaped Convolutional Neural Network with Multi-Scale Information for Image Denoising. Mathematics, 11(12), 2772. https://doi.org/10.3390/math11122772

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop