Next Article in Journal
An ADMM-qSPICE-Based Sparse DOA Estimation Method for MIMO Radar
Next Article in Special Issue
DRL-Based Dynamic Destroy Approaches for Agile-Satellite Mission Planning
Previous Article in Journal
Detection of the Monitoring Window for Pine Wilt Disease Using Multi-Temporal UAV-Based Multispectral Imagery and Machine Learning Algorithms
Previous Article in Special Issue
Dark Spot Detection from SAR Images Based on Superpixel Deeper Graph Convolutional Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

MD3: Model-Driven Deep Remotely Sensed Image Denoising

1
Artificial Intelligence School, Wuchang University of Technology, Wuhan 430223, China
2
School of Electrical and Information Engineering, Wuhan Institute of Technology, Wuhan 430205, China
3
State Key Laboratory of Information Engineering in Surveying, Mapping and Remote Sensing, Wuhan University, Wuhan 430079, China
4
College of Astronautics, Northwestern Polytechnical University (NWPU), Xi’an 710072, China
5
School of Electronic and Information Engineering, Wuhan Donghu University, Wuhan 430212, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(2), 445; https://doi.org/10.3390/rs15020445
Submission received: 30 November 2022 / Revised: 31 December 2022 / Accepted: 2 January 2023 / Published: 11 January 2023
(This article belongs to the Special Issue Reinforcement Learning Algorithm in Remote Sensing)

Abstract

:
Remotely sensed images degraded by additive white Gaussian noise (AWGN) have low-level vision, resulting in a poor analysis of their contents. To reduce AWGN, two types of denoising strategies, sparse-coding-model-based and deep-neural-network-based (DNN), are commonly utilized, which have their respective merits and drawbacks. For example, the former pursue enjoyable performance with a high computational burden, while the latter have powerful capacity in completing a specified task efficiently, but this limits their application range. To combine their merits for improving performance efficiently, this paper proposes a model-driven deep denoising (MD 3 ) scheme. To solve the MD 3 model, we first decomposed it into several subproblems by the alternating direction method of multipliers (ADMM). Then, the denoising subproblems are replaced by different learnable denoisers, which are plugged into the unfolded MD 3 model to efficiently produce a stable solution. Both quantitative and qualitative results validate that the proposed MD 3 approach is effective and efficient, while it has a more powerful ability in generating enjoyable denoising performance and preserving rich textures than other advanced methods.

1. Introduction

Remote sensing imaging technology, based on the theory of electromagnetic waves, is an approach to produce imagery for the visualization of electromagnetic wave information radiated and reflected by long-distance targets [1,2]. Remotely sensed images (RSIs) truly and objectively show the current situation of the distribution of ground objects, the relationship between ground objects or phenomena, and the mutual influences and changes between ground objects [3,4]. Their wide applications, such as real-time monitoring of crop growth, fire detection, geological exploration, dynamic hydrological conditions, and environmental monitoring, have attracted increasing attention. However, they may be degraded by additive white Gaussian noise (AWGN) produced by photo or electronic effects, causing them to have low-level vision, which will cause intelligent systems difficulty in interpreting and analyzing their contents. Assuming an observed RSI f corrupted by AWGN g [5], its degradation process can be easily formulated as [6]
f = x + g ,
where x is a latent clean RSI to be estimated.
To restore the high-level clean RSI x while preserving rich structures from the degraded RSI f for favorably visualizing and understanding its contents, many noise removal methods have been proposed in recent decades and have made considerable progress. One type of traditional denoising method directly uses the spatial statistical information of adjacent pixels to produce approximation results [7]. These approaches are straightforward and easily implemented, but they may generate results with staircase artifacts. To improve the performance and obtain an enjoyable result, Dabov et al. [8] transferred statistical knowledge from adjacent pixels to neighboring patches and proposed a block matching and 3D collaborative filtering (BM3D) scheme. It can efficiently produce a competitive denoising performance, but may over-smooth the fine irregular structures when an image has insignificant self-similarity properties. The other commonly considered denoising strategy is to construct convex optimization mathematical models (COMMs) with the employment of different priors. Such a COMM can be uniformly formulated as
min x 1 2 f x 2 2 + υ Ψ x ,
where the first term · 2 2 is a data fidelity term to guarantee its solution in the degradation process, the second term Ψ · is a regularization prior formed by the properties of the latent image x to ensure its optimal approximation solution, and υ is a weighting parameter to balance the two terms. The most commonly used model is designated as total variation (TV), first proposed by Rudin et al. [9], which penalizes the 1 -norm on the global gradient as a constraint. Then, some other TV-based strategies [10,11,12,13,14,15,16] were proposed to improve the denoising performance with different handcrafted priors to deal with different inverse problems. However, these TV-based methods have some problems: First, their results are sensitive to the selection of a set of parameters and may produce “staircase” results. Second, they may smooth preferable structures due to the piecewise constant assumption. More importantly, these TV-based methods have a heavy computational burden. Figure 1 shows a visual comparison between a model-based scheme and the proposed MD 3 approach, from which we can observe that the denoising performance of the model-based approach is still to be improved and many rich structures in its result are over-smoothed. Unfortunately, it requires more than 780 s of computational time, which is undesirable for its extensive applications.
To improve the efficiency, as well as to have a competitive denoising performance, we further studied the sparse coding model and found that its two subproblems decomposed by the alternating direction method of multipliers (ADMM) [17] are pure AWGN removal problems. As mentioned in [18,19,20,21], such pure AWGN removal problems can be replaced by deep denoisers due to their high flexibility, high efficiency, and powerful modeling capacity. As such, two deep denoiser priors, acting as two independent modulars, are proposed to be plugged into the iterative optimization process of the unfolding sparse COMM to speed up its convergence to a promising approximation solution. The flowchart of the proposed framework is shown in Figure 2.
Our main contributions are summarized as follows:
  • We propose a novel sparse coding denoising strategy, namely model-driven deep denoising (MD 3 ), for the pursuit of pleasing denoising performance efficiently.
  • A learnable iterative soft thresholding algorithm (LISTA) is proposed for pursing the solution of sparse coding coefficient, while a lightweight residual network for learning a dictionary.
  • The quantitative and qualitative results of experiments on both synthetic and real-life remote sensing images validate that the proposed MD 3 approach is effective and even outperforms the state-of-the-art methods.
The remaining parts of our work are organized as follows. Section 2 briefly reviews the previous related work. Section 3 reports the proposed model-driven deep denoising (MD 3 ) strategy and its optimization. Section 4 provides the quantitative and qualitative experimental results. We summarize our work in Section 5.

2. Related Works

In this section, three categories of the previous methods, including sparse-model-based handle-crafted image priors, discriminative learning strategies, and model-guided deep convolutional neural networks (deep unfolding), are briefly introduced.

2.1. Sparse-Model-Based Handle-Crafted Image Priors

Inspired by the performance of BM3D, researchers further studied the neighboring stacked patches having a non-local self-similar (NSS) property that can be further transformed into various forms for model construction. One is low-rank regularization, such as Dong et al. [22], who proposed nonlocal low-rank regularization for compressive sensing. Yang et al. [23] presented a field of expert-regularized nonlocal low-rank matrix approximation (RFoE) denoising models. To make the rank estimation more accurate and obtain a preferable solution, Xue et al. [24] explored two intrinsic priors, including the global correlation across spectrum (GCS) and nonlocal self-similarity (NSS) over space, to avoid tensor rank estimation bias for denoising performance. Liu et al. [25] developed a multigraph-based low-rank tensor approximation for hyperspectral image restoration. Similar to TV problems, the above low-rank matrix factorization (LRMF) problems are also nonconvex optimization problems. There is another direction of research for low-rank matrix approximation: nuclear norm minimization (NNM) [26]. Compared to non-convex LRMF problems, it is the tightest convex relaxation with a certain data fidelity term, so it has attracted great attention and has been extended to many approaches, such as weighted NNM (WNNM) [27], weighted Schatten p-norm minimization (WSNM) [28], and the iterative weighted nuclear norm (IWNN) [29]. These methods assign higher weights to larger singular values, which are flexible for coping with different rank components and generate better approximation solutions than the LRMF problems [30]. Additionally, some methods [31,32,33,34,35] have combined nonlocal low-rank tensor decomposition and total variation regularization for the pursuit of high hyperspectral image denoising performance.
There is also another strategy to represent the non-similar stacked patch, which can be approximated by sparse coding coefficients with a redundant dictionary [36,37,38]. As traditional dictionaries are manually designed and are not flexible in representing the complex image structures, they are commonly learned directly from image data to develop dictionary learning (DL) methods. In sparse DL methods, k-singular-value decomposition (K-SVD) is a well-known approach to alternatively update the sparse coding and dictionary [39]. However, it lacks a shift-invariance property. To cope with this issue, convolutional dictionary learning (CDL) [40,41] replaced matrix multiplication with the convolution operation in signal representation. Inspired by CDL, many CDL-based methods have been proposed and have achieved considerable progress in improving denoising performance [42,43,44,45,46,47,48,49,50]. For instance, convolutional sparse coding (CSC) [42] is proposed to solve the sparse coding problem. Dong et al. [43] proposed a nonlocally centralized sparse representation (NCSR) with the usage of nonlocal self-similar patches for image restoration. Xu et al. [47] developed a trilateral weighted sparse coding (TWSC) strategy by introducing three weight terms into the sparse coding framework. From a multi-scale perspective, an image is separated into high- and low-frequency components, and then, the sparse representation is reconstructed using patch-based structure similarity and singular-value decomposition (SVD) [49]. Ou et al. [50] used multi-scale NSS priors to construct patch groups and proposed a multi-scale weighted group sparse coding model (MS-WGSC) for image denoising with ringing artifact removal while preserving rich details. These methods employ the advantages of the high-order dependency of sparse coefficients for pleasing denoising performance. However, they have their own drawbacks: the weak model representation flexibility of the universal dictionary, several regularization parameters that are required to be manually set, and even a heavy computational time burden.

2.2. Deep Neural Network

In recent years, deep neural network (DNN) approaches have been popularly developed to directly learn a nonlinear mapping function from the space of noisy images to that of clean images [51]. With the employment of its strong learning capacity on large training samples, many DNN-based denoising methods have been proposed [18,19,20,21,52,53,54,55,56,57,58,59,60,61,62,63]. For example, Chen et al. [52] used a loss-based scheme to learn filters from training data and formed a trainable nonlinear reaction diffusion (TNRD) model. Zhang et al. [18] proposed a classical deep image denoising strategy, which is named the deep convolutional neural network (DnCNN), and then proposed a fast and flexible solution for CNN-based image denoising (FFDNet) [21]. Guo et al. [53] presented a convolutional blind denoising network (CBDNet) by incorporating the network architecture, noise modeling, and asymmetric learning to improve the robustness and practicability of deep denoising models. Jia et al. [54] developed a fractional optimal control network (FOCNet) for image denoising. Tian et al. [55] integrated batch renormalization into a deep CNN (DCNN) to obtain good results. Zhang et al. [56] used a memory-efficient hierarchical neural architecture to search for pleasing solutions. COLA-Net [57] and SQAD [58] introduced an attention mechanism into the DNN for denoising performance improvement. The residual network was further extended by Zhang et al. [59] to a residual dense network. Non-local self-similarity [60] and the low rank [25] of stacked patches are both plugged into a convolutional neural network for noise removal. Jia et al. [61], Dong et al. [62], and Xu et al. [63] reduced noise using deep learning technologies from a multi-scale perspective.
The above methods successfully and efficiently achieved leading denoising performance due to their powerful learning capabilities and surpassed the traditional image denoising methods by learning the image denoisers from the training data. However, these DNN-based methods lack good interpretability because they directly map low-quality images to latent noise-free results with a black-box nature, and their application range is greatly limited because they are usually applied to a specialized task.

2.3. Deep Unfolding

To address the above issues of DNN-based methods, deep learning technologies are usually exploited to complete specialized tasks (e.g., denoisers) in the optimization process of sparsity-based models, which can be decoupled by certain algorithms (such as the alternating direction method of multipliers (ADMM) [17] and half-quadratic splitting (HQS) [64]). For instance, Meinhardt et al. [65] replaced the proximal operator of regularization with a denoising neural network. The methods in [66,67,68] unfolded the CSC process and integrated DNN into the CSC problem for an efficient solution. Simon et al. [69] further improved it by using strided convolution. Dong et al. [70] unfolded the iterative process into a feed-forward neural network. Fu et al. [71] proposed a multi-scale feature extraction module for reducing JPEG artifacts. Bertocchi et al. [72] provided a neural network architecture by unfolding a proximal interior point algorithm. Zheng et al. [73] proposed a deep convolutional dictionary learning (DCDicL) framework to learn the priors for both representation coefficients and dictionaries. Sun et al. [74] used low-rank representation and a CNN denoiser (LRR-CNN) to form a hyperspectral image denoising model. Huang et al. [75] proposed a nonlocal self-similar (NSS) block-based deep image denoising scheme, designated the deep low-rank prior (DLRP), to achieve efficient performance. Xu et al. [76] developed an end-to-end deep architecture to follow the process of sparse-representation-based image restoration.
These unfolding iterative denoising strategies successfully improved their interpretability when compared to DNN-based methods. However, the existing deep unfolding methods also have drawbacks as follows: (1) the learning capacity of DNNs is wasted with the usage of fixed priors (without learning the priors from data). (2) The dictionary is universally learned and is not adaptive, resulting in their performance being even lower than that of DNN-based approaches.

3. Model-Driven Deep Denoising

3.1. MD 3 Model Generation

For an exemplar patch of size n × n at position i denoted by x i , its similar neighborhood patches are ordered lexicographically to construct a data matrix:
X i = R i x = x i 1 , x i 2 , , x i m , X i Ω n × m ,
where x i j ( j 1 , m ) is a column vector similar to x i and m is the total number of similar patches that are selected by the following criteria:
Z i = i j | ρ i j > η ,
where Z i is the position set of similar patches and η is a judgment parameter (which is empirically set as 0.4 in the following experiments), and ρ i j , the correlation coefficient between the i-th patch and the j-th patch, is computed by
ρ i j = c o v ( x i , x j ) σ x i σ x j = E ( x i x j ) E ( x i ) E ( x j ) E ( x i 2 ) E 2 ( x i ) E ( x j 2 ) E 2 ( x j ) ,
where E ( · ) is a function to compute the mean value of a matrix. Generally, this constructed data matrix X i is sparse and can be characterized by a coding coefficient vector α i with a given or learned dictionary D [75]. That is, X i = R i x = D α i . Compared to the usage of a given dictionary, a learned dictionary is more flexible to be able to adaptively interpret complex structures. By considering the above constraints and the priors on the unknown variables, we can construct a global objective model as follows:
min x , D , α i 1 2 f x 2 2 + i Γ x , D , α i ,
where Γ x , D , α i = λ 2 R i x D α i 2 2 + β Φ D + γ Ψ α i , λ is a penalty parameter (which has a direct relationship with the noise level σ ), while β and γ are regularization parameters to guarantee that the model has an optimized solution.

3.2. MD 3 Model Optimization

The objective function (6) can be decomposed into the following two subproblems and be solved by alternately minimizing them for the latent clean image x, the sparse representation matrix α i , and dictionary D while keeping the other variables fixed:
{ min x 1 2 f x 2 2 + i λ 2 R i x D α i 2 2 , 7 a min D , α i λ 2 R i x D α i 2 2 + β Φ D + γ Ψ α i . 7 b

3.2.1. Solving x

In the ( k + 1 ) -th iteration, Sub-problem (7a) for x is a quadratic optimization problem and has a fast closed-form solution:
x = I + λ i R i T R i 1 f + λ i R i T D k α i k ,
where I is a unit matrix and R i T is the transpose matrix of R i .

3.2.2. Solving α i and D

By the usage of the alternating direction method of multipliers (ADMM), Sub-problem (7b) can be further separated into two independent subproblems as follows:
{ min α i λ 2 R i x D α i 2 2 + γ Ψ α i , 9 a min D λ 2 R i x D α i 2 2 + β Φ D . 9 b
where the sparse coding α i is estimated by Equation (9a) from X i = R i x with a fixed dictionary D. Alternately, dictionary D is computed by Equation (9b) from X i = R i x with a known sparse coding α i :
  • For sparse coding α i : Equation (9a) is a restoration task for α i from X i = R i x , which can be rewritten as
    min α i 1 2 X i D α i 2 2 + γ λ Ψ α i .
The traditional strategies to construct the sparse coding model usually plug the handle-crafted priors into Problem (10). Specifically, Ψ α i = α i 0 is an 0 -norm, which is an original sparse approximation problem to count the number of non-zero elements in α i . Unfortunately, the 0 -norm is an N-P problem that is difficult to solve. An 1 -norm ( Ψ α i = α i 1 ) is usually employed to replace the 0 -norm to form a convex and continuous problem, with which Problem (10) is equivalently written as
min α i 1 2 X i D α i 2 2 + γ λ α i 1 ,
which can be effectively solved by a popularly representative approach, the iterative soft thresholding algorithm (ISTA) [77], to guarantee its convergence. The solution of the iterative process at the k-th iteration using the ISTA is
α i 0 = 0 α i k + 1 = S θ α i k + 1 τ D T X i D α i k
where θ = γ / λ τ and τ is a parameter that should meet the following criterion: τ > υ m a x , where υ m a x is the largest eigenvalue of D T D . S θ is an elementwise soft-thresholding shrinkage function, which is defined as
S θ κ = s i g n κ max κ θ , 0
to search for the global minimum of the sparse representation model (10). However, this global minimum solution is dependent on the selection of parameters γ and λ . Moreover, the obtained solution of the 1 -norm is theoretically approximated to the desired solution of the 0 -norm, but it is not an exact solution and depends on its sparsity and the properties of dictionary D. Actually, the iterative process in Equation (11) can be equivalently written as
α i k + 1 = S θ I 1 τ D T D α i k + 1 τ D T X i
Let P = I 1 τ D T D and M = 1 τ D T , then Equation (14) can be simplified as
α i k + 1 = S θ P k α i k + M k X i k + 1
which can be implemented by a learnable ISTA (LISTA) encoder network to obtain a fast sparse coding approximation solution of Equation (10). The LISTA encoder network is viewed as a modular part to be plugged into the denoising model, where the thresholding parameter θ is learned for the pursuit of the sparse coding α i , which is then fixed for solving other variables. The framework of the LISTA encoder for Problem (15) is shown in Figure 3.
  • For dictionary D: with the aid of the half-quadratic splitting (HQS) algorithm due to its simplicity and fast convergence in many applications. Equation (9b) can be solved by inducing an auxiliary variable Θ D to convert the constrained problem (9b) into an unconstrained one:
    min D , Θ D λ 2 R i x D α i 2 2 + β Φ Θ D + μ 2 Θ D D 2 2 .
By employing ADMM, Equation (16) is decomposed into the following two independent problems:
{ D ^ = arg min D λ 2 R i x D α i 2 2 + μ 2 Θ D D 2 2 17 a Θ ^ D = arg min Θ D β Φ Θ D + μ 2 Θ D D 2 2 17 b
Similar to solving x in Equation (7a), Equation (17a) also has a closed-form solution:
D k + 1 = μ Θ D k + λ X i k + 1 α i k + 1 T μ I + λ α i k + 1 α i k + 1 T 1
The subproblem (17b) can be further modified as
Θ ^ D = arg min Θ D Φ Θ D + 1 2 β β μ μ 2 Θ D D 2 2 ,
which is a common Gaussian noise removal issue on Θ D with a noise level β β μ μ . For simplicity, it is replaced by a learned deep Gaussian denoiser (DGD) as follows:
Θ D k + 1 = D G D D k + 1 , β β μ μ .
Compared to the traditional approaches with handle-crafted priors Φ · , the DGD has several advantages. First, the unknown image prior Φ · can be implicitly replaced by any Gaussian denoiser. Second, the learned DGD can be jointly utilized as a modular part to efficiently solve many inverse problems due to its high flexibility and powerful modeling capacity.
Figure 4 shows the architecture of the deep dictionary network (DDNet) Φ · . As D has a lighter structure derived from a deep residual network [78], we designed a simple network whose receptive field and learning capability are sufficient to generate a pleasing result. As shown in Figure 4, DDNet has six layers, including five “DConv + PReLU” layers and one “DConv” layer. DConv preserves the merits of 3 × 3 while enlarging the receptive field [79]. PReLU is a parameterized nonlinearity function to generate a high-quality estimation with few filters [80]. The number of channels in each layer is 16.

4. Experimental Results and Analysis

4.1. Experimental Preparation

In this section, we present the experiments conducted on both synthetic and real-life RSIs by state-of-the-art methods to obtain quantitative and qualitative results, which were employed to validate the effectiveness of the proposed MD 3 . All comparable approaches were tested on a PC with an Intel Core i7-5960X CPU 3.0 GHz 16.0 GB memory and a GeForce RTX 2080Ti GPU, and the MATLAB 2014b software was used for the model-based methods, while the TensorFlow software package was exploited for the deep learning methods. All training and testing images were downloaded from [81,82], which were randomly cropped into 735,602 patches of size 128 × 128 with a ratio of 7:3. Note that hyperspectral images were treated as common RSI images in a band-by-band manner to be proceeded. The Adam optimizer [83,84] was used to minimize the L 1 loss function, which was adopted to generate the optimization network parameters shared in each stage. The block size was 8 × 8 , while the dictionary size was 256 × 256 . The learning rate started from 1e-4, decayed by half every 1e5 iterations, and finally, ended after 100 epochs. The noise level ranged from 1 to 50 with a step size of 1, which is enough for training a set of deep denoisers. The iterative optimization process is finished when the following criterion is satisfied: x k + 1 x k 2 2 x k 2 2 < 10 3 . Eight selected test images (as shown in Figure 5) from different datasets were experimented on and presented in this section to verify the effectiveness of the proposed MD 3 strategy. For similarity, the eight test images were named MODIS-A1, MODIS-A2, MODIS-A3, MODIS-T1, MODIS-T2, MODIS-T3, Hyper-WDCM, and Multi-WDC.
Parameters setting: The unknown parameters referred to in Section 3 include four regularization factors { λ , μ , γ , β } and the iteration number K. In order to make the proposed MD 3 method image adaptive, a HyperNet module, taking noise level σ as the inputs, was exploited to learn the four hyper-parameters { λ k , μ k , γ k , β k } for each stage. There are two Conv layers with kernel size 1 and a SoftPlus layer to allow regularization parameters to be positive in the HyperNet module. For noise level σ , the initial value was known in the testing experiments, but it was unknown at other stages and in the experiments on real-life RSIs. To address this issue, we estimated it at each stage by the method presented in [85] to ensure the alternating iterations adaptively converged to a fixed point. The initial D was the DCT dictionary and α 0 = D 0 1 R x 0 .
Quantitative index selection: Except for subjective evaluation, we, in the simulated experiments, employed the peak-signal-to-noise ratio (PSNR) and structural similarity index measure (SSIM) to objectively assess the capacity of the state-of-the-art in AWGN removal and structural preservation, respectively [73]. In real-world experiments, two reference-free metrics, the Q-metric (QM) [4] and natural image quality evaluator (NIQE) [86], were used to evaluate their ability to preserve fine structures and improve estimated image quality, respectively. The higher the PSNR is, the better the AWGN noise suppression. The larger the SSIM and the QM are, the richer the structures. The smaller the NIQE is, the better the denoised image quality.
Comparison approaches’ selection: To check the effectiveness of the proposed MD 3 scheme and its advantages, three types of methods described in Section 2, model-based (NCSR [43] and WSSR [6]), DNN-based ([18]), and deep-unfolding-based (DKSVD [68], DCDicL [73], LRR-CNN [74], and DLRP [75]), were selected to be utilized, and their results were compared regarding quantization and qualification.

4.2. Analysis of Convergence and Intermediate Results

Figure 6 presents the intermediate results generated by the proposed MD 3 method and its convergence on the Hyper-WDCM image with Noise Level 20. With the aid of two learned strategies (including the LISTA encoder and DDNet), the noise in the estimated image (as shown in Figure 6b–d) was more effectively suppressed with the increment of the iterations, while the atoms in the adaptive estimated dictionary (as shown in Figure 6f–h) were much closer to the units of the ground-truth image. The convergence result in Figure 6e shows that the MD 3 method quickly converged to a fixed point with only five iterations, indicating that the two powerful learning modules explored in the optimization process of our method can effectively help it speed up to an enjoyable solution.

4.3. Experiments on the Synthetic RSI Images

The eight synthetic RSI images shown in Figure 5 were simulated and tested to examine the effectiveness of the MD 3 method. Three respective visual comparisons with different noise levels are presented in Figure 7, Figure 8 and Figure 9. By observing these results, we found that NCSR produced the worst maps with distorted structures. WSSR pursues better results with richer textures than the NCSR’s images at the cost of a z huge time burden. DnCNN directly removes noise with the usage of a learned black-block mapping function and produces enjoyable results; however, its interpretability is ambiguous. The deep unfolding methods (DKSVD, DCDicL, LRR-CNN, and DLRP) enhance the interpretability of the DNN and assign its specialized task, but their results are over-smoothed, which are even worse than those produced by DnCNN. The reason may be that there are too many regularization parameters that are not adaptive. In contrast, the details generated by the proposed MD 3 method are better highlighted and are closer to those of the ground-truth images.
The objective values of PSNR and SSIM are shown in Table 1 and Table 2, respectively, from which some observations can be made. First, MD 3 obtains an enjoyable PSNR and SSIM on all RSI images at each noise level, indicating that MD 3 can yield competitive denoising performance and preferable results with plentiful structures. Second, MD 3 generates the best average PSNR, which outperforms the second-best method (deep unfolding DLRP) from 0.18 dB to 0.39 dB, verifying that MD 3 is effective at noise removal. Such conclusions are consistent with the visual comparisons.

4.4. Experiments on Real-World RSI Images

To further examine its advantages, we also applied the state-of-the-art to real-world RSI images (as shown in Figure 10), which are also related in [6]. Their denoised results are compared in Figure 11 and Figure 12, in which we can observe that the details in the results produced by MD 3 are much richer than those generated by the competing methods (the readers can visualize the enlarged parts in these figures by themselves). Meanwhile, the quantitative comparisons are presented in Table 3. For each real-life RSI image, MD 3 achieves the highest QM value, further demonstrating that MD 3 performs extraordinarily well on preferable detail preservation. This is because MD 3 perceives the global information from the input images with an adaptive learning dictionary, making it image-adaptive. More importantly, each result denoised by MD 3 has the smallest NIQE value, validating that its quality is the most comfortable for human visualization.

4.5. Computational Complexity

Efficiency is also an important index to evaluate the possibility for extensive applications. As model-based methods have a high time burden (CPU time), which may be 100-times more than those produced by deep-learning-based approaches, they were not selected for efficiency assessment. Figure 13 shows a testing example on the MODIS-A1 image with size 512 × 512 , from which we can see that MD 3 is slower than DnCNN, DKSVD, LRR-CNN, and DLRP, but is faster than DCDicL with a complex solution process. However, their denoising performance lags behind that of MD 3 .
In sum, taking all comparisons into account, MD 3 with two deep learning modulars is image-adaptive, can effectively and efficiently obtain a stable solution, and is suitable for extensive applications.

5. Conclusions

In the RSI image denoising task, there are three well-known key points, denoising performance, structure preservation, and efficiency, that should all be considered in designing a good interpretability strategy. To address these issues, this paper reported a deep unfolding scheme, namely model-driven deep denoising (MD 3 ), including the following: (1) The MD 3 model was constructed with the rearranged self-similar data matrix, having sparse coding and a learned dictionary. (2) The MD 3 model was separated into two independent subproblems, including the data fidelity term (which has a closed-form solution) for the latent clean image and the regularization term for the sparse coding and dictionary, by using the alternating direction method of multipliers (ADMM). (3) With further separation of the regularization term, the regularization problem was decomposed into two independent subproblems, respectively, for the sparse coding and dictionary. The sparse coding problem was solved by the proposed learnable ISTA (LISTA) encoder network, while the dictionary was learned by a lightweight residual network (namely the deep dictionary network (DDNet)) for adaptive images. (4) The two deep neural networks, acting as two modulars, were plugged into the iterative procedure to speed up the optimization process for a fixed solution. Finally, the proposed MD 3 , as well as the state-of-the-art were evaluated on both synthetic and real-life RSI images, for which the comparison results for the quantitation and qualification validated that it is effective at improving the denoising performance and preserving rich structures. Its satisfactory computation time is beneficial for its extensive application.

Author Contributions

Z.H.: investigation, writing—original draft. Z.Z. and Y.Z.: software. Z.W. and B.X.: visualization, investigation. J.L. and S.L.: writing—review and editing. H.F.: conceptualization, methodology. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China under Grant No. 61901309.

Data Availability Statement

Not applicable.

Acknowledgments

We appreciate the critical and constructive comments and suggestions from the Reviewers, which helped improve the quality of this manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Li, Y.; Zhang, Y.; Zhu, Z. Error-tolerant deep learning for remote sensing image scene classification. IEEE Trans. Cybern. 2021, 51, 1756–1768. [Google Scholar] [CrossRef] [PubMed]
  2. Li, Y.; Zhang, Y. A new paradigm of remote sensing image interpretation by coupling knowledge graph and deep learning. Geomatics Inf. Sci. Wuhan Univ. 2022, 47, 1176–1190. [Google Scholar] [CrossRef]
  3. Huang, Z.; Zhang, Y.; Li, Q.; Zhang, T.; Sang, N.; Hong, H. Progressive dual-domain filter for enhancing and denoising optical remote sensing images. IEEE Geosci. Remote Sens. Lett. 2018, 15, 759–763. [Google Scholar] [CrossRef]
  4. Chang, Y.; Chen, M.; Yan, L.; Zhao, X.; Li, Y.; Zhong, S. Toward universal stripe removal via wavelet-based deep convolutional neural network. IEEE Trans. Geosci. Remote Sens. 2020, 58, 2880–2897. [Google Scholar] [CrossRef]
  5. Rasti, B.; Chang, Y.; Dalsasso, E.; Denis, L.; Ghamisi, P. Image restoration for remote sensing: Overview and toolbox. arXiv 2021, arXiv:2107.00557. [Google Scholar] [CrossRef]
  6. Huang, Z.; Zhang, Y.; Li, Q.; Li, X.; Zhang, T.; Sang, N.; Hong, H. Joint analysis and weighted synthesis sparsity priors for simultaneous denoising and destriping optical remote sensing images. IEEE Trans. Geosci. Remote Sens. 2020, 58, 6958–6982. [Google Scholar] [CrossRef]
  7. Buades, A.; Coll, B.; Morel, J. A non-local algorithm for image denoising. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), San Diego, CA, USA, 20–25 June 2005; Volume 2, pp. 60–65. [Google Scholar]
  8. Dabov, K.; Foi, A.; Katkovnik, V.; Egiazarian, K. Image denoising by sparse 3-D transform-domain collaborative filtering. IEEE Trans. Image Process. 2007, 16, 2080–2095. [Google Scholar] [CrossRef]
  9. Rudin, L.; Osher, S.; Fatemi, E. Nonlinear total variation based noise removal algorithms. Physica D 2007, 60, 259–268. [Google Scholar] [CrossRef]
  10. Bougleux, S.; Peyré, G.; Cohen, L. Non-local regularization of inverse problems. Inverse Probl. Imaging 2011, 5, 511–530. [Google Scholar]
  11. Zuo, W.; Zhang, L.; Song, C.; Zhang, D. Texture enhanced image denoising via gradient histogram preservation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Portland, OR, USA, 23–28 June 2013. [Google Scholar]
  12. Zhu, H.; Cui, C.; Deng, L.; Cheung, R.; Yan, H. Elastic Net Constraint based Tensor Model for High-order Graph Matching. IEEE Trans. Cybern. 2020, 51, 4062–4074. [Google Scholar] [CrossRef]
  13. Li, Z.; Malgouyres, F.; Zeng, T. Regularized non-local total variation and application in image restoration. J. Math. Imaging Vis. 2017, 59, 296–317. [Google Scholar] [CrossRef]
  14. Zhang, H.; Tang, L.; Fang, Z.; Xiang, C.; Li, C. Nonconvex and nonsmooth total generalized variation model for image restoration. Signal Process. 2018, 143, 69–85. [Google Scholar] [CrossRef]
  15. Liu, J.; Ma, R.; Zeng, X.; Liu, W.; Wang, M.; Chen, H. An efficient non-convex total variation approach for image deblurring and denoising. Appl. Math. Comput. 2021, 397, 125977. [Google Scholar] [CrossRef]
  16. Shen, Y.; Liu, Q.; Lou, S.; Hou, Y. Wavelet-based total variation and nonlocal similarity model for image denoising. IEEE Signal Process. Lett. 2017, 24, 877–881. [Google Scholar] [CrossRef]
  17. Boyd, S.; Parikh, N.; Chu, E. Distributed optimization and statistical learning via the alternating direction method of multipliers. Found. Trends® Mach. Learn. 2011, 3, 1–122. [Google Scholar]
  18. Zhang, K.; Zuo, W.; Chen, Y.; Meng, D.; Zhang, L. Beyond a Gaussian denoiser: Residual learning of deep CNN for image denoising. IEEE Trans. Image Process. 2017, 26, 3142–3155. [Google Scholar] [CrossRef] [Green Version]
  19. Zhu, H.; Qiao, Y.; Xu, G.; Deng, L.; Yu, F. DSPNet: A lightweight dilated convolution neural networks for spectral deconvolution with self-paced learning. IEEE Trans. Ind. Inform. 2019, 16, 7392–7401. [Google Scholar] [CrossRef]
  20. Li, Y.; Zhang, Y.; Huang, X.; Zhu, H.; Ma, J. Large-scale remote sensing image retrieval by deep hashing neural networks. IEEE Trans. Geosci. Remote Sens. 2018, 56, 950–965. [Google Scholar] [CrossRef]
  21. Zhang, K.; Zuo, W.; Zhang, L. FFDNet: Toward a fast and flexible solution for CNN-based image denoising. IEEE Trans. Image Process. 2018, 27, 4608–4622. [Google Scholar] [CrossRef] [Green Version]
  22. Dong, W.; Shi, G.; Li, X.; Ma, Y.; Huang, F. Compressive sensing via nonlocal low-rank regularization. IEEE Trans. Image Process. 2014, 23, 3618–3632. [Google Scholar] [CrossRef]
  23. Yang, H.; Lu, J.; Zhang, H.; Luo, Y.; Lu, J. Field of experts regularized nonlocal low rank matrix approximation for image denoising. J. Comput. Appl. Math. 2022, 412, 114244. [Google Scholar] [CrossRef]
  24. Xue, J.; Zhao, Y.; Liao, W.; Chan, C. Nonlocal low-rank regularized tensor decomposition for hyperspectral image denoising. IEEE Trans. Geosci. Remote Sens. 2019, 57, 5174–5189. [Google Scholar] [CrossRef]
  25. Liu, N.; Li, W.; Tao, R.; Du, Q.; Chanussot, J. Multi-graph-based low-rank tensor approximation for Hyperspectral image restoration. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5530314. [Google Scholar]
  26. Donoho, D.; Gavish, M.; Montanari, A. The phase transition of matrix recovery from gaussian measurements matches the minimax mse of matrix denoising. Proc. Natl. Acad. Sci. USA 2013, 110, 8405–8410. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  27. Gu, S.; Zhang, L.; Zuo, W.; Feng, X. Weighted nuclear norm minimization with application to image denoising. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 2862–2869. [Google Scholar]
  28. Xie, Y.; Gu, S.; Liu, Y.; Zuo, W.; Zhang, W.; Zhang, L. Weighted schatten p-norm minimization for image denoising and background subtraction. IEEE Trans. Image Process. 2016, 25, 4842–4857. [Google Scholar] [CrossRef] [Green Version]
  29. Huang, Z.; Li, Q.; Fang, H.; Zhang, T.; Sang, N. Iterative weighted nuclear norm for X-ray angiogram image denoising. Signal Image Video Process. 2017, 11, 1445–1452. [Google Scholar] [CrossRef]
  30. Huang, T.; Dong, W.; Xie, X.; Shi, G.; Bai, X. Mixed noise removal via laplacian scale mixture modeling and nonlocal low-rank approximation. IEEE Trans. Image Process. 2017, 26, 3171–3186. [Google Scholar] [CrossRef]
  31. Zhang, H.; Liu, L.; He, W.; Zhang, L. Hyperspectral image denoising with total variation regularization and nonlocal low-rank tensor decomposition. IEEE Trans. Geosci. Remote Sens. 2020, 58, 3071–3084. [Google Scholar] [CrossRef]
  32. Zhu, H.; Ni, H.; Liu, S.; Xu, G.; Deng, L. Tnlrs: Target-aware non-local low-rank modeling with saliency filtering regularization for infrared small target detection. IEEE Trans. Image Process. 2020, 29, 9546–9558. [Google Scholar] [CrossRef]
  33. Li, Y.; Chen, W.; Huang, X.; Gao, Z.; Li, S.; He, T.; Zhang, Y. Mfvnet: Deep adaptive fusion network with multiple field-of-views for remote sensing image semantic segmentation. Sci. China Inf. Sci. 2022. [Google Scholar] [CrossRef]
  34. Zhu, H.; Zhang, J.; Xu, G.; Deng, L. Tensor field graph-cut for image segmentation: A non-convex perspective. IEEE Trans. Circuits Syst. Video Technol. 2020, 31, 1103–1113. [Google Scholar] [CrossRef]
  35. Kong, X.; Zhao, Y.; Xue, J.; Chan, J.C.W.; Zang, J. Hyperspectral image denoising based on nonlocal low-rank and TV regularization. Remote Sens. 2020, 12, 1956. [Google Scholar] [CrossRef]
  36. Li, Y.; Kong, D.; Zhang, Y.; Tan, Y.; Chen, L. Robust deep alignment network with remote sensing knowledge graph for zero-shot and generalized zero-shot remote sensing image scene classification. ISPRS J. Photogramm. Remote Sens. 2021, 179, 145–158. [Google Scholar] [CrossRef]
  37. Zhu, H.; Liu, S.; Deng, L.; Li, Y.; Xiao, F. Infrared Small Target Detection via Low Rank Tensor Completion with Top-Hat Regularization. IEEE Trans. Geosci. Remote Sens. 2020, 58, 1004–1016. [Google Scholar] [CrossRef]
  38. Li, Y.; Zhou, Y.; Zhang, Y.; Zhong, L.; Wang, J.; Chen, J. DKDFN: Domain knowledge-guided deep collaborative fusion network for multimodal unitemporal remote sensing land cover classification. ISPRS J. Photogramm. Remote Sens. 2022, 186, 170–189. [Google Scholar] [CrossRef]
  39. Elad, M.; Aharon, M. Image denoising via sparse and redundant representation over learned dictionaries. IEEE Trans. Image Process. 2006, 15, 3736–3745. [Google Scholar] [CrossRef] [PubMed]
  40. Bristow, H.; Eriksson, A.; Lucey, S. Fast convolutional sparse coding. In Proceedings of the 2013 IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013; pp. 391–398. [Google Scholar]
  41. Cardona, C.G.; Wohlberg, B. Convolutional dictionary learning: A comparative review and new algorithms. IEEE Trans. Comput. Imaging 2018, 4, 366–381. [Google Scholar] [CrossRef] [Green Version]
  42. Rey-Otero, I.; Sulam, J.; Elad, M. Variations on the CSC model. IEEE Trans. Image Process. 2020, 68, 519–528. [Google Scholar] [CrossRef]
  43. Dong, W.; Zhang, L.; Shi, G.; Li, X. Nonlocally centralized sparse representation for image restoration. IEEE Trans. Image Process. 2013, 22, 1620–1630. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  44. Huang, Z.; Li, Q.; Zhang, T.; Sang, N.; Hong, H. Iterative weighted sparse representation for X-ray cardiovascular angiogram image denoising over learned dictionary. IET Image Process. 2018, 12, 254–261. [Google Scholar] [CrossRef]
  45. Li, Y.; Chen, W.; Zhang, Y.; Tao, C.; Xiao, R.; Tan, Y. Accurate cloud detection in high-resolution remote sensing imagery by weakly supervised deep learning. Remote Sens. Environ. 2020, 250, 112045. [Google Scholar] [CrossRef]
  46. Zhu, H.; Peng, H.; Xu, G.; Deng, L.; Cheng, Y.; Song, A. Bilateral weighted regression ranking model with spatial-temporal correlation filter for visual tracking. IEEE Trans. Multimed. 2021, 24, 2098–2111. [Google Scholar] [CrossRef]
  47. Xu, J.; Zhang, L.; Zhang, D. A trilateral weighted sparse coding scheme for real-world image denoising. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 21–38. [Google Scholar]
  48. Peng, C.; Liu, Y.; Chen, Y.; Wu, X.; Cheng, A.; Kang, Z.; Chen, C.; Cheng, Q. Hyperspectral image denoising using non-convex local low-rank and sparse separation with spatial-spectral total variation regularization. arXiv 2022, arXiv:2201.02812. [Google Scholar]
  49. Shi, M.; Zhang, F.; Wang, S.; Zhang, C.; Li, X. Detail preserving image denoising with patch-based structure similarity via sparse representation and svd. Comput. Vis. Image Underst. 2021, 206, 103173. [Google Scholar]
  50. Ou, Y.; Swamy, M.N.S.; Luo, J.; Li, B. Single image denoising via multi-scale weighted group sparse coding. Signal Process. 2022, 200, 108650. [Google Scholar] [CrossRef]
  51. Sun, L.; Dong, W.; Li, X.; Wu, J.; Li, L.; Shi, G. Deep maximum a posterior estimator for video denoising. Int. J. Comput. Vis. 2021, 129, 2827–2845. [Google Scholar] [CrossRef]
  52. Chen, Y.; Pock, T. Trainable nonlinear reaction diffusion: A flexible framework for fast and effective image restoration. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1256–1272. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  53. Guo, S.; Yan, Z.; Zhang, K.; Zuo, W.; Zhang, L. Toward convolutional blind denoising of real photographs. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 1712–1722. [Google Scholar]
  54. Jia, X.; Liu, S.; Feng, X.; Zhang, L. FOCNet: A fractional optimal control network for image denoising. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 6054–6063. [Google Scholar]
  55. Tian, C.; Xu, Y.; Zuo, W. Image denoising using deep cnn with batch renormalization. Neural Netw. 2017, 121, 461–473. [Google Scholar] [CrossRef]
  56. Zhang, H.; Li, Y.; Chen, H.; Shen, C. Memory-efficient hierarchical neural architecture search for image denoising. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 3657–3666. [Google Scholar]
  57. Mou, C.; Zhang, J.; Fan, X.; Liu, H.; Wang, R. COLA-Net: Collaborative attention network for image restoration. IEEE Trans. Multimed. 2021, 24, 1366–1377. [Google Scholar] [CrossRef]
  58. Pan, E.; Ma, Y.; Mei, X.; Fan, F.; Huang, J.; Ma, J. SQAD: Spatial-spectral quasi-attention recurrent network for hyperspectral image denoising. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–14. [Google Scholar] [CrossRef]
  59. Zhang, Y.; Tian, Y.; Kong, Y.; Zhong, B.; Fu, Y. Residual dense network for image restoration. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 2480–2495. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  60. Wang, Z.; Ng, M.K.; Zhuang, L.; Gao, L.; Zhang, B. Nonlocal self-similarity-based hyperspectral remote sensing image denoising with 3-D convolutional neural network. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–17. [Google Scholar] [CrossRef]
  61. Jia, X.; Peng, Y.; Ge, B.; Li, J.; Liu, S.; Wang, W. A multi-scale dilated residual convolution network for image denoising. Neural Process. Lett. 2022. [Google Scholar] [CrossRef]
  62. Dong, X.; Lin, J.; Lu, S.; Wang, H.; Li, Y. Multiscale spatial attention network for seismic data denoising. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–17. [Google Scholar] [CrossRef]
  63. Xu, J.; Deng, X.; Xu, M. Revisiting convolutional sparse coding for image denoising: From a multi-scale perspective. IEEE Signal Process. Lett. 2022, 29, 1202–1206. [Google Scholar] [CrossRef]
  64. Zhang, K.; Gool, L.-V.; Timofte, R. Deep unfolding network for image super-resolution. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 3217–3226. [Google Scholar]
  65. Meinhardt, T.; Moeller, M.; Hazirbas, C.; Cremers, D. Learning proximal operators: Using denoising networks for regularizing inverse imaging problems. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 1799–1808. [Google Scholar]
  66. Sreter, H.; Giryes, R. Learned convolutional sparse coding. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; pp. 2191–2195. [Google Scholar]
  67. Zhu, H.; Zhang, J.; Xu, G.; Deng, L. Balanced ring top-hat transformation for infrared small-target detection with guided filter kernel. IEEE Trans. Aerosp. Electron. Syst. 2021, 56, 3892–3903. [Google Scholar] [CrossRef]
  68. Scetbon, M.; Elad, M.; Milanfar, P. Deep K-SVD Denoising. IEEE Trans. Image Process. 2021, 30, 5944–5955. [Google Scholar] [CrossRef]
  69. Simon, D.; Elad, M. Rethinking the csc model for natural images. Adv. Neural Inf. Process. Syst. 2019, 32, 2274–2284. [Google Scholar]
  70. Dong, W.; Wang, P.; Yin, W.; Shi, G.; Wu, F.; Lu, X. Denoising prior driven deep neural network for image restoration. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 40, 2305–2318. [Google Scholar] [CrossRef] [Green Version]
  71. Fu, X.; Zha, Z.J.; Wu, F.; Ding, X.; Paisley, J. Jpeg artifacts reduction via deep convolutional sparse coding. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 2501–2510. [Google Scholar]
  72. Bertocchi, C.; Chouzenoux, E.; Corbineau, M.C.; Pesquet, J.C.; Prato, M. Deep unfolding of a proximal interior point method for image restoration. Inverse Probl. 2020, 36, 034005. [Google Scholar] [CrossRef] [Green Version]
  73. Zheng, H.; Yong, H.; Zhang, L. Deep convolutional dictionary learning for image denoising. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021. [Google Scholar]
  74. Sun, H.; Liu, M.; Zheng, K.; Yang, D.; Li, J.; Gao, L. Hyperspectral image denoising via low-rank representation and CNN denoiser. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 716–728. [Google Scholar] [CrossRef]
  75. Huang, Z.; Wang, Z.; Zhu, Z.; Zhang, Y.; Fang, H.; Shi, Y.; Zhang, T. DLRP: Learning deep low-rank prior for remotely sensed image denoising. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
  76. Xu, W.; Zhu, Q.; Qi, N.; Chen, D. Deep sparse representation based image restoration with denoising prior. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 6530–6542. [Google Scholar] [CrossRef]
  77. Beck, A.; Teboulle, M. A fast iterative shrinkage-thresholding algorithm for linear inverse problems. SIAM J. Imaging Sci. 2009, 2, 183–202. [Google Scholar] [CrossRef] [Green Version]
  78. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 27–30. [Google Scholar]
  79. Yu, F.; Koltun, V. Multi-scale context aggregation by dilated convolutions. In Proceedings of the 4th International Conference on Learning Representations, San Juan, Puerto Rico, 2–4 May 2016. [Google Scholar]
  80. He, K.; Zhang, X.; Ren, S.; Sun, J. Delving deep into rectifiers: Surpassing human-level performance on ImageNet classification. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 1026–1034. [Google Scholar]
  81. MODIS Data. Available online: https://modis.gsfc.nasa.gov/data/ (accessed on 30 January 2018).
  82. A Freeware Multispectral Image Data Analysis System. Available online: https://engineering.purdue.edu/~biehl/MultiSpec/hyperspectral.html (accessed on 30 January 2018).
  83. Kingma, D.; Ba, J. Adam: A method for stochastic optimization. In Proceedings of the 3rd International Conference on Learning Representations, Banff, AB, Canada, 14–16 April 2014. [Google Scholar]
  84. Huang, Z.; Wang, L.; An, Q.; Zhou, Q.; Hong, H. Learning a contrast enhancer for intensity correction of remotely sensed images. IEEE Signal Process. Lett. 2022, 29, 394–398. [Google Scholar] [CrossRef]
  85. Pyatykh, S.; Hesser, J.; Zheng, L. Image noise level estimation by principal component analysis. SIAM J. Imaging Sci. 2013, 22, 687–699. [Google Scholar] [CrossRef]
  86. Huang, Z.; Zhu, Z.; An, Q.; Wang, Z.; Zhou, T.; Alshomrani, A. Luminance learning for remotely sensed image enhancement guided by weighted least squares. IEEE Geosci. Remote Sens. Lett. 2021, 22, 1–5. [Google Scholar] [CrossRef]
Figure 1. An example of a visual comparison between the model-based method (WSSR) (PSNR|SSIM: 27.2 dB|0.7155) and the proposed MD 3 approach (PSNR|SSIM: 27.63 dB|0.7352) on the MODIS-A1 image with size 512 × 512 and noise level σ = 25 . (a) Ground-truth. (b) Noised. (c) WSSR [6]. (d) Proposed.
Figure 1. An example of a visual comparison between the model-based method (WSSR) (PSNR|SSIM: 27.2 dB|0.7155) and the proposed MD 3 approach (PSNR|SSIM: 27.63 dB|0.7352) on the MODIS-A1 image with size 512 × 512 and noise level σ = 25 . (a) Ground-truth. (b) Noised. (c) WSSR [6]. (d) Proposed.
Remotesensing 15 00445 g001
Figure 2. The flowchart of the proposed model-driven deep remotely sensed image denoising.
Figure 2. The flowchart of the proposed model-driven deep remotely sensed image denoising.
Remotesensing 15 00445 g002
Figure 3. Framework of LISTA encoder for updating sparse coding α i .
Figure 3. Framework of LISTA encoder for updating sparse coding α i .
Remotesensing 15 00445 g003
Figure 4. The flowchart of the deep dictionary network Φ · . “DConv” denotes dilated convolution, while “PReLU” represents the parametric rectified linear unit. “⨁” is a pixelwise operator used to add feature maps.
Figure 4. The flowchart of the deep dictionary network Φ · . “DConv” denotes dilated convolution, while “PReLU” represents the parametric rectified linear unit. “⨁” is a pixelwise operator used to add feature maps.
Remotesensing 15 00445 g004
Figure 5. Test images. Images in (ah) are from the Aqua MODIS dataset, Terra MODIS dataset, hyperspectral dataset, and multispectral dataset, respectively.
Figure 5. Test images. Images in (ah) are from the Aqua MODIS dataset, Terra MODIS dataset, hyperspectral dataset, and multispectral dataset, respectively.
Remotesensing 15 00445 g005
Figure 6. The convergence curve and intermediate results generated by the proposed MD 3 method. (a) Noised ( σ = 20 ). (b) x 1 , PSNR: 26.21 dB. (c) x 3 , PSNR: 28.43 dB. (d) x 5 , PSNR: 28.44 dB. (e) Convergence. (f) D 1 . (g) D 3 . (h) D 5 .
Figure 6. The convergence curve and intermediate results generated by the proposed MD 3 method. (a) Noised ( σ = 20 ). (b) x 1 , PSNR: 26.21 dB. (c) x 3 , PSNR: 28.43 dB. (d) x 5 , PSNR: 28.44 dB. (e) Convergence. (f) D 1 . (g) D 3 . (h) D 5 .
Remotesensing 15 00445 g006
Figure 7. Visual comparison of experiments on the MODIS-T2 image. (a) Ground-truth. (b) Noised ( σ = 25 ). (c) NCSR [43]. (d) WSSR [6]. (e) DnCNN [18]. (f) DKSVD [68]. (g) DCDicL [73]. (h) LRR-CNN [74]. (i) DLRP [75]. (j) Proposed.
Figure 7. Visual comparison of experiments on the MODIS-T2 image. (a) Ground-truth. (b) Noised ( σ = 25 ). (c) NCSR [43]. (d) WSSR [6]. (e) DnCNN [18]. (f) DKSVD [68]. (g) DCDicL [73]. (h) LRR-CNN [74]. (i) DLRP [75]. (j) Proposed.
Remotesensing 15 00445 g007
Figure 8. Visual comparison of experiments on the MODIS-A3 image. (a) Ground-truth. (b) Noised ( σ = 30 ). (c) NCSR [43]. (d) WSSR [6]. (e) DnCNN [18]. (f) DKSVD [68]. (g) DCDicL [73]. (h) LRR-CNN [74]. (i) DLRP [75]. (j) Proposed.
Figure 8. Visual comparison of experiments on the MODIS-A3 image. (a) Ground-truth. (b) Noised ( σ = 30 ). (c) NCSR [43]. (d) WSSR [6]. (e) DnCNN [18]. (f) DKSVD [68]. (g) DCDicL [73]. (h) LRR-CNN [74]. (i) DLRP [75]. (j) Proposed.
Remotesensing 15 00445 g008
Figure 9. Visual comparison of experiments on the Multi-WDC image. (a) Ground-truth. (b) Noised ( σ = 40 ). (c) NCSR [43]. (d) WSSR [6]. (e) DnCNN [18]. (f) DKSVD [68]. (g) DCDicL [73]. (h) LRR-CNN [74]. (i) DLRP [75]. (j) Proposed.
Figure 9. Visual comparison of experiments on the Multi-WDC image. (a) Ground-truth. (b) Noised ( σ = 40 ). (c) NCSR [43]. (d) WSSR [6]. (e) DnCNN [18]. (f) DKSVD [68]. (g) DCDicL [73]. (h) LRR-CNN [74]. (i) DLRP [75]. (j) Proposed.
Remotesensing 15 00445 g009
Figure 10. Three real-life noisy RSI images used for further testing. (a) Mountain. (b) City-wall. (c) Factory.
Figure 10. Three real-life noisy RSI images used for further testing. (a) Mountain. (b) City-wall. (c) Factory.
Remotesensing 15 00445 g010
Figure 11. Visual comparison of experiments on the real-life Factory image. (a) NCSR [43]. (b) WSSR [6]. (c) DnCNN [18]. (d) DKSVD [68]. (e) DCDicL [73]. (f) LRR-CNN [74]. (g) DLRP [75]. (h) Proposed.
Figure 11. Visual comparison of experiments on the real-life Factory image. (a) NCSR [43]. (b) WSSR [6]. (c) DnCNN [18]. (d) DKSVD [68]. (e) DCDicL [73]. (f) LRR-CNN [74]. (g) DLRP [75]. (h) Proposed.
Remotesensing 15 00445 g011
Figure 12. Visual comparison of experiments on the real-life Mountain image. (a) NCSR [43]. (b) WSSR [6]. (c) DnCNN [18]. (d) DKSVD [68]. (e) DCDicL [73]. (f) LRR-CNN [74]. (g) DLRP [75]. (h) Proposed.
Figure 12. Visual comparison of experiments on the real-life Mountain image. (a) NCSR [43]. (b) WSSR [6]. (c) DnCNN [18]. (d) DKSVD [68]. (e) DCDicL [73]. (f) LRR-CNN [74]. (g) DLRP [75]. (h) Proposed.
Remotesensing 15 00445 g012
Figure 13. Computational time vs. PSNR for different approaches on the MODIS-A1 image with size 512 × 512 ; the noise level is 25.
Figure 13. Computational time vs. PSNR for different approaches on the MODIS-A1 image with size 512 × 512 ; the noise level is 25.
Remotesensing 15 00445 g013
Table 1. PSNR (dB) comparisons of the state-of-the-art. For each image at each level, the best result is marked with bold font.
Table 1. PSNR (dB) comparisons of the state-of-the-art. For each image at each level, the best result is marked with bold font.
MethodsNoise Level σ = 15 Noise Level σ = 20
MODIS-Hyper-WDCMMulti-WDCAve.MODIS-Hyper-WDCMMulti-WDCAve.
A1A2A3T1T2T3A1A2A3T1T2T3
NCSR28.928.8129.6930.3735.8334.6728.5331.0330.9827.527.5328.072934.6133.4127.1530.3429.7
WSSR29.5829.3230.4731.2136.4835.3229.6232.6331.8528.2228.228.9429.6435.273427.8331.430.44
DnCNN29.3929.2430.3731.0236.3235.1929.532.6131.7128.0628.0628.6729.4435.0433.927.5931.130.23
DKSVD29.3629.1730.2230.7736.2135.1729.4932.4331.627.9627.9428.6629.4135.0333.8527.5131.0430.18
DCDicL29.4529.3430.3330.8236.4135.2529.5232.5131.728.0728.0628.7929.5535.1933.9727.731.1930.32
LRR-CNN29.4229.3830.5430.8936.4835.2729.3732.5731.7428.1128.1128.8729.5735.2433.9927.7531.3330.37
DLRP29.5829.530.5631.0936.6135.3829.6532.6531.8828.328.252929.6935.2934.1528.1731.4430.54
Proposed29.7629.7230.7931.2536.7835.6329.7932.8732.0628.5428.6129.3729.935.6634.5428.4431.7330.85
Noise level σ = 25 Noise level σ = 30
NCSR26.8726.9527.0828.1833.8632.725.9929.5528.926.0326.2326.0827.5533.2531.8824.828.5328.04
WSSR27.227.2827.6828.6934.4133.0726.7730.3329.4326.4426.6626.8228.0733.7532.3525.6529.4828.65
DnCNN27.1227.227.5628.5134.2333.0226.630.0629.2926.1426.5126.5727.9633.4532.2225.5329.1628.44
DKSVD26.9527.0427.5228.4734.2133.0126.7229.9229.3226.1526.426.5727.7833.4232.225.3629.0628.37
DCDicL27.1427.2427.6128.6434.3833.0426.7730.0929.3626.226.6326.6528.133.632.2925.5529.2528.66
LRR-CNN27.1127.2527.6728.6734.3933.0726.7430.1129.3826.3526.6526.8228.0333.7432.325.6429.3928.62
DLRP27.3127.3227.7528.8134.4733.1626.8130.3229.4926.4726.6926.8328.3633.7932.3625.729.5228.72
Proposed27.6327.6928.1729.1435.0133.6927.0330.6829.8826.7526.9527.1128.373432.862629.728.97
Noise level σ = 35 Noise level σ = 40
NCSR25.3625.4825.3227.0632.3631.3724.2328.0527.424.9525.0725.0426.5332.0230.6723.6227.5426.93
WSSR25.8526.1426.1127.4633.0631.6525.0828.7628.0125.4125.625.52732.4631.1324.2628.1827.44
DnCNN25.8225.8925.8927.2532.931.524.9228.4527.8325.1325.3825.2626.8132.1731.1324.0227.9527.23
DKSVD25.6525.7725.8627.2432.6931.4824.9428.3927.7525.0825.2925.1826.7232.1130.8823.8927.7127.11
DCDicL25.8125.9825.8927.4332.9931.5125.0128.5127.8925.3425.4725.3126.8732.3431.0524.1128.0327.32
LRR-CNN25.8326.0126.0827.4432.9931.5725.0428.6527.9525.3725.4125.4926.9332.3631.1124.2128.0727.37
DLRP25.9526.1926.1527.533.1631.6625.1228.8728.0825.4125.7225.5427.0532.6831.1624.2728.1227.49
Proposed26.1126.426.3227.7633.3432.0225.2129.0128.2725.6225.9725.5827.2232.8631.2924.6528.5527.72
Table 2. SSIM comparisons of the state-of-the-art. For each image at each level, the best result is marked with bold font.
Table 2. SSIM comparisons of the state-of-the-art. For each image at each level, the best result is marked with bold font.
MethodsNoise Level σ = 15 Noise Level σ = 20
MODIS-Hyper-WDCMMulti-WDCAve.MODIS-Hyper-WDCMMulti-WDCAve.
A1A2A3T1T2T3A1A2A3T1T2T3
NCSR0.8050.7680.91080.84010.88530.88010.89120.8610.85520.75490.69730.88510.76920.87880.8690.83410.82870.8146
WSSR0.82720.79830.91760.84820.9080.89980.90140.86470.87080.77550.71920.88440.79310.88920.87510.84930.83850.828
DnCNN0.81270.77630.91570.8320.90840.89890.89370.86190.86250.76480.71780.87950.78050.88930.8740.84740.83390.8234
DKSVD0.82250.77780.91840.84280.90780.89660.90030.86940.86160.75080.69290.88640.77910.890.87350.84530.82340.8177
DCDicL0.81560.78120.91530.83720.90770.89980.85780.86660.86550.75810.70990.88310.78870.88840.87550.85350.83220.8237
LRR-CNN0.82060.78660.91440.84140.90940.89710.89340.86960.86660.75850.71220.88270.78750.89140.87430.84490.83610.8235
DLRP0.82670.79710.91630.84750.91260.90160.90670.87160.87250.77170.72710.8870.79810.89190.88040.85180.83810.8308
Proposed0.83860.79920.92110.85170.91480.90690.90930.87430.87670.79520.74880.90080.81720.90250.89380.87230.85370.848
Noise level σ = 25 Noise level σ = 30
NCSR0.64820.63810.83810.72680.86950.84730.79520.78580.76860.62860.58620.82850.68250.85210.82410.74840.74970.7175
WSSR0.71550.66910.85350.74720.87470.85780.81040.80650.79180.66370.62420.82480.72270.86720.8420.76980.77910.7617
DnCNN0.69950.65150.84810.73210.87440.85660.80850.80250.78420.63140.59620.81730.69580.86230.83950.7690.76530.7471
DKSVD0.68370.6420.84530.73230.87320.85290.80230.78950.77770.62970.5970.81550.69110.86060.83830.7580.75790.7435
DCDicL0.65690.66360.85630.74910.87550.8540.81170.80020.78340.65710.61850.8240.71610.86140.83740.76580.77390.7568
LRR-CNN0.70360.66580.85630.74540.87680.85730.80940.80630.79010.66190.6210.82370.71710.86440.84070.76920.77910.7596
DLRP0.72970.66860.85330.7590.88010.86020.81130.80760.79620.66450.6260.82990.71650.86270.83920.77310.78030.7615
Proposed0.73520.68840.86670.77080.88750.87090.81780.81030.8060.67420.63190.83280.73750.86980.85310.790.78750.7721
Noise level σ = 35 Noise level σ = 40
NCSR0.56790.54360.77780.6580.83920.81810.71150.72570.70520.52340.51370.74740.63330.8320.79780.65970.69720.6756
WSSR0.6170.59030.80270.69450.85550.82180.73160.75360.73340.57640.55630.77150.6640.84320.81320.69780.7320.7068
DnCNN0.58420.56560.790.66220.84830.81940.71930.74050.71620.54540.52490.75950.6420.83730.80550.67570.7120.6878
DKSVD0.57990.54830.78340.6620.84650.82220.71330.73510.71130.53330.51620.75930.63320.83440.80560.67650.70320.6827
DCDicL0.62030.58860.78850.68880.85550.82680.73350.74930.73140.58470.54520.7660.6610.83620.80780.69610.72750.7031
LRR-CNN0.61730.58990.79410.68010.84980.82420.73450.75350.73040.58180.56010.7710.65540.83330.80570.69980.72840.7044
DLRP0.62140.58960.79960.68620.85560.82550.72920.7590.73330.5850.55980.77880.66630.84630.81070.70220.73060.71
Proposed0.6410.60640.80550.7030.85680.83450.74030.76670.74430.60360.58360.78320.68540.85110.8160.7260.74710.7245
Table 3. Quantitative comparisons of the QM and NIQE values produced by the state-of-the-art. For each image, the best result is marked with bold font.
Table 3. Quantitative comparisons of the QM and NIQE values produced by the state-of-the-art. For each image, the best result is marked with bold font.
MethodsMountainCity-WallFactory
QMNIQEQMNIQEQMNIQE
NCSR18.654.6316.567.0833.235.13
WSSR22.723.6121.053.5739.354.09
DnCNN21.623.9320.055.4337.764.61
DKSVD20.754.4319.325.0736.534.64
DCDicL23.53.7721.483.6140.364.31
LRR-CNN22.073.5619.783.6938.124.14
DLRP22.333.7119.74.0638.993.99
Proposed25.343.524.463.4244.193.83
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Huang, Z.; Zhu, Z.; Zhang, Y.; Wang, Z.; Xu, B.; Liu, J.; Li, S.; Fang, H. MD3: Model-Driven Deep Remotely Sensed Image Denoising. Remote Sens. 2023, 15, 445. https://doi.org/10.3390/rs15020445

AMA Style

Huang Z, Zhu Z, Zhang Y, Wang Z, Xu B, Liu J, Li S, Fang H. MD3: Model-Driven Deep Remotely Sensed Image Denoising. Remote Sensing. 2023; 15(2):445. https://doi.org/10.3390/rs15020445

Chicago/Turabian Style

Huang, Zhenghua, Zifan Zhu, Yaozong Zhang, Zhicheng Wang, Biyun Xu, Jun Liu, Shaoyi Li, and Hao Fang. 2023. "MD3: Model-Driven Deep Remotely Sensed Image Denoising" Remote Sensing 15, no. 2: 445. https://doi.org/10.3390/rs15020445

APA Style

Huang, Z., Zhu, Z., Zhang, Y., Wang, Z., Xu, B., Liu, J., Li, S., & Fang, H. (2023). MD3: Model-Driven Deep Remotely Sensed Image Denoising. Remote Sensing, 15(2), 445. https://doi.org/10.3390/rs15020445

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop