Next Article in Journal
Short-Term vs. Long-Term: A Critical Review of Indoor Radon Measurement Techniques
Previous Article in Journal
Predefined-Time Adaptive Neural Tracking Control for a Single Link Manipulator with an Event-Triggered Mechanism
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Hierarchical Low-Rank Denoising Model for Remote Sensing Images Based on Deep Unfolding

1
China Electronic Technology Group Corporation, Beijing 100846, China
2
School of Electronic and Optical Engineering, Nanjing University of Science and Technology, Nanjing 210094, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(14), 4574; https://doi.org/10.3390/s24144574
Submission received: 13 June 2024 / Revised: 30 June 2024 / Accepted: 13 July 2024 / Published: 15 July 2024
(This article belongs to the Section Remote Sensors)

Abstract

:
Recently, the low-rank representation (LRR) model has been widely used in the field of remote sensing image denoising due to its excellent noise suppression capability. However, those low-rank-based methods always discard important edge details as residuals, leading to a common issue of blurred edges in denoised results. To address this problem, we take a new look at low-rank residuals and try to extract edge information from them. Therefore, a hierarchical denoising framework was combined with a low-rank model to extract edge information from low-rank residuals within the edge subspace. A prior knowledge matrix was designed to enable the model to learn necessary structural information rather than noise. Also, such traditional model-driven approaches require multiple iterations, and the solutions may be very complex and computationally intensive. To further enhance the noise suppression performance and computing efficiency, a hierarchical low-rank denoising model based on deep unrolling (HLR-DUR) was proposed, integrating deep neural networks into the hierarchical low-rank denoising framework to expand the information capture and representation capabilities of the proposed shallow model. Sufficient experiments on optical images, hyperspectral images (HSI), and synthetic aperture radar (SAR) images showed that HLR-DUR achieved state-of-the-art (SOTA) denoising results.

1. Introduction

Remote sensing data acquisition often relies on sensors mounted on satellites or aircraft, which can capture electromagnetic spectra in different bands, including visible light, infrared, microwaves, etc. Due to its capability to cover vast geographical areas and provide continuous observations, remote sensing technology has important applications in numerous fields such as agriculture, forestry, geology, meteorology, environmental protection, and urban planning [1]. However, remote sensing images are inevitably affected by various factors during acquisition and transmission, leading to the presence of noise in the images [2].
In recent years, the rapid development of deep learning has injected new vitality into various fields [3,4], particularly in the domain of image processing [5,6,7]. Utilizing the powerful learning and representational capabilities of deep neural networks can more effectively solve many challenges and problems in image processing [8], which also brings inspiration to the issue of remote sensing image denoising. Compared to traditional image processing methods, deep learning, through an end-to-end training process, can automatically learn high-level, semantically rich feature representations from a large amount of data. These learned features not only have better robustness and generalization capability for image-denoising tasks but also provide richer image descriptions and understanding.
Deep learning methods have been proven to have significant effects in the task of denoising remote sensing images. For optical remote sensing images, a method of remote sensing image denoising based on a Generative Adversarial Network (GAN) named “Restoration Generative Adversarial Network with ResNet and DenseNet” (RRDGAN) was proposed [9]. This method acquires better-quality images by incorporating denoising and SR into a unified framework. Huang et al. integrated deep convolutional neural network (DCNN) denoiser prior into a unidirectional variation (UV) model, named UV-DCNN, to simultaneously destripe and denoise optical remote sensing images [10]. For hyperspectral images (HSI) images, Xie et al. introduced the Deep Spatio-Spectral Bayesian Posterior Network (DSSBPNet) for hyperspectral images [11]. DSSBPNet combines Bayesian variational posterior with deep neural networks, where a deep spatial-spectral network divides the input image into three parts, generating spatial-spectral gradients for each. Different convolutions are used in different parts of the DSS network. Meanwhile, noise estimation of the original data, noise distribution, and sparse noise gradients constitute the Bayesian posterior approach. Finally, the forward-backward propagation method is used to connect DSS with the Bayesian posterior. A novel, deep-learning framework for 3-D HSI denoising [12] was proposed, which decomposes 3-D filtering into 2-D spatial filtering and 1-D spectral filtering. This method can achieve substantial savings on the number of network parameters to keep the computational complexity low. For synthetic aperture radar (SAR) images, a novel two-component deep learning (DL) network was proposed for SAR despeckling [13]. First, the texture estimation subnetwork is constructed to produce the texture level map (TLM), which evaluates the randomness and scale of the texture distribution. Then, the noise removal subnetwork learns a spatially variable mapping between the noise and clean images with the help of TLM.
However, we cannot ignore the important role of traditional models in the field of denoising. Low-rank representation (LRR) models have been widely explored and improved in the field of remote sensing image denoising [14,15]. A method called NAILRMA [16] proposed an iterative regularization framework aiming to achieve denoising of signal subspaces. In NAILRMA, an adaptive iteration factor selection method based on noise variance was applied to each band of the hyperspectral image (HSI). Additionally, some tensor-based low-rank denoising methods [17,18,19] have also demonstrated satisfactory results. The Fisher–Tippett distribution-based WNNM (FT-WNNM) [20] attempts to recover underlying low-rank components from patch group matrices. The Composite Regularization Method for Spot Noise Reduction (CRM-SR) [21] introduces regularization terms, including TV regularization and NLLR regularization. However, determining the regularization parameters involved in the CRM-SR model is challenging. The NLLR framework [22] decomposes speckle noise into low-rank components with proposed truncation and weighted nuclear norm regularization. Considering the effectiveness of the LRR model, more and more denoising algorithms have combined low-rank models with deep learning denoising methods.
Zhang et al. proposed a Deep Low-Rank Prior (DLRP) [23] for remote sensing image denoising, firstly using the low-rank characteristics of nearby non-local self-similar blocks arranged in dictionary order to model the Global Objective Function (GOF). Secondly, with the help of an alternating iterative strategy, GOF can be easily decomposed into two independent sub-problems. Among them, the low-rank minimization denoising problem can be solved through learning with a deep convolutional neural network. Nguyen et al. proposed an HSI denoising method using deep image prior with sparse and low-rank prior [24]. However, hyperparameter selection in this method is still an open question. Sun et al. proposed a HSI denoising via LRR and CNN denoiser [25]. The sparse-based low-rank representation can explore the global correlations in both the spatial and spectral domains, and the CNN-based denoiser can represent the deep prior, which cannot be designed by traditional restoration models. However, such denoising methods often have a common problem, that is, they will filter out significant edge signals as noise, which will degrade the quality of images. Zhang et al. developed a self-supervised HSI denoising method via integrating a model-driven strategy with a data-driven strategy [26]. The proposed framework simultaneously cooperates with the spectral low-rankness prior and deep spatial prior (SLRP-DSP) for HSI self-supervised denoising. Regarding the characteristics of SAR images, more novel deep networks were proposed. Xiong et al. decomposed the SAR image in the SAR imaging-despeckling observation model into a sparse matrix and a low-rank matrix, and then established an optimization problem with the corresponding sparse and low-rank priors [27]. A despeckling model was proposed that uses deeper convolutional neural networks [28], which was never used before, as far as authors are concerned, for diminishing speckles in noisy SAR images. Multiple skip connections from the ResNet model are also employed in the authors’ proposed architecture. However, these denoising methods often do not take into account the issue of edge preservation in the denoising results, leading to the loss of some important structural information in the denoised images.
In this paper, we combine the shallow low-rank denoising model with an autoencoder, following the method in [29], to construct a two-level deep encoder. By applying deep unfolding, the shallow LRR denoising model was introduced into the realm of deep learning. Thus, a hierarchical low-rank denoising model based on deep unrolling (HLR-DUR) was proposed. Integrating the characteristics of autoencoders, a two-level deep autoencoder was established to optimize the iterative solution process for the low-rank and edge components of remote sensing images. Utilizing the advantages of deep neural networks, the model captures more complex patterns in the data, thereby enhancing the effectiveness of image denoising. Finally, experiments conducted on optical remote sensing images, hyperspectral images, and SAR images demonstrated that HLR-DUR significantly outperforms shallow models across various metrics.
The main contributions can be summarized as follows:
  • We think the residuals of traditional low-rank models contain not only noise but also high-frequency edge information that has been filtered out. Therefore, we tried to re-extract edge information from the residuals to construct a hierarchical low-rank model for denoising remote sensing images. To better achieve edge extraction, we introduced manifold learning into the model, ensuring that useful structural information, rather than noise, is extracted in the edge subspace. Additionally, a new prior knowledge regulation was designed to distinguish between clean pixels and noisy pixels, guiding the model to learn useful information during the denoising process;
  • Due to the number of iterations required, traditional LRR models usually take a long time to process large-sized remote sensing images. To tackle this problem, we designed a hierarchical denoising model based on deep unfolding by combining shallow models with deep autoencoders. By mapping traditional iterative optimization algorithms onto the structure of neural networks, each iterative step was transformed into a layer of the network. This technique enhances the interpretability of neural networks, making each layer no longer a black box model but rather corresponding to a specific optimization step;
  • Experiments conducted on three types of images—optical remote sensing images, hyperspectral images, and SAR images—demonstrated the effectiveness of our proposed HLR-DUR model. HLR-DUR not only enhances denoising performance but also better preserves sufficient edge detail compared to existing methods, achieving better edge-preserving denoising.

2. Related Work

Wang et al. proposed that a low-rank prior and a fully convolutional Auto Encoder (AE) can be incorporated through modeling an energy minimization problem [30], which is similar to the Robust Principal Component Analysis (RPCA) model.
Given a data matrix X, it can be decomposed as [31]:
X = Z + E
where Z represents the low-rank part and E is a sparse matrix, which represents residual. Z and E can then be optimized by solving the following energy minimization problem:
min Z , E Z * + λ E 1 s . t . X = Z + E
where · * denotes nuclear norm, · 1 denotes l 1 norm, and λ is the regularization parameter.
Based on this model, the low-rank prior and a fully convolutional AE are incorporated by modeling an energy minimization problem, which can be mathematically depicted as follows:
min Z , θ 1 2 ( f θ ( Y ) X ) W 2 + λ Z * s . t . Z = f θ ( Y )
where Y has the same dimensionality as the hyperspectral image, f θ ( Y ) is the network which uses the uniform distribution as a prior to generate a reconstruction of the background, and W is the adaptive-weighted map.
Then, the problem can be solved by the Alternating Direction Method of Multipliers (ADMM). The model can be described as
min Z , θ 1 2 ( f θ ( Y ) X ) W 2 + λ Z * + μ 2 Z f θ ( Y ) 2 + S T ( Z f θ ( Y ) )
where S stands for the Lagrange multiplier, and μ is a penalty parameter.
Finally, the above-shown optimization problem can be split into three subproblems, and each subproblem can be solved alternately in an iterative procedure.

3. Proposed Methodology

As previously mentioned, the presence of noisy data can destroy the low-rank structure of the data. So, the LRR model splits the data into clean parts and noisy parts, separating the noisy data while maintaining the low-rank structure of the data. However, the LRR model has a drawback that cannot be ignored. Specifically, it strives to maintain a low-rank structure by filtering out useful high-frequency information, which then causes important image edges to be discarded as part of residuals. Motivated by this phenomenon, we first try to construct a hierarchical low-rank model to extract edge details from residuals. Then, we try to combine the proposed hierarchical model with deep neural networks to further improve the denoising effect.

3.1. Hierarchical Low-Rank Model

3.1.1. Model Formulation

To address the issue of traditional low-rank models failing to preserve structural edge information, we modify the traditional low-rank model using a prior knowledge matrix. This modification allows the model to better distinguish between clean and noisy pixels. The model can be represented as
min Z , E Z + λ E 2 , 1 s . t .   H X = H Z + E
where X represents the original noisy image, Z represents the low-rank component, and E represents the low-rank residual. H is the prior knowledge matrix used in this model, “◦” denotes the Hadamard product, and λ is an adjustable parameter used to balance the residual term.
It is crucial to extract the edge information that has been filtered out from the residuals. Building on (5), we innovatively introduce a constraint term H E = H E 1 + H E 2 , where E 1 represents the edge part extracted from the residual, and E 2 represents the noise component in the residual. By further decomposing the residual, the filtered edge information can be obtained, thereby enhancing the quality of the denoising results. However, considering the different requirements for extracting residuals and low-rank components, we further set up hierarchical prior knowledge information, implemented through different coefficients, to meet different requirements. Specifically, when extracting the low-rank part, the goal is to achieve as clean denoising results as possible, that is, the less noise, the better, even if it means that the low-rank part of the image may have blurred edges and lose some detail. When extracting edge structures, the focus is more on effectively extracting rich edge information, even if some noise is introduced in the process of extracting edges. To achieve this goal, we further propose a hierarchical prior knowledge to better extract the edge details from the residuals, with the model presented as follows:
min Z , E , E 1 , E 2 Z + λ E 2 , 1 s . t .   H 1 X = H 1 Z + E     H 2 E = H 2 E 1 + E 2
It is noteworthy that in the constraints of (6), H matrices with different subscripts appear, as a special design of the hierarchical model. Here, H 1 is used to obtain the clean low-rank part from the noisy image, while H 2 is used to find more edge details from the residuals. As mentioned above, H 1 should be more stringent to ensure that as much noise as possible is eliminated. In contrast, the constraints on H 2 are relatively lenient because H 2 needs to extract more useful information from the residuals.
However, extracting sufficient edge information from residuals containing a large amount of noise is not an easy task. To ensure that our model learns the required edge information rather than noise, and to ensure that good data geometric structures are captured within the model, a manifold learning framework was also introduced to the proposed model [32]. Manifold learning is a machine learning technique that explores and exploits the underlying low-dimensional structures of high-dimensional data, based on the idea that while data may be presented in high-dimensional spaces, they actually distribute along some low-dimensional manifolds. The purpose of manifold learning is to reveal these low-dimensional structures to better understand the essential characteristics of the data. By introducing manifold learning, robust subspace projections V = ( v 1 , v 2 , , v r ) can be learned from the residuals E 1 , allowing for the extraction of edge subspaces. By introducing dynamic affinity graph regularization, manifold learning can be incorporated into the original model, with the formula represented as follows:
min V , Z , W , E , E 1 , E 2 1 2 i , j V T e 1 , i V T e 1 , j 2 2 W i j + λ W F 2 s . t .   H 1 X = H 1 Z + E     H 2 E = H 2 E 1 + E 2     V T V = I     W 1 = 1 , W i i = 0 , W 0
where e 1 , i and e 1 , j represent the values of corresponding samples in the residual E 1 , and 1 / 2 i , j V T e 1 , i V T e 1 , j 2 2 is an increased coupling term. It is a part of our model because we do not know which part of the data is affected by noise. W is the dynamic affinity graph matrix, and λ is an adjustable parameter used to balance the dynamic affinity graph.
For two similar samples x i and x j , if the values of corresponding e 1 , i and e 1 , j are very small, this also suggests that the distance between them will be very small. Therefore, it can be considered that they primarily contain edge information, which are the edge details to be extracted from the residual E . Conversely, if they are part of the noise, the distance between them will be increased. Based on this assumption, a weighting coefficient W i j is introduced to adaptively learn projections V with similar residuals. When two residual samples are similar, W i j is set to non-zero. Otherwise, W i j is set to zero to mitigate the impact of noise. At the same time, three non-negative constraints W i i = 0 , W 0 , and W 1 = 1 are also added to the model, and these three constraints are used to ensure that the affinity graph will be normalized. Finally, the complete model is as follows:
min V , Z , W , E , E 1 , E 2 1 2 i , j V T e 1 , i V T e 1 , j 2 2 W i j + λ 1 Z + λ 2 E 2 , 1 + λ 3 W F 2 s . t .   H 1 X = H 1 Z + E     H 2 E = H 2 E 1 + E 2     V T V = I     W 1 = 1 , W i i = 0 , W 0
where λ 1 , λ 2 , λ 3 are adjustable parameters used to balance the different terms.
To solve (8), we employed the Augmented Lagrange Multiplier method (ALM). The coupling term 1 / 2 i , j V T e 1 , i V T e 1 , j 2 2 can be rewritten as t r ( V T ( F L E ^ T ) V ) , where L = D W is the graph Laplacian operator of W , and D = d i a g i W i j is the diagonal matrix. Then, a relaxation variable S to denote Z , a relaxation variable E 1 to denote E ^ and a relaxation variable R to denote E are introduced to make (8) easier to solve. Additionally, another variable F is introduced to avoid the Sylvester equation during the solution process for E . Therefore, (8) is reformulated as
min V , Z , W , E , E 1 , E 2 t r ( V T ( F L E ^ T ) V ) + λ 1 Z + λ 2 E 2 , 1 + λ 3 W F 2 s . t .   H 1 X = H 1 Z + E     H 2 E = H 2 E 1 + E 2     V T V = I     W 1 = 1 , W i i = 0 , W 0     S = Z , E 1 = E ^ , F = E ^ , R = E

3.1.2. Prior Knowledge Regulation Construction

Firstly, we constructed the basic prior knowledge matrix based on the work of ROLD [33]. ROLD is a rank-ordered logarithmic difference, which can identify more noisy pixels with fewer false hits. With such statistics as cornerstones, we tried to construct our prior knowledge matrix. The uncorrupted probability h i of each pixel k i using ROLD is as follows:
h i = exp ( α R O L D ( k i ) )
where α is the coefficient used to obtain image features under different requirements. When α is large, the difference between uncontaminated pixels and contaminated pixels will be widened, and the final result will filter most speckle noise, but more details such as edges will also be filtered out together. On the contrary, when we set α small, it is more likely to retain the edge details of images even though some speckle noise may also be left on images.
Furthermore, we adapted the abovementioned prior knowledge matrix to make it more compatible with SAR images. The backscattered signatures from point targets are dominated by a small number of strong scatterers, which will make them classified as noisy pixels. Therefore, we adopted the method of the Enhanecd Lee filter (EnLee) [34] to divide all pixels into three categories, and combined it with (10) to make the prior knowledge matrix more suitable for SAR images. Through using the hierarchical prior knowledge matrices, our method can better learn the intrinsics of the low-rank part and edge part, and obtain clearer results.
Usually, two classes are considered in traditional filters: homogeneous and heterogeneous. Enhanced Lee filter presented that isolated point targets need special treatment as a third class. For an NLOOK intensity image, it is filtered as follows:
R ^ ( t ) = { I ¯ , C i < C min I ¯ ( t ) + ω ( t ) ( I ( t ) I ¯ ( t ) ) I , C i > C max , C min C i C max
where C i = s t d I ¯ , std is the standard deviation within the filter window. C min = 1 N L O O K , C max = 1 + 2 N L O O K .
The weight function can be written as
ω ( t ) = exp ( ( C i C min ) ( C max C i ) )
Learning from EnLee, we divide all pixels into three categories to obtain the weight matrix ε :
(1)
C i > C max , which means that this pixel is a point target and needs to be preserved. So, we set the corresponding value in ε to 0.9 (a very big value).
(2)
C min < C i < C max , update ε as 1 ω according to (12).
(3)
C i < C min , which represents that this pixel is a noisy pixel and needs to be smoothed. So, we set the corresponding value in ω to 0.1 (a respectively small value).
Then, the new prior knowledge matrix is constructed as follows:
H i = h i ε
Finally, H i needs to be normalized to represent the probability of 0–1.

3.2. Deep Unfolding Model HLR-DUR Based on Autoencoder

In recent years, deep low-rank representation models have received increasing attention as a potential solution for designing interpretable neural networks. By mapping traditional iterative optimization algorithms into the structure of neural networks, each iterative step is transformed into a layer of the network, enhancing the interpretability of neural networks. Each layer of the network is no longer a black-box model but corresponds to a specific optimization step. In the unfolded network, models based on traditional physical meanings are used as the basic architecture, and the parameters that need manual adjustment are transformed into trainable parameters of the network. The core idea of the deep unfolding technique is to utilize deep neural networks to simulate and optimize the iterative process of low-rank representation. Therefore, based on (9), the new model can be described as
min V , Z , S , W , E , E 1 , E 2 , E ^ , F , R t r ( V T ( F L E ^ T ) V ) + λ 1 S + γ 1 Z f d e 1 ( X ; θ 1 ) F 2 + λ 2 E 2 , 1 + γ 2 E 1 f d e 2 ( E ; θ 2 ) F 2 + λ 3 W F 2 s . t .   H 1 X = H 1 Z + E , H 2 E = H 2 E 1 + E 2     E 1 = E ^ , S = Z , R = E     V T V = I , W 1 = 1 , W i i = 0 , W 0
where f d e 1 ( X ; θ 1 ) = g ( W L g ( W i g ( W 2 X ) ) ) denotes a multi-layer deep autoencoder, θ 1 = { W 2 , , W L } is a set of learning parameters used for learning the low-rank component. Additionally, f d e 2 ( E ; θ 2 ) is used to extract edge part, and θ 2 is also a set of learning parameters.
Before solving for the optimal solution of each variable in (14), it is necessary to first obtain its Augmented Lagrangian Multiplier (ALM) function; the complete equation is as follows:
min V , Z , S , W , E , E ^ , E 1 , E 2 , F , R t r ( V T ( F L E ^ T ) V ) + λ 1 S + λ 2 E 2 , 1 + λ 3 W F 2 + γ 1 Z f d e 1 ( X ; θ 1 ) F 2 + γ 2 E 1 f d e 2 ( E ; θ 2 ) F 2 + t r ( Y 1 T ( H 1 X H 1 Z E ) ) + t r ( Y 2 T ( H 2 R H 2 E 1 E 2 ) ) + t r ( Y 3 T ( S Z ) ) + t r ( Y 4 T ( E ^ E 1 ) ) + t r ( Y 5 T ( F E ^ ) ) + t r ( Y 6 T ( R E ) ) + μ 2 ( H 1 X H 1 Z E F 2 + H 2 R H 2 E 1 E 2 F 2 + S Z F 2 + E ^ E 1 F 2 + F E ^ F 2 + R E F 2 ) s . t .   V T V = I     W 1 = 1 , W i i = 0 , W 0
where Y 1 , Y 2 , Y 3 , Y 4 , Y 5 and Y 6 are the introduced Lagrange multiplier.
By establishing a two-level deep encoder, the traditional hierarchical low-rank model is extended to a deep learning network. In traditional iterative algorithms, the denoising task is typically addressed through multiple iterative steps, with each step being computed based on the output of the previous one. In deep unfolding, these iterative steps are “unfurled” into a sequence of layers, each layer corresponding to an iterative step. Within these unfolded layers, the iterative process of low-rank representation is simulated and optimized through a deep autoencoder network.
The flowchart of the proposed model is depicted in Figure 1. Initially, the noisy image undergoes preprocessing with a prior knowledge matrix. Subsequently, the preprocessed noisy image is fed into HLR-DUR to obtain the low-rank and edge components. Finally, the denoised edge and low-rank components are merged to produce a clean image.
The above iterative process is unfolded into a neural network, whose general architecture and components are shown in Figure 2. It consists of four main modules: the low-rank module, the edge module, the update module for the other parameters, and the Lagrangian multiplier module. Each module corresponds to the subproblem.
Next, we introduce each module.
  • low-rank module
The low-rank module is used to compute the clean low-rank component extracted from noisy data. θ 1 is computed by minimizing L θ 1 with respect to the weight parameters θ 1 while keeping other parameters fixed. The optimal solution is obtained by solving the following subproblem:
L θ 1 = Z f d e 1 ( X ; θ 1 ) F 2
The optimization is performed using the Gradient Descent (GD) algorithm, and the formula can be written as follows:
θ 1 = θ 1 ζ L θ 1 / θ 1
where ζ is the learning rate and L θ 1 / θ 1 represents the gradient of L θ 1 .
  • edge module
The edge module is used to extract useful structural edge information from the residual part. Its solution process is similar to that of the low-rank module, which can be described as
L θ 2 = E 1 f d e 2 ( E ; θ 2 ) F 2
θ 2 = θ 2 ζ L θ 2 / θ 2
  • other variables
This module is used to update other variables apart from the low-rank part and the edge part. We have eight variables that need to be updated: V , S , W , E , E ^ , E 2 , F , R . We adopt the alternating direction method of multipliers (ADMM) [35] strategy to accelerate convergence.
V subproblem:
Firstly, update V by fixing the other variables, and this subproblem can be expressed as
min V t r ( V T ( F L E ^ T ) V ) s . t .   V T V = I
Denote F L E ^ T = Q , then we can obtain
min V t r ( V T Q V ) s . t .   V T V = I
V is obtained using the above-shown eigenvalue decomposition problem whose optimal solution involves the set of k eigenvectors of the topmost k smallest eigenvalues of Q .
S subproblem:
min S λ 1 S + μ 2 S Z + Y 3 μ F 2
Using the singular value thresholding (SVT), the optimal S is obtained as
S = Φ λ 1 μ ( Z Y 3 μ )
where Φ represents the soft-thresholding operator.
W subproblem:
min W 1 2 i , j V T e 1 , i V T e 1 , j 2 2 W i j + λ 3 W F 2 s . t .   W 1 = 1 , W i i = 0 , W 0
We can find that (24) is independent for each i, so W i can be solved separately as follows:
min W i 1 2 i , j V T e 1 , i V T e 1 , j 2 2 W i j + λ 3 W i F 2 s . t .   W 1 = 1 , W i i = 0 , W 0
E subproblem:
min E λ 2 E 2 , 1 + μ 2 H 1 X H 1 Z E + Y 1 μ F 2 + μ 2 R E + Y 6 μ F 2
Denote T = 1 2 ( H 1 X H 1 Z + R + Y 1 μ + Y 6 μ ) , τ = λ 2 μ .
We can obtain the i-th column of the variable E as
E ( : , i ) = { t i τ t i t i   ,   i f   τ < t i 0   ,   o t h e r w i s e
E ^ subproblem:
min E ^ t r ( V T ( F L E ^ T ) V ) + μ E ^ E 1 + Y 4 μ F 2 + μ F E ^ + Y 5 μ F 2
Set / E ^ = 0 , the optimized result can be obtained as
E ^ = 2 μ E 1 V V T F L 2 Y 4 + 2 μ F + 2 Y 5 4 μ
E 2 subproblem:
min E 2 H 2 R H 2 E 1 E 2 + Y 2 μ F 2
E 2 = ( H 2 R ) ( H 2 E 1 ) + Y 2 μ
F subproblem:
min F t r ( V T ( F L E ^ T ) V ) + μ F E ^ + Y 5 μ F 2
F = 2 μ E V V T E ^ L T 2 Y 5 2 μ
R subproblem:
min R H 2 R H 2 E 1 E 2 + Y 2 μ F 2 + R E + Y 6 μ F 2
R = ( H 2 H 2 E 1 + H 2 E 2 H 2 Y 2 μ + E Y 6 μ ) ÷ ( 1 + H 2 H 2 )
where ÷ represents element-wise division.
  • Lagrange multipliers
In this section, we compute and update the Lagrange multipliers.
Y 1 = Y 1 + μ ( H 1 X H 1 Z E )
Y 2 = Y 2 + μ ( H 2 R H 2 E 1 E 2 )
Y 3 = Y 3 + μ ( S Z )
Y 4 = Y 4 + μ ( E ^ E 1 )
Y 5 = Y 5 + μ ( F E ^ )
Y 6 = Y 6 + μ ( R E )

4. Results

To better compare the denoising results of deep learning methods on remote sensing images, we selected the following classical deep learning algorithms as comparative methods: DnCNN [36], FFDNet [37], HSID-CNN [38], GRN [39], DD-CNN [40], and PDSNet [41].
In Section 4.1, we use the Mean Peak Signal-to-Noise Ratio (MPSNR) and Mean Structural Similarity Index (MSSIM) as evaluation metrics. In Section 4.3, we employ the Equivalent Number of Looks (ENL) and Edge Preservation Index (EPI) for evaluation. Our specific parameter settings are as follows: batch size of 150, ADAM optimizer, and learning rate of 5 × 10−3. For selecting parameters such as rank threshold, hierarchical levels, and regularization terms, we integrated our proposed HLR-DUR model into an AutoML tool. Across different datasets, we employed cross-validation to automatically search for optimal parameters.

4.1. Experiments on Optical Images

4.1.1. Dataset Description

In this section, we selected the NWPU-RESISC45 dataset as the experimental subject. The NWPU-RESISC45 dataset is an optical remote sensing image dataset, which contains 45 different types of natural and man-made scenes, such as airports, bridges, forests, lakes, ports, etc., with approximately 700 different images for each type. These images are known for their high resolution and diversity, making them highly suitable for training and testing computer vision algorithms for image recognition, classification, and denoising. In order to conduct our experiments, the dataset was converted into single-channel grayscale images, and Gaussian noise with different variances of 0.01, 0.02, 0.04, 0.06, 0.08, and 0.1 was added to conduct the denoising experiments.

4.1.2. Experiments and Analysis

Noise with a variance of 0.04 is added to the first column of images in Figure 3 and Figure 4 show the denoising results for each comparison algorithm In Figure 4f, it is evident that DD-CNN exhibits the poorest visual performance, characterized by noticeable noise artifacts in the image and significant blur and deformation along the edges of the airplane. While DnCNN managed to recover the overall structure of the noise-free image, it suffered severe blurring in edge details. The lines in the top-left corner of the image were nearly entirely filtered out, and the airplane’s edges in the top-right corner appeared notably blurred. GRN retained some crucial edge information compared to the others but still exhibited noticeable noise residues, resulting in subpar visual quality. In contrast, FFDNet, HSID-CNN, PDSNet, and HLR-DUR achieved comparatively better denoising results, preserving more structural information in the image.
However, it can be seen that among all compared algorithms, HLR-DUR was the only method that clearly and completely preserved the lines in the upper left corner, indicating a significant advantage of HLR-DUR in edge preservation. Additionally, the shape and structure of the three airplanes in the denoising results of the HLR-DUR were closest to the original image. Although some texture information was missing in the denoised result of HLR-DUR, the overall edge structure was the most completely preserved. It is evident that introducing the hierarchical denoising model into the deep learning domain significantly improved the model’s edge preservation ability and denoising effect.
In this section, we evaluate the denoising results of the dataset under different noise levels using MPSNR and MSSIM as evaluation metrics. These metrics were chosen because they are commonly used to assess denoising performance across entire datasets, providing a comprehensive and reliable evaluation of algorithm performance. From the data in Table 1, it is evident that DnCNN, despite its poor visual effect in Figure 4, performed well on both MPSNR and MSSIM metrics, with its MPSNR results being second only to the HLR-DUR. GRN and DD-CNN achieved competitive denoising results at noise variances of 0.01, 0.02, and 0.04, but their MPSNR and MSSIM values dropped faster than other comparative algorithms when the noise pollution level became severe, indicating that these two algorithms may not effectively remove noise under high noise conditions. Meanwhile, FFDNet and HSID-CNN showed more stable performance but did not achieve the best results in numerical evaluation. HLR-DUR achieved the highest metric values in most cases, especially in PSNR. Even with a noise variance of 0.1, the MPSNR value obtained by HLR-DUR remained higher than 30, showing a clear advantage over other algorithms. As for the MSSIM metric, HLR-DUR achieved the best values under different levels of noise, indicating HLR-DUR’s superiority in structure preservation.

4.2. Experiments on Hyperspectral Images

4.2.1. Dataset Description

The AVIRIS Indian Pines dataset was acquired in 1992 by the Airborne Visible/Infrared Imaging Spectrometer (AVIRIS) sensor over the Indian Pines test site in northwestern Indiana. The data size is 145 × 145 × 220, and the pseudo-color image is shown in Figure 5. Several bands in this dataset contain mixed noise, including impulse noise and Gaussian noise.

4.2.2. Experiments and Analysis

Firstly, Figure 6 shows the denoising results of all algorithms on the Pavia city center dataset. FFDNet, HSID-CNN, and GRN all achieved good denoising results for hyperspectral images, while DnCNN caused obvious edge blurring and DD-CNN failed to effectively remove noise. Furthermore, both PDSNet and our proposed HLR-DUR model excelled in removing noise from homogeneous regions while preserving edge information, achieving superior visual outcomes.
Figure 7 shows the auto-correlation curves of the denoised images. The auto-correlation curve depicts the correlation of the image at different lags, reflecting the consistency and repetitive pattern of the image texture. The closer the autocorrelation curve is to that of the original image, the better the denoising algorithm performs in preserving the image texture and structure. This proximity indicates that the algorithm can effectively remove noise while preserving edge details and texture features, which is crucial for maintaining image quality. Although the HLR-DUR model and the PDSNet model showed comparable visual effects, the results in Figure 6 clearly demonstrate that our proposed HLR-DUR model better preserves the details of the original image, as evidenced by its autocorrelation curve closely resembling that of the original image. Other algorithms showed a certain degree of broadening, which is typically the result of over-smoothing. Over-smoothing leads to the loss of unique texture details in the image, creating an unnatural visual effect, which is undesirable in image processing. The broadening observed in the comparative algorithms indicates that while noise was removed, a significant amount of texture information was also discarded. In summary, the analysis of the auto-correlation curves revealed that HLR-DUR exhibited superior ability in preserving the original texture and structure of the image.

4.3. Experiments on SAR Images

4.3.1. Dataset Description

In the experiments of this section, we used the Virtual SAR dataset as the experimental subject. The Virtual SAR dataset is a simulated SAR image dataset, which randomly adds different levels of speckle noise based on the NWPU-RESISC45 dataset. The Virtual SAR dataset consists of 31,500 images, each with a size of 256 × 256. Each image has a corresponding noise-free version and a simulated noise-added version, as shown in Figure 8.

4.3.2. Experiments and Analysis

As shown in Figure 9, taking data number 00072 from the Virtual SAR dataset as an example, the denoising results of various comparative algorithms are presented. Among all the methods, the denoising result obtained by the GRN algorithm still had obvious noise affecting the image quality, while the FFDNet and HSID-CNN algorithms failed to smoothly denoise the homogeneous areas, resulting in some noisy spots in the image. DnCNN, DD-CNN, and PDSNet, which are better at smoothing noise, did not adequately preserve edge information, with the runway lines in the upper left corner of these two images almost completely filtered out. In other algorithms, these lines were partially or completely preserved. Visually, HLR-DUR still achieved the best effect. It is observed that the structural outline of the airplane was clearer in the result of HLR-DUR, and the denoising effect in the homogeneous areas was also better.
To further evaluate the edge preservation capabilities of different denoising algorithms on SAR images, we used the Canny edge detection operator to extract edges from the denoised results of each comparative algorithm, as shown in Figure 10. The Canny edge detection operator is widely used in various image processing and computer vision tasks, including image segmentation, feature extraction, and pattern recognition. Although it is not the most advanced algorithm, due to its reliability, it remains one of the most popular edge detection methods in image processing. In the extracted edge results, it can be seen that due to incomplete noise removal in GRN, its edge extraction was also affected. While other comparative algorithms generally extracted clear and complete target edges, specifically the edge extracted by HLR-DUR was the most complete and closest to the original image edges.
Table 2 shows the numerical evaluation of the denoising performance of various comparative algorithms on five images from the Virtual SAR dataset. Because we were evaluating the denoising performance of SAR images, we employed ENL and EPI as evaluation metrics to assess the despeckling and edge preservation capabilities. It can be seen that on the SAR dataset, the HLR-DUR model performed well on both ENL and EPI metrics, meaning it effectively maintains the image’s edges and details while reducing noise. Other algorithms such as DnCNN and DD-CNN exhibited poorer performance in edge preservation compared to FFDNet, HSID-CNN, GRN, and PSDNet. While FFDNet, HSID-CNN, GRN, and PSDNet also achieved respectable results in terms of ENL and EPI, overall, there remains a gap compared to HLR-DUR.
In Figure 11, we present the denoising results obtained from a real SAR dataset, SARBuD. The SARBuD dataset is a synthetic aperture radar dataset designed for benchmarking despeckling algorithms. It includes a variety of simulated SAR images with different noise levels, providing a comprehensive resource for evaluating the performance of image-denoising techniques. It can be observed that FFDNet, HSID-CNN, and GRN algorithms exhibit suboptimal performance with residual noise remaining in uniform regions. While DnCNN, DD-CNN, and PDSNet demonstrate relatively effective denoising, they tend to blur or even filter out image edges. HLR-DUR achieved the best visual results among the methods evaluated.
As shown in Table 3, it is evident that the HLR-DUR model consistently outperformed other algorithms when faced with real SAR image datasets. This indicates that it achieves superior evaluation metrics in most scenarios, suggesting its capability to effectively balance denoising and edge preservation in real SAR images.
The HLR-DUR model showed significant advantages through many evaluation metrics. By conducting a comprehensive analysis of the denoising results for virtual SAR images and real SAR images, it can be seen that compared to other commonly used denoising algorithms such as DnCNN, FFDNet, HSID-CNN, GRN, DD-CNN, and PDSNet, HLR-DUR more effectively preserves the image’s edge parts and detail information while reducing noise levels. Especially in dealing with complex and diverse image denoising problems, HLR-DUR demonstrated its comprehensive advantages, achieving the goal of edge preservation and denoising, and holds great potential in the field of image denoising.

4.4. Processing Speed and Ablation Experiments

4.4.1. Ablation Experiment and Accuracy and Loss Curves

Figure 12 illustrates the training and validation loss results of HLR-DUR, demonstrating its rapid convergence and minimal loss values, indicating effective denoising performance.
We conducted ablation experiments on the deep model. The primary objective of these experiments was to assess the impact of the edge modules incorporated within the model. The results presented in Table 4 unequivocally demonstrate that integrating the edge module significantly improved both the denoising efficacy and the model’s ability to preserve edges. This outcome robustly validates the effectiveness of the edge module in enhancing overall model performance.

4.4.2. Results of the Processing Speed and Parameters Experiment

We employed four NVIDIA® GeForce RTX 3090 GPUs which are manufactured by NVIDIA Corporation, a technology company based in Santa Clara, California, United States.to evaluate the processing speed and parameter count of training models using 256 × 256 SAR images as input. Table 5 presents the parameter sizes and processing details of seven models. Among them, the model with the largest parameter count was the enhanced DD-CNN, which totaled 7,955,552 parameters. The model requiring the longest computational time was PDSNet, with an approximate execution time of 76.26 ms. Our proposed HLR-DUR model had 28,311 parameters and took about 1.2 ms per epoch to compute.

5. Conclusions

In this paper, a new deep low-rank model for remote sensing image denoising named HLR-DUR was proposed. Compared with existing denoising methods, HLR-DUR combines the advantages of the traditional LRR model and deep learning, enabling it to extract edge information using prior knowledge regulations while further enhancing the model’s denoising performance and computational efficiency through the deep autoencoder. HLR-DUR combines the shallow low-rank denoising model with an autoencoder to construct a two-level deep encoder. It uses deep unfolding to extract both the low-rank and edge components. By integrating traditional iterative algorithms with the deep learning framework, the iterative steps are unfolded into multiple layers, which are then aggregated to obtain the final denoised result. Experiments conducted on three types of remote sensing datasets, optical remote sensing images, hyperspectral images, and SAR images, demonstrated that HLR-DUR achieves better edge-preserving denoising results than SOTA models and significantly improves the denoising effect.

Author Contributions

Conceptualization, F.S. and T.Z.; methodology, F.S. and S.T.; software, X.F. and T.Z.; Data collection, X.F.; validation, F.S. and X.F.; investigation, S.T.; writing—original draft preparation, T.Z.; writing—review and editing, F.S. and S.T.; supervision, F.S. and S.T.; project administration, F.S.; funding acquisition, F.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The original data presented in the study are openly available in [IEEEDataport and Puedue University Research Repository] at [https://ieee-dataport.org/open-access/virtual-sar-synthetic-dataset-deep-learning-based-speckle-noise-reduction-algorithms and https://purr.purdue.edu/publications/1947/1].

Acknowledgments

The authors would like to thank the Editor and the anonymous Referees for their instructive comments.

Conflicts of Interest

Author Fanqi Shao was employed by the company China Electronics Technology Group Corporation. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Sabins, F.F., Jr.; Ellis, J.M. Remote Sensing: Principles, Interpretation, and Applications; Waveland Press: Long Grove, IL, USA, 2020. [Google Scholar]
  2. Rasti, B.; Chang, Y.; Dalsasso, E.; Denis, L.; Ghamisi, P. Image restoration for remote sensing: Overview and toolbox. IEEE Geosci. Remote Sens. Mag. 2021, 10, 201–230. [Google Scholar] [CrossRef]
  3. Li, G.; Gu, X.; Chen, C.; Zhou, C.; Xiao, D.; Wan, W.; Cai, H. Low-Frequency Magnetotelluric Data Denoising Using Improved Denoising Convolutional Neural Network and Gated Recurrent Unit. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5909216. [Google Scholar] [CrossRef]
  4. Li, G.; Wu, S.; Cai, H.; He, Z.; Liu, X.; Zhou, C.; Tang, J. IncepTCN: A new deep temporal convolutional network combined with dictionary learning for strong cultural noise elimination of controlled-source electromagnetic data. Geophysics 2023, 88, E107–E122. [Google Scholar] [CrossRef]
  5. Han, L.; Zhao, Y.; Lv, H.; Zhang, Y.; Liu, H.; Bi, G. Remote sensing image denoising based on deep and shallow feature fusion and attention mechanism. Remote Sens. 2022, 14, 1243. [Google Scholar] [CrossRef]
  6. Huang, Z.; Zhu, Z.; Wang, Z.; Shi, Y.; Fang, H.; Zhang, Y. DGDNet: Deep gradient descent network for remotely sensed image denoising. IEEE Geosci. Remote Sens. Lett. 2023, 20, 6002405. [Google Scholar] [CrossRef]
  7. Tian, C.; Fei, L.; Zheng, W.; Xu, Y.; Zuo, W.; Lin, C.-W. Deep learning on image denoising: An overview. Neural Netw. 2020, 131, 251–275. [Google Scholar] [CrossRef] [PubMed]
  8. Jiao, L.; Zhao, J. A survey on the new generation of deep learning in image processing. IEEE Access 2019, 7, 172231–172263. [Google Scholar] [CrossRef]
  9. Feng, X.; Zhang, W.; Su, X.; Xu, Z. Optical remote sensing image denoising and super-resolution reconstructing using optimized generative network in wavelet transform domain. Remote Sens. 2021, 13, 1858. [Google Scholar] [CrossRef]
  10. Huang, Z.; Zhang, Y.; Li, Q.; Li, Z.; Zhang, T.; Sang, N.; Xiong, S. Unidirectional variation and deep CNN denoiser priors for simultaneously destriping and denoising optical remote sensing images. Int. J. Remote Sens. 2019, 40, 5737–5748. [Google Scholar] [CrossRef]
  11. Xie, Y.; Qu, Y.; Tao, D.; Wu, W.; Yuan, Q.; Zhang, W. Hyperspectral image restoration via iteratively regularized weighted schatten $ p $-norm minimization. IEEE Trans. Geosci. Remote Sens. 2016, 54, 4642–4659. [Google Scholar] [CrossRef]
  12. Dong, W.; Wang, H.; Wu, F.; Shi, G.; Li, X. Deep spatial–spectral representation learning for hyperspectral image denoising. IEEE Trans. Comput. Imaging 2019, 5, 635–648. [Google Scholar] [CrossRef]
  13. Gu, F.; Zhang, H.; Wang, C. A two-component deep learning network for SAR image denoising. IEEE Access 2020, 8, 17792–17803. [Google Scholar] [CrossRef]
  14. Feng, X.; Tian, S.; Abhadiomhen, S.E.; Xu, Z.; Shen, X.; Wang, J.; Zhang, X.; Gao, W.; Zhang, H.; Wang, C. Edge-Preserved Low-Rank Representation via Multi-Level Knowledge Incorporation for Remote Sensing Image Denoising. Remote Sens. 2023, 15, 2318. [Google Scholar] [CrossRef]
  15. Chen, Y.; Huang, T.-Z.; He, W.; Zhao, X.-L.; Zhang, H.; Zeng, J. Hyperspectral image denoising using factor group sparsity-regularized nonconvex low-rank approximation. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5515916. [Google Scholar] [CrossRef]
  16. He, W.; Zhang, H.; Zhang, L.; Shen, H. Hyperspectral image denoising via noise-adjusted iterative low-rank matrix approximation. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2015, 8, 3050–3061. [Google Scholar] [CrossRef]
  17. Ma, T.-H.; Xu, Z.; Meng, D. Remote sensing image denoising via low-rank tensor approximation and robust noise modeling. Remote Sens. 2020, 12, 1278. [Google Scholar] [CrossRef]
  18. Zhang, H.; Liu, L.; He, W.; Zhang, L. Hyperspectral image denoising with total variation regularization and nonlocal low-rank tensor decomposition. IEEE Trans. Geosci. Remote Sens. 2019, 58, 3071–3084. [Google Scholar] [CrossRef]
  19. Li, C.; Ma, Y.; Huang, J.; Mei, X.; Ma, J. Hyperspectral image denoising using the robust low-rank tensor recovery. J. Opt. Soc. Am. A 2015, 32, 1604–1612. [Google Scholar] [CrossRef] [PubMed]
  20. Guan, D.; Xiang, D.; Tang, X.; Kuang, G. SAR image despeckling based on nonlocal low-rank regularization. IEEE Trans. Geosci. Remote Sens. 2019, 57, 3472–3489. [Google Scholar] [CrossRef]
  21. Chen, G.; Li, G.; Liu, Y.; Zhang, X.-P.; Zhang, L. SAR image despeckling based on combination of fractional-order total variation and nonlocal low rank regularization. IEEE Trans. Geosci. Remote Sens. 2020, 58, 2056–2070. [Google Scholar] [CrossRef]
  22. Geng, J.; Fan, J.; Ma, X.; Wang, H.; Cao, K. An iterative low-rank representation for SAR image despeckling. In Proceedings of the 2016 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Beijing, China, 10–15 July 2016; pp. 72–75. [Google Scholar]
  23. Baier, G.; He, W.; Yokoya, N. Robust nonlocal low-rank SAR time series despeckling considering speckle correlation by total variation regularization. IEEE Trans. Geosci. Remote Sens. 2020, 58, 7942–7954. [Google Scholar] [CrossRef]
  24. Zhang, H.; Chen, H.; Yang, G.; Zhang, L. LR-Net: Low-rank spatial-spectral network for hyperspectral image denoising. IEEE Trans. Image Process. 2021, 30, 8743–8758. [Google Scholar] [CrossRef]
  25. Nguyen, H.V.; Ulfarsson, M.O.; Sigurdsson, J.; Sveinsson, J.R. Deep sparse and low-rank prior for hyperspectral image denoising. In Proceedings of the IGARSS 2022–2022 IEEE International Geoscience and Remote Sensing Symposium, Kuala Lumpur, Malaysia, 17–22 July 2022; pp. 1217–1220. [Google Scholar]
  26. Sun, H.; Liu, M.; Zheng, K.; Yang, D.; Li, J.; Gao, L. Hyperspectral image denoising via low-rank representation and CNN denoiser. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 15, 716–728. [Google Scholar] [CrossRef]
  27. Zhang, Q.; Yuan, Q.; Song, M.; Yu, H.; Zhang, L. Cooperated spectral low-rankness prior and deep spatial prior for HSI unsupervised denoising. IEEE Trans. Image Process. 2022, 31, 6356–6368. [Google Scholar] [CrossRef]
  28. Xiong, K.; Zhao, G.; Wang, Y.; Shi, G. SAR imaging and despeckling based on sparse, low-rank, and deep CNN priors. IEEE Geosci. Remote Sens. Lett. 2021, 19, 4501205. [Google Scholar] [CrossRef]
  29. Monga, V.; Li, Y.; Eldar, Y.C. Algorithm unrolling: Interpretable, efficient deep learning for signal and image processing. IEEE Signal Process. Mag. 2021, 38, 18–44. [Google Scholar] [CrossRef]
  30. Passah, A.; Amitab, K.; Kandar, D. SAR image despeckling using deep CNN. IET Image Process. 2021, 15, 1285–1297. [Google Scholar] [CrossRef]
  31. Wang, S.; Wang, X.; Zhang, L.; Zhong, Y. Deep low-rank prior for hyperspectral anomaly detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5527017. [Google Scholar] [CrossRef]
  32. Candès, E.J.; Li, X.; Ma, Y.; Wright, J. Robust principal component analysis? J. ACM 2011, 58, 11. [Google Scholar] [CrossRef]
  33. Dong, Y.; Chan, R.H.; Xu, S. A detection statistic for random-valued impulse noise. IEEE Trans. Image Process. 2007, 16, 1112–1120. [Google Scholar] [CrossRef]
  34. Lopes, A.; Touzi, R.; Nezry, E. Adaptive speckle filters and scene heterogeneity. IEEE Trans. Geosci. Remote Sens. 1990, 28, 992–1000. [Google Scholar] [CrossRef]
  35. Hong, M.; Luo, Z.-Q. On the linear convergence of the alternating direction method of multipliers. Math. Program. 2017, 162, 165–199. [Google Scholar] [CrossRef]
  36. Huang, Z.; Wang, Z.; Zhu, Z.; Zhang, Y.; Fang, H.; Shi, Y.; Zhang, T. DLRP: Learning deep low-rank prior for remotely sensed image denoising. IEEE Geosci. Remote Sens. Lett. 2022, 19, 6508905. [Google Scholar] [CrossRef]
  37. Zhang, K.; Zuo, W.; Zhang, L. FFDNet: Toward a fast and flexible solution for CNN-based image denoising. IEEE Trans. Image Process. 2018, 27, 4608–4622. [Google Scholar] [CrossRef] [PubMed]
  38. Yuan, Q.; Zhang, Q.; Li, J.; Shen, H.; Zhang, L. Hyperspectral image denoising employing a spatial–spectral deep residual convolutional neural network. IEEE Trans. Geosci. Remote Sens. 2018, 57, 1205–1218. [Google Scholar] [CrossRef]
  39. Cao, X.; Fu, X.; Xu, C.; Meng, D. Deep spatial-spectral global reasoning network for hyperspectral image denoising. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5504714. [Google Scholar] [CrossRef]
  40. Shan, W.; Liu, P.; Mu, L.; Cao, C.; He, G. Hyperspectral image denoising with dual deep CNN. IEEE Access 2019, 7, 171297–171312. [Google Scholar] [CrossRef]
  41. Lin, C.; Qiu, C.; Jiang, H.; Zou, L. A Deep Neural Network Based on Prior-Driven and Structural Preserving for SAR Image Despeckling. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2023, 16, 6372–6392. [Google Scholar] [CrossRef]
Figure 1. The flowchart of the proposed method.
Figure 1. The flowchart of the proposed method.
Sensors 24 04574 g001
Figure 2. The structure of the HLR-DUR model.
Figure 2. The structure of the HLR-DUR model.
Sensors 24 04574 g002
Figure 3. NWPU-RESISC45 dataset.
Figure 3. NWPU-RESISC45 dataset.
Sensors 24 04574 g003
Figure 4. The denoising results of each comparative algorithm in Figure 3.
Figure 4. The denoising results of each comparative algorithm in Figure 3.
Sensors 24 04574 g004
Figure 5. AVIRIS Indian Pines dataset (R: 29, G: 42, B: 89).
Figure 5. AVIRIS Indian Pines dataset (R: 29, G: 42, B: 89).
Sensors 24 04574 g005
Figure 6. The denoising results of all algorithms on the Pavia city center dataset.
Figure 6. The denoising results of all algorithms on the Pavia city center dataset.
Sensors 24 04574 g006
Figure 7. The auto-correlation curves of the denoised images.
Figure 7. The auto-correlation curves of the denoised images.
Sensors 24 04574 g007
Figure 8. The samples of the Virtual SAR dataset.
Figure 8. The samples of the Virtual SAR dataset.
Sensors 24 04574 g008
Figure 9. Denoised results of the Virtual SAR dataset.
Figure 9. Denoised results of the Virtual SAR dataset.
Sensors 24 04574 g009
Figure 10. Comparison of the edge extraction results after denoising by comparative algorithms.
Figure 10. Comparison of the edge extraction results after denoising by comparative algorithms.
Sensors 24 04574 g010
Figure 11. Denoised results of the SARBuD dataset.
Figure 11. Denoised results of the SARBuD dataset.
Sensors 24 04574 g011aSensors 24 04574 g011b
Figure 12. Training accuracy and loss curves.
Figure 12. Training accuracy and loss curves.
Sensors 24 04574 g012
Table 1. Evaluation of denoising results on the NWPU-RESISC45 dataset.
Table 1. Evaluation of denoising results on the NWPU-RESISC45 dataset.
Noise VarianceEvaluation MetricsDnCNNFFDNetHSID-CNNGRNDD-CNNPDSNetHLR-DUR
0.01MPSNR33.856229.155825.169623.465524.726828.456736.8927
MSSIM0.76270.79240.72480.70850.73760.70730.7706
0.02MPSNR31.955727.577825.002922.998122.952725.453936.4328
MSSIM0.72500.71760.72220.67490.67780.70180.7625
0.04MPSNR29.755626.026124.427621.177419.592323.584635.6956
MSSIM0.67720.64450.71060.45520.39930.69360.7433
0.06MPSNR28.363725.192823.843917.416916.855620.397734.9296
MSSIM0.65620.58380.66510.23290.19860.57250.6706
0.08MPSNR27.716424.197923.134117.325714.892818.263332.9710
MSSIM0.63660.51860.59730.23120.12500.55910.6290
0.1MPSNR27.059223.404922.333214.586214.296318.036130.8986
MSSIM0.55750.46810.52140.14910.10990.48260.5928
Table 2. Evaluation of denoising results on the Virtual SAR dataset.
Table 2. Evaluation of denoising results on the Virtual SAR dataset.
Noisy
Image
DnCNNFFDNetHSID-
CNN
GRNDD-CNNPDSNetHLR-DUR
Image 1ENL5.5820103.5885118.0177114.4488125.616498.0918116.3475125.9335
EPI-0.58390.66740.69370.70220.58030.63090.7356
Image 2ENL9.321396.228595.7900101.6122104.617384.0133107.3329115.3813
EPI-0.52360.59370.67960.67250.61160.56820.6823
Image 3ENL7.6195107.8969116.3891113.5986116.4556100.3012114.2744124.5229
EPI-0.63650.87790.87270.89910.60130.89170.9164
Image 4ENL7.792776.631073.961280.832880.938160.464683.653481.9254
EPI-0.59290.73980.86340.89290.70130.79720.9147
Image 5ENL8.8592103.7392109.3430116.1278110.9981101.4649117.5947125.6757
EPI-0.70300.85620.85190.82440.77780.81390.8662
Table 3. Evaluation of denoising results on the SARBuD dataset.
Table 3. Evaluation of denoising results on the SARBuD dataset.
Noisy
Image
DnCNNFFDNetHSID-
CNN
GRNDD-CNNPDSNetHLR-DUR
Image 1ENL10.772767.441555.939054.130257.715958.712765.989272.2878
EPI-0.42010.54130.52650.51590.53370.57120.6175
Image 2ENL12.696462.511067.880363. 427663.524569.712577. 063692.3050
EPI-0.65210.70190.70600.82260.67150.80640.8626
Image 3ENL10.640856.497063.185970.137766.988954.073582.401078.4357
EPI-0.63610.718450.64230.72230.65310.73120.7907
Table 4. Results of the ablation experiments.
Table 4. Results of the ablation experiments.
Evaluation Metrics0.010.020.040.060.080.1
Without Edge ModuleMPSNR35.392735.023334.620133.293031.922329.1146
MSSIM0.61110.60320.57250.55490.70220.5803
HLR-DURMPNSR36.892736.432835.695634.929632.971030.8986
MSSIM0.70230.68820.66200.63210.62900.6116
Table 5. Different model training parameters and time consumption comparison.
Table 5. Different model training parameters and time consumption comparison.
MethodsDnCNNFFDNetHSID-
CNN
GRNDD-CNNPDSNetHLR-DUR
Time(ms)22.5150.1347.1367.7863.1676.261.2
Parameters667,008485,316556,0971,322,2517,955,5521,366,11128,311
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Shao, F.; Feng, X.; Tian, S.; Zhang, T. A Hierarchical Low-Rank Denoising Model for Remote Sensing Images Based on Deep Unfolding. Sensors 2024, 24, 4574. https://doi.org/10.3390/s24144574

AMA Style

Shao F, Feng X, Tian S, Zhang T. A Hierarchical Low-Rank Denoising Model for Remote Sensing Images Based on Deep Unfolding. Sensors. 2024; 24(14):4574. https://doi.org/10.3390/s24144574

Chicago/Turabian Style

Shao, Fanqi, Xiaolin Feng, Sirui Tian, and Tianyi Zhang. 2024. "A Hierarchical Low-Rank Denoising Model for Remote Sensing Images Based on Deep Unfolding" Sensors 24, no. 14: 4574. https://doi.org/10.3390/s24144574

APA Style

Shao, F., Feng, X., Tian, S., & Zhang, T. (2024). A Hierarchical Low-Rank Denoising Model for Remote Sensing Images Based on Deep Unfolding. Sensors, 24(14), 4574. https://doi.org/10.3390/s24144574

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop