Next Article in Journal
Person Recognition Based on Deep Gait: A Survey
Previous Article in Journal
Assessment of a UWB Real Time Location System for Dairy Cows’ Monitoring
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Compressed Imaging Reconstruction Based on Block Compressed Sensing with Conjugate Gradient Smoothed l0 Norm

1
School of Optics and Photonics, Beijing Institute of Technology, Beijing 100081, China
2
MOE Key Laboratory of Optoelectronic Imaging Technology and System, Beijing Institute of Technology, Beijing 100081, China
3
Luoyang Electro Optic Equipment Research Institute of AVIC, Luoyang 471000, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(10), 4870; https://doi.org/10.3390/s23104870
Submission received: 29 March 2023 / Revised: 11 May 2023 / Accepted: 16 May 2023 / Published: 18 May 2023
(This article belongs to the Section Sensing and Imaging)

Abstract

:
Compressed imaging reconstruction technology can reconstruct high-resolution images with a small number of observations by applying the theory of block compressed sensing to traditional optical imaging systems, and the reconstruction algorithm mainly determines its reconstruction accuracy. In this work, we design a reconstruction algorithm based on block compressed sensing with a conjugate gradient smoothed l 0 norm termed BCS-CGSL0. The algorithm is divided into two parts. The first part, CGSL0, optimizes the SL0 algorithm by constructing a new inverse triangular fraction function to approximate the l 0 norm and uses the modified conjugate gradient method to solve the optimization problem. The second part combines the BCS-SPL method under the framework of block compressed sensing to remove the block effect. Research shows that the algorithm can reduce the block effect while improving the accuracy and efficiency of reconstruction. Simulation results also verify that the BCS-CGSL0 algorithm has significant advantages in reconstruction accuracy and efficiency.

1. Introduction

With the development of information technology and the continuous improvement of related requirements in the military and civilian fields, higher standards for the imaging resolution of optical systems have been put forward. The traditional optical imaging system improves resolution mainly by increasing the focal length and aperture of the optical system, reducing the size of detector pixels, and increasing the number of pixels. In practical engineering, the aperture and focal length of optical systems are challenging to improve and limited by observation conditions with stricter requirements. At present, the size of the detector pixels is close to the physical limit, making subsequent improvements difficult. Moreover, increasing the number of detector pixels will increase power consumption, volume, weight, and data storage and processing complexity for the system [1,2]. Therefore, seeking a new image data acquisition and processing method is necessary to improve the imaging system’s resolution.
The emergence of compressed sensing (CS) [3,4] theory provides a new idea for improving the resolution of optical imaging systems. According to CS theory, signals with sparse characteristics can be compressed and projected into a low-dimensional space by a specific observation matrix to obtain a small number of projected signals that can reconstruct the original signal using the corresponding reconstruction algorithm. Compressed imaging super-resolution reconstruction technology applies CS theory to traditional optical systems to achieve super-resolution reconstruction and improve the resolution of imaging systems. This reconstructs higher-resolution images with a small number of observation results. This method has excellent application prospects, reducing the information for image storage, transmission, and processing in optical systems and the requirements of the detector. It transfers the difficulty of improving the system resolution from the hardware to the software side, thus decreasing the equipment cost.
The observation matrix and the reconstruction algorithm determine the reconstruction accuracy of the compressed imaging super-resolution reconstruction technique. The reconstruction algorithm is the primary determinant factor. At present, reconstruction algorithms can be broadly divided into three directions, including the minimization of the l 0 , l 1 , and l p norms. The solving model of the reconstruction algorithm, based on the l 0 norm minimization, is min s 0 , s . t . y = A s , which is an intractable NP−hard problem [5] and must be transformed. The standard methods include the greedy [6] and SL0 algorithms [7]. The reconstruction algorithm, based on the l 1 norm minimization, uses the l 1 norm to approximate the l 0 norm, whose solution model is min s 1 , s . t . y = A s . The common methods include the Basis Pursuit algorithm [8] and the Split-Bregman algorithm [9]. The reconstruction algorithm based on the l p norm minimization uses the l p norm to approximate the l 0 norm. The solution model is min s p , s . t . y = A s . The commonly used algorithms include the iterative reweighted least squares [10] and the FOCUSS algorithms [11].
Among the three types of algorithms mentioned above, the algorithm based on the minimum l 1 norm that introduced the convex programming methods has excellent reconstruction precision. However, its time complexity is much larger than others, with refactoring needed over a long time. At the same time, this method requires more observations because the l 1 norm cannot accurately represent the signal’s sparsity. The algorithm based on the minimum l p norm reduces the number of required observations, but the reconstruction time is still long, and the practicability is not higher. The algorithm based on the minimum l 0 norm is divided into the greedy and SL0 algorithms. The greedy algorithm requires a small number of calculations and has a fast reconstruction speed, but the reconstruction accuracy is poor, and there are reconstruction errors.
The SL0 algorithm, as a reconstruction algorithm based on the l 0 norm minimization, approximates the l 0 norm through the constructor, combines the idea of convex programming with the greedy algorithm, and greatly reduces the reconstruction time while ensuring a high reconstruction accuracy. Therefore, this paper focuses on the improvement of the SL0 algorithm.
Based on the SL0 algorithm, this paper adopts the inverse trigonometric fraction function approximation to express the l 0 norm and combines the modified conjugate gradient method to improve the SL0 algorithm, which is called the CGSL0 (Conjugate Gradient Smooth l0 Norm) algorithm. On this basis, to eliminate the image block effect after reconstruction, this paper proposes a BCS reconstruction algorithm based on the CGSL0 and BCS-SPL [12] algorithms termed the BCS-CGSL0 (Block Compressed Sensing with Conjugate Gradient Smooth l 0 Norm) algorithm, combining the CGSL0 and Smoothed Projected Landweber for image reconstruction.

2. Background

2.1. Compressed Sensing Theory

The theory of compressed sensing projects a signal with sparse characteristics onto a low-dimensional space to realize the sampling and compression of the signal at the same time [13]. It accurately reconstructs the original signal with fewer sample values, and its basic model is as follows:
y = Φ x
where y R m represents the result of compressed sampling, Φ R m × n represents the observation matrix, and x R n represents the original signal. The crux of the problem of CS is how to reconstruct the original signal x accurately from the compressed signal y.
The signal x in CS theory must be sparse [14]. In this way, the observed values after a random sub-sampling of the observation matrix can break the Nyquist sampling theorem and be reconstructed accurately. Most natural signals in the time domain are non-sparse but may be sparse in some transformation domains. That is to say that non-sparse signals can be converted into sparse signals using the sparse transformation basis [15]:
x = Ψ s
where Ψ R n × n is the sparse transformation basis and s R n is the signal after sparse transformation. At this point, the basic model of CS can be expressed as follows:
y = Φ x = Φ Ψ s = A s
where A = Φ Ψ is the perception matrix that is the product of the observation matrix and the sparse basis; this model meets the requirements of CS theory. Sparse signals s can be obtained using the reconstruction algorithm, and then reconstructed signals x can be obtained using the sparse transformation x = Ψ s , which is the reconstruction process of compressed sensing.
The common sparse transformation basis includes the discrete Fourier transform basis (DFT) [16], discrete cosine transform basis (DCT) [17], and discrete wavelet transform basis (DWT) [18]. The DCT can reflect the correlation of image signals, so it is mainly used for sparse representations of image signals.
In (3), this paper takes the random Gaussian measurement matrix as the observation matrix and the discrete cosine transform basis as the sparse basis to verify the performance of these reconstruction algorithms under the same conditions.

2.2. Compressed Imaging System and Reconstruction

Different from traditional image compression, which compresses and encodes complete image data, the compressed imaging system combines compressed sensing theory into the traditional optical imaging system and collects incomplete data. In this system, a coding mask is added to the imaging focal plane to encode the target scene, which is then collected by a low-resolution detector. The reconstruction algorithm is used to reconstruct the low-resolution images observed many times to obtain the corresponding high-resolution target scene. Its optical imaging system is shown in Figure 1. The imaging principle can be summarized as a light emphasis on the encoding mask and a down-sampling of the detector. Finally, the low-resolution image is the obtained observation value that adds detector noise. It can be reconstructed using the theory in Section 2.1.
The original signal x in (3) is a one−dimensional signal. This must be converted when applied to the compressed imaging system in Figure 1.
For a two−dimensional image signal of size M × M and a detector of size N × N , the single sampling process is as follows:
  • Coding modulation. The coding board used in this system is a 0–1 coding board. The 0 position corresponds to no light, and the 1 position corresponds to light.
  • down-sampling. The size of the image block of each pixel of the detector detected is B × B , where B = N / M , and the sum of the corresponding image block signals is the detection value of the detector.
It is expanded into a one−dimensional column vector x R B 2 × 1 for the image block and into a one−dimensional row vector ψ R 1 × B 2 for the corresponding block of the encoding mask. We discover that y = ϕ x represents a single sampling process of the corresponding image block position. Each image block uses the same encoding mask block, and all image encoding processes can be expressed as follows:
y i = ϕ x i , i = 1 , 2 , , M 2
where y i R 1 is the observation value of the i th detector pixel.
The result of sampling the compression imaging system p B 2 times, using a different coding mask each time, is as follows:
Y i = Φ X i , i = 1 , 2 , , M 2
where Y i R p B 2 × 1 are the observations of the i th detector pixel, X i R B 2 × 1 is the column vector of the image block signal, and Φ R p B 2 × B 2 is the observation matrix. In this case, the compressed sensing reconstruction algorithm can be used for super-resolution from low-resolution images obtained by multiple sampling into high-resolution images. This is because the model in (5) meets the compressed sensing theoretical model in (3). This process is the multi-image super-resolution reconstruction technology based on compressed sensing theory.

2.3. BCS-SPL Algorithm

As can be seen in Section 2.2, the core of the compressed imaging reconstruction lies in block compressed sensing (BCS) [19]. The idea of the block also brings some adverse effects, such as the block effect in the reconstructed image. The block effect of the image under a DCT transformation is pronounced. Therefore, [12] proposed the Smooth Projection Landweber (SPL) algorithm, which deals with the image’s block effect and noise. The core idea of the SPL algorithm is two−step iterations starting from the initial sparseness on the transform domain. The process is as follows:
θ [ n + 1 ] = θ [ n ] + 1 ζ Ψ 1 Φ T y Φ Ψ θ [ n ]
θ ^ [ n + 1 ] = θ [ n + 1 ] , θ [ n + 1 ] τ 0 , o t h e r w i s e
where θ [ n ] is the reconstructed signal in the n th iteration, ζ is the largest singular value of Φ Φ T , and τ = λ 2 l o g e N M E D θ ^ [ n ] 0.6754 is the iteration threshold.
Mun S. et al. [12] proposed the reconstruction algorithm called BCS-SPL by applying the SPL algorithm to CS theory, which can remove block effects and impose smoothness. The BCS-SPL algorithm has been verified to have a good reconstruction effect, and its steps are as follows:
  • Obtain the initial solution as x i [ 0 ] = Φ 1 y ;
  • Smooth the reconstructed signal of the n th iteration as x i [ n ] = w i e n e r ( x i [ n ] ) , where w i e n e r ( ) is the Wiener filtering;
  • Calculate according to (6) as x i ^ [ n ] = x i [ n ] + Φ T y i Φ x i [ n ] ;
  • Transform the reconstructed signal to thw Ψ domain and obtain s i ^ [ n ] = Ψ 1 x i ^ [ n ] ;
  • Calculate according to (7) as
    s i ^ [ n ] = s i ^ [ n ] , s i ^ [ n ] τ 0 , o t h e r w i s e ;
  • Transform s i ^ [ n ] back to the spatial domain x i ^ [ n ] = Ψ s i ^ [ n ] ;
  • Calculate according to (6) as x i [ n + 1 ] = x i ^ [ n + 1 ] + Φ T y i Φ x i ^ [ n ] ;
  • If the termination condition D [ n + 1 ] D [ n ] < 10 4 is satisfied, where
    D [ n ] = x i [ n + 1 ] x i [ n ] 2 / N
    then the iteration terminates. Otherwise, go to Step 2.

2.4. SL0 Algorithm

The solution of (3) is essentially to reconstruct the original signal with the least number of non-zeros based on the observation value and the perception matrix to obtain the most sparse solution. After that, the original signal in the natural domain can be obtained by sparse transformation, and the CS algorithm can be summarized as follows:
s = min s 0 , s . t . y = A s
x = Ψ s
where s 0 = i = 1 n s i 0 is the l 0 norm of s, which represents the number of non-zero elements in the vector and the sparsity of the vector. The sparsest solution can be obtained by solving its minimum value. However, it can be seen that this is an NP-hard problem and difficult to solve when n is large.
Using a smooth continuous function to approximate the l 0 norm of the original sparse signal discontinued was proposed for the SL0 algorithm by Mohimani et al. [7]. The smoothing function adopted is the standard Gaussian function:
f σ s i = 1 e s i 2 2 σ 2
It is easier for us to derive the following:
lim σ 0 f σ s i = 0 , s i = 0 1 , s i 0
After bringing F s = i = 1 n f σ s i into (9), the approximation of the l 0 norm can be obtained as follows:
s 0 = i = 1 n s i 0 n lim σ 0 F σ s
where σ is the smoothness factor, whose value determines the smoothness degree of F σ ( s ) . The larger σ is, the smoother F σ ( s ) is, and the lower the degree of approximation to the l 0 norm is. Otherwise, the l 0 norm of the signal s can be approximated by (10) when σ is approximately equal to zero.
The model in (8) can be transformed into an optimization problem to solve continuous signals. On this basis, the fastest descent method and gradient projection principle are used to gradually approach the optimal solution of the continuous function through several iterations.
The search direction of the optimal value is expressed as follows:
d = F σ ( s ) = s 1 e s 1 2 2 σ 2 , s 2 e s 2 2 2 σ 2 , , s n e s n 2 2 σ 2
The effect called “sawtooth” will appear in the search for the optimal value of the fastest descent method, which causes the global optimal value not to be obtained and the estimation accuracy of the l 0 norm to reduce. For these reasons, the algorithm needs to be improved.

3. Materials and Methods

3.1. Construct an Approximate Estimation Function of the l 0 Norm

The accuracy of the approximate l 0 norm is the key factor for improving the effectiveness of the SL0 algorithm. The closer the continuous smoothing function constructed is to the l 0 norm, the more accurate the result of the algorithm reconstruction will be. In this paper, the CGSL0 algorithm uses the following inverse trigonometric fraction function. It is proposed as the approximate estimation function of the l 0 norm:
f σ ( s i ) = 4 π arctan s i 2 s i 2 + ρ σ 2
where ρ and σ are the parameters of controlling the steepness of the smoothing function and s i is the component of the sparse vector s. Therefore, the l 0 norm is approximately expressed as follows:
s 0 = i = 1 n s i 0 lim σ 0 F σ ( s ) = lim σ 0 i = 1 n f σ ( s i )
As shown in Figure 2, the Gaussian function, hyperbolic tangent function, and compound trigonometric function are, respectively, used to approximate the SL0 algorithm [7], NSL0 algorithm [20], and DNSL0 algorithm [21]. It can be observed that in the case of ρ = 0.05 and σ = 0.01 , for the interval [−0.1 0.1], the inverse trigonometric fraction function proposed in this paper is steeper than other algorithms and has a better approximation to the l 0 norm; that is, the sparsity of the signals can be expressed more accurately. It can improve the accuracy of the optimization algorithm.

3.2. CGSL0 Algorithm

The SL0 algorithm has a “sawtooth” effect using the Gaussian function to approximate the l 0 norm. The NSL0 algorithm is solved by the modified Newton method. It needs to calculate the first- and second-order derivatives of the iterative points. At the same time, the Hesse matrix of the objective function must be positive definite, which has high requirements for the objective function. The inverse matrix of the second-order Hesse matrix requires a large amount of calculation, which presents some problems for practical use.
The conjugate gradient method is an unconstrained optimization method between the fastest descent method and the Newton method. It has a superlinear convergence speed, a simple algorithm structure, and easy programming implementation. This method only uses the first derivative to avoid the calculation of the second derivative, reducing the amount of calculation and storage, much like the fastest descent method. For the sake of the above, we propose an algorithm combining the smoothing function and the conjugate gradient method in Section 3.1, BCS-CGSL0.
According to (14), the reconstruction model of the CS algorithm is as follows:
s = min i = 1 n 4 π arctan s i 2 s i 2 + ρ σ 2 , s . t . y = A s
The solution can be divided into two steps:
  • Calculate the iteration direction using the conjugate gradient method and search for the optimal value;
  • Project the results of the conjugate gradient method into the feasible set using constraints.

3.2.1. Conjugate Gradient Method to Find the Optimal Solution

The idea of the conjugate gradient method is to generate the conjugate direction of the Hesse matrix of the convex quadratic function by using the fastest descending direction at the current step at each iteration step. Firstly, choose the least squares solution as the initial value:
s 0 = ( A T A ) 1 A T y
The gradient direction of the fastest descent can be obtained according to (14):
g k = F σ ( s ) = f σ ( s 1 ) s 1 , , f σ ( s 1 ) s 1 = ρ σ 2 s 1 s 1 2 + ρ σ 2 s 1 2 + ρ 2 σ 4 , , ρ σ 2 s 1 s n 2 + ρ σ 2 s n 2 + ρ 2 σ 4
In this method, we assume that the current iteration value is x k and the next iteration value is x k + 1 = x k + α k d k , where α k is the step of the current iteration and d k represents the search direction of the minimum value. The conjugate gradient method can then be expressed as follows:
d k = g k , k = 1 g k + β k d k 1 , k > 1
where g k represents the gradient of the function and β k can be solved using different conjugate gradient methods. The classical solution expressions are as follows:
β k F R = g k 2 g k 1 2 , β k P R P = g k T ( g k g k 1 ) g k 1 2 , β k D Y = g k 2 d k 1 T ( g k g k 1 )
Different gradient conjugate methods usually have different performance results in different scenarios, among which FR and PRP conjugate gradient methods are more commonly used. Therefore, the hybrid conjugate gradient method can be used to modify it to a certain extent. In this paper, the conjugate gradient method of the FR and PRP hybrid is adopted, and the hybridization method is shown in (19):
β k = β k P R P , 0 β k P R P β k F P β k F R , o t h e r w i s e
The conjugate gradient method of the FR and PRP hybridization, which can avoid the disadvantage of producing continuous small steps, is an algorithm with a better all around performance among conjugate gradient methods. For this reason, it is selected as the search direction for calculating the minimum value.

3.2.2. Project the Optimal Solution into the Feasible Set

The solution obtained above is s, which needs to be projected as follows to ensure that the solution is in the feasible set X | Y = A X limited by the constraints.
s p = min s p s 2 , s . t . s p X = s A T ( A A T ) 1 ( A s y )
where s p is the solution in the feasible set.

3.3. BCS-CGSL0 Algorithm

Combining the BCS-SPL algorithm introduced in Section 2.3 and the CGSL0 algorithm introduced in Section 3.2, we propose a BCS-CGSL0 algorithm, which presents the iterative idea of SL0 under the framework of BCS-SPL. Not only does the BCS-CGSL0 algorithm ensure the accuracy and efficiency of reconstruction, but it also removes the block effect caused by block compressed sensing. The step-by-step process of the BCS-CGSL0 algorithm is as Algorithm 1:
Algorithm 1: BCS-CGSL0 Algorithm
Input: measure signal y, measurement matrix Φ , transform domain basis Ψ , block size B
initialization:
  •     x i [ 0 ] = Φ 1 y i ;
  •     σ = [ σ 1 , σ 2 , , σ J ] , where σ j = ρ σ j 1 , 0 < ρ < 1 and σ 1 = 2 max s i ^ [ 0 ] ;
for j = 0 , 1 , , J 1 do
  •     x ^ i [ j ] = w i e n e r ( x i [ j ] ) , where w i e n e r ( ) represents the smoothing filter;
  •     x ˜ i [ j ] = x ^ i [ j ] + Φ T ( y Φ x ^ i [ j ] ) and s i [ j ] = Ψ 1 x ˜ i [ j ] ;
  •     σ = σ j and s = s i [ j ] ;
    for k = 1 , 2 , , L do
  •      g k = ρ σ 2 s 1 s 1 4 + ρ σ 2 s 1 2 + ρ 2 σ 4 , , ρ σ 2 s n s n 4 + ρ σ 2 s n 2 + ρ 2 σ 4 , where n = B 2
  •      d k = g k , k = 1 g k + β k d k 1 , k > 1 , where β k = β k P R P , 0 β k P R P β k F P β k F P , o t h e r w i s e ;
  •      s s + μ d k and s p s A T ( A A T ) 1 ( A s y ) ;
   end
  •     s ^ i [ j ] = s and s ^ i [ j ] = s ^ i [ j ] , s ^ i [ j ] τ 0 , o t h e r w i s e ;
  •     x ˇ i [ j ] = Ψ s ^ i [ j ] and x i [ j ] = x ˇ i [ j ] + Φ T ( y i Φ x ˇ i [ j ] ) ;
end (
Output: the reconstructed image x = x i [ j ]

4. Experiments and Results

In the experiment, the size of the original images is 512 × 512 . Due to the characteristics of block compressed sensing, a too large block size will greatly increase the reconstruction time, and a too small block size will reduce the reconstruction accuracy. Therefore, we choose the block size B = 16 after comprehensive consideration. The observation matrix is an orthogonalized random Gaussian matrix. The sparsity transform basis is a discrete cosine transform (DCT). All experiments were performed using MATLAB 2022a on a computer equipped with an Intel Core TM i9, 3.7 GHz processor, with 32GB of RAM and running on Windows 10. In the BCS-CGSL0 algorithm, we set the decreasing factor as ρ = 0.6 , the iteration number as J = 200 , L = 3 , and the step size as μ = 0.01 . The threshold τ can be calculated according to (7), where λ = 6 .
Two sets of experiments are described in this paper. The first experiment compares the reconstruction effects of the BCS-CGSL0 and SL0 series algorithms at a 0.1–0.5 sampling rate. The second experiment compares the reconstruction effects of the BCS-CGSL0 and no-SL0 series algorithms at a 0.1–0.5 sampling rate. Figure 3 presents the original and observed images of Goldhill and Clown.
The evaluation indicator in the experiment includes the peak signal-to-noise ratio (PSNR), structural similarity (SSIM), and reconstruction time. The results of the two groups of experiments are in Figure 3.

4.1. The Comparison of the BCS-CGSL0 and SL0 Series Algorithms

In the first experiment, we evaluated the effectiveness of the proposed BCS-CGSL0 and SL0 series of algorithms, including the SL0 algorithm [7], NSL0 algorithm [20], DNSL0 algorithm [21], and CGSL0 algorithm proposed in this paper.
Table 1 lists these algorithms’ PSNR, SSIM, and reconstruction time for the Goldhill and Clown images when the rate is 0.5. As can be seen, the SL0 series algorithms have a fast reconstruction speed while ensuring a high reconstruction accuracy. Compared with the other four algorithms, the BCS-CGSL0 algorithm proposed in this paper reduces the reconstruction time while improving the reconstruction accuracy (PSNR and SSIM). The BCS-CGSL0 algorithm has the best effect and reasonable practicability when combining the three evaluation indicators.
Table 2 lists the reconstruction effect of these algorithms for the Goldhill image when the rate is between 0.1 and 0.5. It can be seen that the effect of the SL0 series algorithm is very poor at low sampling rates. For example, the reconstruction effect of the SL0 algorithm is worse when the rate is 0.1, making the reconstructed image unusable. This demonstrates that the reconstruction accuracy of the BCS-CGSL0 algorithm is the highest; the PSNR is increased by more than 2 dB, and the SSIM is higher than the other four algorithms.
Figure 4 compares the PSNR and SSIM of the BCS-CGSL0 and SL0 series algorithms for the Goldhill image when the rate is between 0.1 and 0.8. It can be seen that the SL0 series algorithms are greatly affected by the sampling rate, and the effect is extremely poor when the sampling rate is low. However, the BCS-CGSL0 algorithm overcomes this shortcoming and achieves good reconstruction results when the rate is 0.1, making the BCS-CGSL0 algorithm adaptable to more scenarios.

4.2. The Comparison of The BCS-CGSL0 and non-SL0 Series Algorithms

In the second experiment, we evaluated the effectiveness of the proposed BCS-CGSL0 and no-SL0 series of algorithms, including OMP algorithm [6], Split-Bregman algorithm [9], IRLS algorithm [10], FOCUSS algorithm [11], BCS-SPL algorithm [12], and BCS-TVAL3 algorithm [22]. Table 3 lists these algorithms’ PSNR, SSIM, and reconstruction time for the Goldhill and Clown images when the rate is 0.5. Compared with Table 1, it can be seen that these typical no-SL0 series algorithms have a good reconstruction accuracy. However, most take a long time and are unsuitable for real-time scenarios. In contrast, the BCS-CGSL0 algorithm has good practicability since it can complete the reconstruction in a short time and have a better reconstruction accuracy.
Table 4 lists the reconstruction effect of these algorithms for the Goldhill image when the rate is between 0.1 and 0.5. Compared with Table 1, it can be seen that the no-SL0 series algorithms still have a relatively stable reconstruction effect at low resolutions, which is better than the SL0 series algorithms. However, their disadvantage is that the reconstruction time is too long for adaptation to real-time scenarios. Although the BCS-SPL algorithm’s reconstruction speed is fast, the reconstruction accuracy still needs to be improved. In contrast, the advantages of the BCS-CGSL0 algorithm are more obvious, including a higher reconstruction accuracy and shorter reconstruction time.
Figure 5 compares the PSNR and SSIM of the BCS-CGSL0 and no-SL0 series algorithms for the Goldhill image when the rate is between 0.1 and 0.8. It intuitively demonstrates that the reconstruction effect of no-SL0 series algorithms is more stable at different sampling rates and that the BCS-CGSL0 algorithm has an improved performance.

5. Conclusions

This study proposes a compressed sensing reconstruction algorithm based on block compressed sensing with a conjugate gradient smoothed l 0 norm, BCS-CGSL0. The core of our method is to combine the idea of the SL0 series algorithms with the BCS-SPL algorithms, to remove the block effect of the reconstructed image and improve the reconstruction accuracy and speed. The main contributions of this paper are as follows:
  • We propose a new function called the inverse trigonometric fraction function, which approximates the l 0 norm better than similar functions;
  • We propose a method for optimizing the SL0 algorithm (CGSL0), using the inverse trigonometric fraction function to approximate the l 0 norm and the modified conjugate gradient method to solve the optimization problem;
  • We propose a reconstruction algorithm that combines CGSL0 and BCS-SPL, which has a high reconstruction accuracy and removes the blockiness of reconstructed images.
Through the simulation experiment verification of encoding low-resolution images, compared with SL0 series and no-SL0 series algorithms, this algorithm can ensure a good reconstruction speed when improving the reconstruction accuracy, which ensures that the algorithm has great value in practice. In future work, we plan to further study improvements to the SL0 series of algorithms, improve their accuracy and speed, and enhance their applicability in practical scenarios.

Author Contributions

Methodology and software, Y.Z.; validation and formal analysis, Y.Z. and C.Z.; writing—original draft preparation, Y.Z.; writing—review and editing, X.C. and K.G.; visualization, Y.Z.; supervision and project administration, S.L. and K.G. All authors have read and agreed to the published version of the version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China under Grant U2241275 and Grant 61827814, in part by the Beijing Natural Science Foundation under Grant Z190018, and in part by the China high-resolution Earth Observation System Project under Grant 52.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
SL0Smoothed l 0 Norm
IRLSIteratively Reweighted Least Squares
BCS-SPLBlock Compressed Sensing with Smoothed Projected Landweber
BCS-CGSL0Block Compressed Sensing with Conjugate Gradient Smoothed l 0 Norm

References

  1. Kozlov, D.; Ott, P. CFAR Detector for Compressed Sensing Radar Based on l1-norm Minimisation. In Proceedings of the 2020 28th European Signal Processing Conference (EUSIPCO), Amsterdam, Netherlands, 18–21 January 2021. [Google Scholar]
  2. Porta, C.; Chang, C.I. Progressive Compressively Sensed Band Processing for Hyperspectral Classification. IEEE Trans. Geosci. Remote Sens. 2020, 99, 1–13. [Google Scholar]
  3. Wei, J.; Huang, Y.; Ke, L.; Wang, L. Nonlocal Low-Rank-Based Compressed Sensing for Remote Sensing Image Reconstruction. IEEE Geosci. Remote Sens. Lett. 2016, 13, 1557–1561. [Google Scholar] [CrossRef]
  4. Zhu, Y.; Liu, W.; Shen, Q. Adaptive Algorithm on Block-Compressive Sensing and Noisy Data Estimation. Electronics 2019, 8, 753. [Google Scholar] [CrossRef]
  5. Kumamoto, K.; Matsuo, S.; Kuroki, Y. A Convolutional Dictionary Learning based l1 Norm Error with Smoothed l0 Norm Regression. In Proceedings of the 2019 International Symposium on Intelligent Signal Processing and Communication Systems (ISPACS), Taipei, Taiwan, 3–6 December 2019. [Google Scholar]
  6. Adcock, B.; Brugiapaglia, S.; King-Roskamp, M. Iterative and greedy algorithms for the sparsity in levels model in compressed sensing. In Proceedings of the Wavelets and Sparsity XVIII, San Diego, CA, USA, 9 September 2019. [Google Scholar]
  7. Mohimani, H.; Babaie-Zadeh, M.; Jutten, C. A Fast Approach for Overcomplete Sparse Decomposition Based on Smoothed L-0 Norm. IEEE Trans. Signal Process. 2009, 57, 289–301. [Google Scholar] [CrossRef]
  8. Wang, Y.; Yin, W. Sparse signal reconstruction via iterative support detection. SIAM J. Imaging Sci. 2010, 3, 462–491. [Google Scholar] [CrossRef]
  9. Goldstein, T.; Osher, S. The split Bregman method for L1-regularized problems. SIAM J. Imaging Sci. 2009, 2, 323–343. [Google Scholar] [CrossRef]
  10. Rodriguez, P.; Wohlberg, B. An iteratively reweighted norm algorithm for total variation regularization. In Proceedings of the 2006 Fortieth Asilomar Conference on Signals, Systems and Computers, Pacific Grove, CA, USA, 29 October–1 November 2006. [Google Scholar]
  11. Gorodnitsky, I.F.; Rao, B.D. Sparse signal reconstruction from limited data using FOCUSS: A re-weighted minimum norm algorithm. IEEE Trans. Signal Process. 2002, 45, 600–616. [Google Scholar] [CrossRef]
  12. Mun, S.; Fowler, J.E. Block compressed sensing of images using directional transforms. In Proceedings of the 2010 Data Compression Conference, Snowbird, UT, USA, 22 April 2010. [Google Scholar]
  13. Wang, X.D.; Li, Y.H.; Wang, Z.; Liu, W.G.; Wang, J.N. Self-adaptive block-based compressed sensing imaging for remote sensing applications. J. Appl. Remote Sens. 2020, 14, 016513. [Google Scholar] [CrossRef]
  14. Li, L.; Liu, P.; Wu, J.; Wang, L.; He, G. Spatiotemporal Remote-Sensing Image Fusion With Patch-Group Compressed Sensing. IEEE Access 2020, 8, 209199–209211. [Google Scholar] [CrossRef]
  15. Zou, F.; Liu, X.; Lai, Z.; Li, J.; Luo, K. Redundant Dictionary based Pulse Wave Compressed Sensing. In Proceedings of the 2020 Chinese Automation Congress (CAC), Shanghai, China, 6–8 November 2020. [Google Scholar]
  16. Zhang, H.; Shan, T.; Liu, S.; Tao, R. Performance Evaluation and Parameter Optimization of Sparse Fourier Transform. Signal Process 2021, 179, 107823. [Google Scholar] [CrossRef]
  17. Wang, Z.; Liu, L.; Chen, S.; Feng, M. Secure image block compressive sensing using chaotic DCT sparse basis and partial chaotic DHT measurement matrix. Sens. Imaging 2020, 21, 1–23. [Google Scholar] [CrossRef]
  18. Wang, Z.; Hussein, Z.S.; Wang, X. Secure compressive sensing of images based on combined chaotic DWT sparse basis and chaotic DCT measurement matrix. Opt. Lasers Eng. 2020, 134, 106246. [Google Scholar] [CrossRef]
  19. Figueiredo, M.; Nowak, R.D.; Wright, S.J. Gradient Projection for Sparse Reconstruction: Application to Compressed Sensing and Other Inverse Problems. IEEE J. Sel. Top. Sign. Proces. 2008, 1, 586–597. [Google Scholar] [CrossRef]
  20. Lin, W.J.; Zhao, R.Z.; Li, H. The NSL0 Algorithm for Compressive Sensing Signal Reconstruction. J. New Ind. 2011, 1, 78–84. (In Chinese) [Google Scholar]
  21. Yang, L.L.; Zhao, S.M.; Zheng, B.Y.; Tang, W.J. The Improved Reconstruction Algorithm for Compressive Sensing on SL0. Signal Process 2012, 28, 834–841. (In Chinese) [Google Scholar]
  22. Li, C. An Efficient Algorithm for Total Variation Regularization with Applications to the Single Pixel Camera and Compressive Sensing; Department of Computational and Applied Mathematics: Houston, TX, USA, 2009; Volume 9. [Google Scholar]
Figure 1. Compressed imaging system.
Figure 1. Compressed imaging system.
Sensors 23 04870 g001
Figure 2. Smooth function model comparison diagram. The most central curve is the l 0 norm. The remaining curves from inside to outside are smooth function approximation models of SL0, NSL0, DNSL0, and CGSL0 algorithms. It can be seen that the CGSL0 smooth function has the highest approximation to the l 0 norm.
Figure 2. Smooth function model comparison diagram. The most central curve is the l 0 norm. The remaining curves from inside to outside are smooth function approximation models of SL0, NSL0, DNSL0, and CGSL0 algorithms. It can be seen that the CGSL0 smooth function has the highest approximation to the l 0 norm.
Sensors 23 04870 g002
Figure 3. Images used in the experiment and their observations: (a) original image of Goldhill with a size of 512 × 512; (b) observed image of Goldhill with a size of 32 × 32; (c) original image of Clown with a size of 512 × 512; (d) observed image of Clown with a size of 32 × 32.
Figure 3. Images used in the experiment and their observations: (a) original image of Goldhill with a size of 512 × 512; (b) observed image of Goldhill with a size of 32 × 32; (c) original image of Clown with a size of 512 × 512; (d) observed image of Clown with a size of 32 × 32.
Sensors 23 04870 g003
Figure 4. Comparison of the reconstruction effect of the BCS-CGSL0 and SL0 series algorithms for Goldhill image: (a) comparison of PSNR; (b) comparison of SSIM.
Figure 4. Comparison of the reconstruction effect of the BCS-CGSL0 and SL0 series algorithms for Goldhill image: (a) comparison of PSNR; (b) comparison of SSIM.
Sensors 23 04870 g004
Figure 5. Comparison of the reconstruction effect of the BCS-CGSL0 and no-SL0 series algorithms for Goldhill image: (a) comparison of PSNR. (b) comparison of SSIM.
Figure 5. Comparison of the reconstruction effect of the BCS-CGSL0 and no-SL0 series algorithms for Goldhill image: (a) comparison of PSNR. (b) comparison of SSIM.
Sensors 23 04870 g005
Table 1. Comparison of reconstruction quality for Goldhill and Clown images with SL0, NSL0, DNSL0, CGSL0, and BCS-CGSL0 algorithms.
Table 1. Comparison of reconstruction quality for Goldhill and Clown images with SL0, NSL0, DNSL0, CGSL0, and BCS-CGSL0 algorithms.
ImageAlgorithmPSNR/dBSSIMTims/s
GoldhillSL032.19600.89130.5027
NSL034.81570.93470.9250
DNSL035.00880.93712.1053
CGSL035.27360.94030.8343
BCS-CGSL037.55920.96670.8078
ClownSL030.95150.90380.5037
NSL034.63060.95420.9064
DNSL034.99980.95721.9203
CGSL035.61380.96170.7518
BCS-CGSL038.53110.97780.8574
Table 2. Reconstruction effect of SL0, NSL0, DNSL0, CGSL0, and BCS-CGSL0 algorithms for Goldhill image when the rate is between 0.1 and 0.5.
Table 2. Reconstruction effect of SL0, NSL0, DNSL0, CGSL0, and BCS-CGSL0 algorithms for Goldhill image when the rate is between 0.1 and 0.5.
Sample RateAlgorithmPSNR/dBSSIMTims/s
0.1SL014.52550.14740.2132
NSL020.26830.35170.2832
DNSL022.31200.58871.1115
CGSL024.80580.61041.1550
BCS-CGSL028.07090.79881.0696
0.2SL016.20640.37540.2998
NSL022.59540.65570.3816
DNSL023.43650.69401.3254
CGSL026.96540.81691.2732
BCS-CGSL030.95860.87870.9225
0.3SL021.58780.59760.4207
NSL028.79020.86590.6539
DNSL029.20170.87521.6780
CGSL030.34020.89991.6083
BCS-CGSL033.38830.92401.1859
0.4SL028.06170.75420.4612
NSL032.90410.90270.7291
DNSL033.06160.90561.8192
CGSL033.27960.90960.7126
BCS-CGSL035.58130.95080.9762
0.5SL032.19600.89130.5027
NSL034.81570.93470.9250
DNSL035.00880.93712.1053
CGSL035.27360.94030.8343
BCS-CGSL037.55920.96670.8078
Table 3. Comparison of the reconstruction quality for Goldhill and Clown images with the OMP, IRLS, Split-Bregman, BCS-SPL, BCS-TVAL3, and BCS-CGSL0 algorithms.
Table 3. Comparison of the reconstruction quality for Goldhill and Clown images with the OMP, IRLS, Split-Bregman, BCS-SPL, BCS-TVAL3, and BCS-CGSL0 algorithms.
ImageAlgorithmPSNR/dBSSIMTims/s
GoldhillOMP32.50680.89855.2730
Split-Bregman34.61230.934313.0204
IRLS35.01870.938732.6246
FOCUSS34.40920.926092.4273
BCS-SPL34.90580.93560.4938
BCS-TVAL335.62390.93208.4257
BCS-CGSL037.55920.96670.8078
ClownOMP33.35860.94055.3338
Split-Bregman34.66070.956111.8532
IRLS35.26170.961632.4094
FOCUSS34.49670.953379.2619
BCS-SPL35.36010.93480.7055
BCS-TVAL336.22450.95539.9349
BCS-CGSL038.53110.97780.8574
Table 4. The reconstruction effect of the OMP, Split-Bregman, IRLS, FOCUSS, BCS-SPL, BCS-TVAL3, and BCS-CGSL0 algorithms for Goldhill image when the rate is between 0.1 and 0.5.
Table 4. The reconstruction effect of the OMP, Split-Bregman, IRLS, FOCUSS, BCS-SPL, BCS-TVAL3, and BCS-CGSL0 algorithms for Goldhill image when the rate is between 0.1 and 0.5.
Sample RateAlgorithmPSNR/dBSSIMTims/s
0.1OMP23.21510.51870.2837
Split-Bregman24.15450.5978110.5349
IRLS23.54540.57348.0813
FOCUSS23.90180.573720.9178
BCS-SPL25.41810.63450.8154
BCS-TVAL326.03470.761810.8333
BCS-CGSL028.07090.79881.0696
0.2OMP25.99410.66770.7621
Split-Bregman27.62780.756455.2424
IRLS27.64650.757112.6816
FOCUSS27.41720.741140.7951
BCS-SPL27.41420.72310.7601
BCS-TVAL329.69010.863910.0676
BCS-CGSL030.95860.87870.9225
0.3OMP28.40800.77811.7354
Split-Bregman30.24470.845132.7416
IRLS30.35400.848518.1794
FOCUSS30.02290.832956.2880
BCS-SPL29.95530.81910.7631
BCS-TVAL332.41810.91239.8397
BCS-CGSL033.38830.92401.1859
0.4OMP30.51070.85253.0058
Split-Bregman32.68240.903119.7576
IRLS32.86250.902324.7105
FOCUSS32.44330.894775.5162
BCS-SPL32.35050.89080.6507
BCS-TVAL334.59260.94079.7414
BCS-CGSL035.58130.95080.9762
0.5OMP32.50680.89855.2730
Split-Bregman34.61230.934313.0204
IRLS35.01870.938732.6246
FOCUSS34.40920.926092.4273
BCS-SPL34.90580.93560.4938
BCS-TVAL336.22450.95539.9349
BCS-CGSL037.55920.96670.8078
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, Y.; Chen, X.; Zeng, C.; Gao, K.; Li, S. Compressed Imaging Reconstruction Based on Block Compressed Sensing with Conjugate Gradient Smoothed l0 Norm. Sensors 2023, 23, 4870. https://doi.org/10.3390/s23104870

AMA Style

Zhang Y, Chen X, Zeng C, Gao K, Li S. Compressed Imaging Reconstruction Based on Block Compressed Sensing with Conjugate Gradient Smoothed l0 Norm. Sensors. 2023; 23(10):4870. https://doi.org/10.3390/s23104870

Chicago/Turabian Style

Zhang, Yongtian, Xiaomei Chen, Chao Zeng, Kun Gao, and Shuzhong Li. 2023. "Compressed Imaging Reconstruction Based on Block Compressed Sensing with Conjugate Gradient Smoothed l0 Norm" Sensors 23, no. 10: 4870. https://doi.org/10.3390/s23104870

APA Style

Zhang, Y., Chen, X., Zeng, C., Gao, K., & Li, S. (2023). Compressed Imaging Reconstruction Based on Block Compressed Sensing with Conjugate Gradient Smoothed l0 Norm. Sensors, 23(10), 4870. https://doi.org/10.3390/s23104870

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop