1. Introduction
In recent years, with the continuous development of artificial intelligence, smart devices have been continuously updated and applied across various industries, making a variety of sensors an important bridge connecting these smart devices. Sensors collect data information and perform calculations, extractions, reconstructions, and other processes to achieve control over smart devices. In light of the vast amounts of data that need to be stored and transmitted, efficient and high-quality storage and transmission have garnered the attention of researchers.
The Nyquist sampling theorem is one of the most important theoretical foundations of modern signal processing. To prevent distortion or aliasing in the reconstructed signal, the sampling rate must exceed twice the highest frequency of the sampled signal [
1]. However, high sampling rates and the process of compressed transmission waste a significant amount of resources. The emergence of compressed sensing provides a new approach to effectively address this issue. Compressed sensing can capture all the information in the original signal effectively with a lower sampling rate, enabling the reconstruction of the original signal and significantly reducing the amount of data required for storage or transmission [
2]. Therefore, it has been widely applied in fields such as image encryption [
3], medical diagnosis [
4], radar imaging [
5], and superharmonic detection [
6].
The key to reconstructing signals with compressed sensing technology lies in the choice of reconstruction algorithm [
7]. Currently, common algorithms for compressed sensing reconstruction are divided into three categories: convex optimization algorithms [
8,
9], greedy algorithms [
10], and combinatorial algorithms [
11]. Convex optimization algorithms can accurately reconstruct the original signal under certain conditions; however, their high computational complexity makes them difficult to apply to large-scale or high-dimensional signals [
12]. Combinatorial algorithms achieve the reconstruction of the original signal by combining different algorithms. However, compared to single algorithms, this approach requires more computational resources, making it challenging to implement in resource-constrained situations [
13]. In contrast, greedy algorithms have garnered widespread attention due to their low computational complexity and short reconstruction time [
14]. The Matching Pursuit (MP) algorithm has been the most widely used greedy algorithm for signal reconstruction in recent years, leading to the development of many improved algorithms based on MP, such as OMP, ROMP, and STOMP. However, these algorithms require the sparsity of the signal as prior information, which is often difficult to ascertain, thus reducing the applicability of many MP algorithms. The emergence of the SAMP algorithm has broken this deadlock, as it is widely used due to its characteristic of not requiring sparsity as prior information; researchers have also proposed many improved algorithms based on SAMP, such as CCSAMP and VSSAMP. Although the aforementioned methods yield good reconstruction results under certain conditions, they are somewhat limited by their need for prior information or fixed step sizes, often resulting in unsatisfactory reconstruction outcomes. To address this, this paper proposes a variable-step sparse adaptive matching pursuit algorithm based on the Spearman correlation coefficient, which introduces an adaptive step size adjustment method without requiring prior information, enabling the high-quality reconstruction of one-dimensional sparse signals and two-dimensional image signals.
The organization of the rest of this paper is as follows:
Section 2 discusses related work on reconstruction algorithms, while
Section 3 covers the theoretical background of compressed sensing and the process of the SAMP algorithm.
Section 4 presents the algorithm proposed in this paper,
Section 5 details the experimental results of the proposed method under different signals, and finally,
Section 6 concludes the paper.
2. Related Work
Reliable reconstruction algorithms not only improve reconstruction accuracy but also exhibit good work efficiency, and greedy algorithms possess this characteristic; therefore, this subsection will focus on greedy algorithms.
The main idea of the greedy algorithm is to approximate the original signal during the iterative process, resulting in a reconstructed signal with high precision. The Matching Pursuit (MP) [
15] and Orthogonal Matching Pursuit (OMP) [
16] algorithms are among the earliest greedy algorithms; OMP effectively resolves the issue of atom redundancy in MP by ensuring the orthogonality between atoms selected in successive iterations, thus providing greater robustness. However, this orthogonality also introduces a series of problems, for instance, the OMP algorithm may become trapped in local optima and struggle to converge in certain cases [
17]. To address this, Needell et al. proposed the Regularized Orthogonal Matching Pursuit (ROMP) algorithm [
18], which introduces randomness in atom selection, thereby mitigating the issues arising from orthogonality and reducing computational complexity while ensuring algorithm performance. For the aforementioned algorithms, if the selected atoms do not contain information from the original signal, the reconstruction quality will also diminish. In response, Dileep et al. proposed the Compressed Sampling Matching Pursuit (CoSaMP) algorithm, which incorporates a backtracking approach to identify and reselect atoms that do not meet termination conditions, effectively enhancing algorithm performance by choosing columns most relevant to the current residual. Similarly, the Subspace Pursuit (SP) [
19] algorithm also incorporates the backtracking concept.
However, the aforementioned algorithms require the sparsity of the signal as prior information, limiting their applicability. In response, Dohono et al. proposed the Phase Orthogonal Matching Pursuit (StOMP) algorithm [
20], which does not require signal sparsity as an input parameter and is advantageous when sparsity is unknown. Subsequently, to address the issue of slow convergence speed due to strong orthogonality, Blumensath et al. introduced the Phase-Controlled Weak Orthogonal Matching Pursuit (SWOMP) [
21], which incorporates a weak selection process and sets a threshold for atom selection to ensure reconstruction accuracy, though the choice of threshold remains an unresolved issue. Similarly, the Sparsity Adaptive Matching Pursuit (SAMP) [
22] algorithm also does not require sparsity as prior information; it is not constrained by preset thresholds and gradually approximates the original signal through continuous iteration with a fixed step size. However, the choice of an appropriate step size affects the reconstruction performance; a smaller step size may slow convergence, increasing iterations and time, while a larger step size may skip the optimal solution, reducing reconstruction accuracy [
23]. In response, researchers improved the SAMP algorithm by introducing a variable step size strategy to reduce iteration counts and enhance reconstruction accuracy. Huang et al. proposed the Adaptive Regularized Matching Pursuit Algorithm for Compressed Sensing with Sparsity and Step Size [
24], which uses a coefficient
to dynamically adjust the step size. Wang et al. introduced the Improved Sparsity Adaptive Matching Pursuit Algorithm based on Compressed Sensing [
25], which dynamically adjusts the step size by incorporating a nonlinear function and a variable step size correction factor, enhancing reconstruction accuracy. The Variable Step Size Sparsity Adaptive Matching Pursuit (VSSAMP) algorithm estimates sparsity and determines the step size based on the energy difference between two residuals, optimizing reconstruction performance for one-dimensional signals and two-dimensional images while reducing reconstruction time [
26]. The Correlation Coefficient Sparsity Adaptive Matching Pursuit (CCSAMP) Algorithm [
27] identifies a linear relationship between the signal recovered from the support set and the candidate set. It uses the Pearson correlation coefficient (PCC) to measure the linear relationship between two vectors, enabling dynamic adjustment of the step size. However, the Pearson correlation coefficient requires variables to follow a normal distribution, which may limit its applicability to non-normally distributed data. Additionally, the Pearson correlation coefficient is highly sensitive to outliers, which may cause results to deviate from the true relationship.
To address the aforementioned issues, this paper proposes a variable step size sparsity adaptive Matching Pursuit Algorithm based on the Spearman correlation coefficient, which measures the similarity between the candidate set and the input signal, adapting the step size through correlation-based partitioning. This algorithm achieves high reconstruction accuracy for one-dimensional sparse signals and two-dimensional image signals.
4. Proposed Method
The SAMP algorithm estimates the sparsity of each iteration by comparing the residual value, gradually approaching the optimal sparse solution, and finally successfully reconstructing the original signal. However, due to the use of fixed step size during iteration, the accuracy of sparsity estimation is compromised. Therefore, the Spearman correlation coefficient SAMP (SCCSAMP) algorithm is proposed to dynamically change the step size according to the correlation between the compressed signal and the candidate set.
The initial parameters are expressed as follows:
After parameters initialization,
u is defined as the dot product of sensing matrix
A and residual
r. Then,
L maximum values are selected from
u to form
St, which corresponds to column set. Then, the inner product of the sensing matrix
A and the residual
r is calculated, the largest
L values are selected, and the relevant column sequence number in
St is stored:
The index set
Λt−1 of the previous iteration was merged with the
St of the current iteration to construct the candidate set
Ct of the
t-th iteration:
. The sparse signal of the
t-th iteration is calculated using the least squares method as follows:
The
L-th term with the largest absolute value from
θt is selected and labeled as
θtL, while the
L-th column corresponding to
At is selected and denoted as
AtL, and the set of sequence numbers of
AtL is reserved in the support set
Ft, The residuals are obtained as follows:
The residual r is used to judge whether the stop iteration condition is satisfied; the iteration is terminated if it is satisfied, and the iteration is continued if it is not satisfied.
In the SAMP algorithm, we construct the vector
ytL,
ytL that was generated by the candidate set
Ft.
As the number of iterations increases, the rnew tends to zero; therefore, the vectors y and ytL are gradually close to each other. Compressed sensing theory states that the sparse signal x can be recovered from the small number of linear measurements , requiring the perception matrix A to satisfy the restricted isometry property (RIP).
During the recovery of sparse signals, the selection of atoms most relevant to vector y is achieved by maximizing , in other words, the projection of the vector y onto the column space of the sensing matrix A is actually calculated. Therefore, selecting the set of atoms Ft corresponding to the L component with the largest absolute value means selecting the atoms that contribute the most to vector y. This selection strategy ensures that the atoms selected in each iteration will contribute the most to signal y, so that the vector ytL contains more useful atoms and minimizes the reconstruction error at each step. This theoretically increases the representativeness and information content of the selected atoms, and better preserves the geometric properties of the signal, making the selected atoms more orthogonal and independent, thus increasing the likelihood that the sensing matrix A satisfies the RIP.
Therefore, the algorithm proposed in this paper deals with the step size in segments by measuring the similarity between the vectors y and ytL. A higher similarity means that atoms closer to the sparse solution are selected. By judging the similarity, different step sizes can be selected to effectively approximate the sparse solution.
Common correlation analysis methods include covariance, the Pearson correlation coefficient, and the Spearman correlation coefficient. The calculation result of covariance is only used to judge whether the two variables are correlated, and its value is not easy to define the degree of correlation. The Pearson coefficient can be used to analyze the degree of correlation, but it is only suitable for data that conform to the normal distribution. The Spearman correlation coefficient can be used for the correlation analysis of any two groups of data with the same length. Spearman’s correlation coefficient reflects the correlation between variables by using monotone function, which is not affected by data dimension and is not sensitive to outliers [
31]. Therefore, the Spearman correlation coefficient is selected for correlation analysis in this paper, and its calculation method is as follows:
where
yi and (
ytL)
i are the observed rank of two well-ordered variables,
and
are the average rank of two variables, and
N is the total number of each variable.
The range of the Spearman’s correlation coefficient ρ is [–1, 1]. A negative value indicates a negative correlation between two variables, while a positive value indicates a positive correlation between two variables. The absolute value of the calculated results can be used to characterize the degree of correlation between the two variables. The usual judgment criteria are as follows: 0.8~1.0 indicates a very strong correlation, 0.6~0.8 indicates a strong correlation, 0.4~0.6 indicates a moderate correlation, 0.2~0.4 indicates a weak correlation, and 0.0~0.2 indicates a very weak correlation or no correlation.
Assuming that
ρ represents the Spearman correlation between
y and
ytL, the formula is as follows:
According to Spearman’s criterion, we can process step size in different steps. When 0 < ρ < 0.4, it is considered that the correlation between the two is weak and the step length should be taken. If 0.4 < ρ < 0.8, it is considered that the correlation between the two is above medium, and a small length is taken; if ρ > 0.8, it is considered that the correlation between the two vectors is very strong, and the sparse solution can be gradually approached with step size 1.
We calculate the complexity of the method in this paper according to the steps of the pseudocode. The pseudocode of the method proposed in this article is as follows (Algorithm 2, based on Algorithm 1):
Algorithm 2: SCCSAMP |
Input: Compressed signal , sensing matrix , step size . Output: Sparse coefficient of the estimated signal .
- 1:
Initialization: - 2:
; ; ; ; ; - 3:
Repeat: - 4:
- 5:
- 6:
- 7:
- 8:
- 9:
- 10:
- 11:
If then - 12:
- 13:
else if then - 14:
- 15:
else - 16:
- 17:
end if - 18:
If then - 19:
Output - 20:
else if - 21:
- 22:
- 23:
- 24:
else - 25:
- 26:
- 27:
- 28:
end if - 29:
Until - 30:
Output: - 31:
|
In the code above, Step 4 computes the inner product and sorts to extract the top L values. The complexity of the inner product is O(MN), and the complexity of the descending sort is O(NlogN). Therefore, the overall complexity is (MN + NlogN).
In Step 5, the process of establishing the candidate set Ct requires the most N operations, resulting in a complexity of O(N).
In Step 6, the calculation of the least squares solution involves matrix–matrix multiplication, with a complexity of O(L2). The complexity for matrix inversion is O(L3), leading to a total complexity of O(ML2 + L3).
In Step 7, the complexity of constructing F is O(L).
In Steps 8 and 9, the calculation of the residual involves a matrix–vector multiplication with a time complexity of O(ML).
In Step 10, calculating the Spearman correlation coefficient has a complexity of O(MlogM).
Based on the above analysis, after k iterations, the overall complexity can be summarized as O(K(MN + NlogN + N + L2M + L3 + L + ML + MlogM)). Since L is significantly smaller than both M and N, the total complexity for the SCCSAMP algorithm can be approximated as O(k(MN + NlogN)).