Next Article in Journal
A Robust and High-Dimensional Clustering Algorithm Based on Feature Weight and Entropy
Previous Article in Journal
A Fragile Image Watermarking Scheme in DWT Domain Using Chaotic Sequences and Error-Correcting Codes
Previous Article in Special Issue
Optimal Passive Source Localization for Acoustic Emissions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Compressive Sensing via Variational Bayesian Inference under Two Widely Used Priors: Modeling, Comparison and Discussion

by
Mohammad Shekaramiz
1,* and
Todd K. Moon
2
1
Machine Learning & Drone Lab, Electrical and Computer Engineering Program, Engineering Department, Utah Valley University, 800 West University Parkway, Orem, UT 84058, USA
2
Electrical and Computer Engineering Department, Utah State University, 4120 Old Main Hill, Logan, UT 84322, USA
*
Author to whom correspondence should be addressed.
Entropy 2023, 25(3), 511; https://doi.org/10.3390/e25030511
Submission received: 30 January 2023 / Revised: 6 March 2023 / Accepted: 14 March 2023 / Published: 16 March 2023

Abstract

:
Compressive sensing is a sub-Nyquist sampling technique for efficient signal acquisition and reconstruction of sparse or compressible signals. In order to account for the sparsity of the underlying signal of interest, it is common to use sparsifying priors such as Bernoulli–Gaussian-inverse Gamma (BGiG) and Gaussian-inverse Gamma (GiG) priors on the components of the signal. With the introduction of variational Bayesian inference, the sparse Bayesian learning (SBL) methods for solving the inverse problem of compressive sensing have received significant interest as the SBL methods become more efficient in terms of execution time. In this paper, we consider the sparse signal recovery problem using compressive sensing and the variational Bayesian (VB) inference framework. More specifically, we consider two widely used Bayesian models of BGiG and GiG for modeling the underlying sparse signal for this problem. Although these two models have been widely used for sparse recovery problems under various signal structures, the question of which model can outperform the other for sparse signal recovery under no specific structure has yet to be fully addressed under the VB inference setting. Here, we study these two models specifically under VB inference in detail, provide some motivating examples regarding the issues in signal reconstruction that may occur under each model, perform comparisons and provide suggestions on how to improve the performance of each model.

1. Introduction

Compressive sensing (CS) involves efficient signal acquisition and reconstruction techniques in a sub-Nyquist sampling sense. The CS framework can capture the vital information of the underlying signal via a small number of measurements while retaining the ability to reconstruct the signal. CS operates under the assumption that the signal is compressible or sparse, and the number and location of dominating components are unknown in most cases [1,2,3]. Compressibility or sparsity means that the signal has few dominating elements under some proper basis. CS has been used in a variety of applications such as the single-pixel camera, missing pixels and inpainting removal of images, biomedical such as heart rate estimation, internet of things (IoT), geostatistical data analysis, seismic tomography, communications such as blind multi-narrowband signals sampling and recovery, the direction of arrival (DoA) estimation, spectrum sharing of radar and communication signals, wireless networks and many more [4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27]. In the linear CS framework, the problem is posed as
y = A x s + e ,
where y R M contains the measurements, x s R N is the sparse signal of interest, e is the noise representing either the measurement noise or the insignificant coefficients of x s and, generally, M N [1,2]. The measurement matrix can be defined as A = Φ Ψ , where Φ is the sensing design matrix and Ψ is a proper sparsifying basis. There exist various approaches to solve for x s in (1) including greedy-based, convex-based, thresholding-based and sparse Bayesian learning (SBL) algorithms [27,28,29,30,31,32,33,34,35,36,37,38,39,40,41,42,43,44,45,46,47,48,49,50,51,52,53,54,55,56,57,58,59,60,61,62,63,64]. Typically, the performance of CS reconstruction is determined in terms of the mean-squared reconstruction error. In this paper, we are also interested in the more demanding requirements of the probability of detection and the false alarm of the nonzero components. This is of more interest to CS applications such as blind multinarrowband signals, spectrum sharing RADAR, etc. [11,12,13,14,15].
The focus of this paper is on sparse Bayesian learning (SBL) for the CS problem. Bayesian learning models are flexible in incorporating prior knowledge of the characteristics of the underlying signal into the model. Bayesian learning also provides a distribution of the hidden variables, which is more informative than the point estimate approaches. A prior favoring the sparsity or compressibility in x s can be represented in the SBL framework via Gaussian-inverse Gamma (GiG), Laplace-inverse Gamma (LiG), Bernoulli–Gaussian-inverse Gamma (BGiG), often referred to as spike-and-slab prior, etc. [27,46,47,48,49,50,51,52,53,54,55,56,57,58,59]. The inference on parameters and hidden variables in these models is usually made using Markov chain Monte Carlo (MCMC) and variational Bayes (VB) [27,45,46,47,48,49,50,51,52]. In this paper, we focus on the two most commonly used SBL prior models for solving the inverse problem of compressive sensing: Bernoulli–Gaussian-inverse Gamma (BGiG) prior and Gaussian-inverse Gamma (GiG). These models have been widely used, along with some additional priors, for sparse recovery of signals or images with block-sparsity/clustering patterns, sparse signals with correlated coefficients or other structured patterns [26,27,48,49,50,51,62,63].
We use VB inference to estimate the variables and parameters of the SBL model. VB is preferred over MCMC because MCMC is computationally expensive, though it can numerically approximate exact posterior distributions with a sufficient amount of computation. The convergence diagnostic of MCMC requires additional work, such as measuring the potential scale reduction factor (PSRF) for all the hidden variables and parameters of the model or monitoring their trace plots [45,50,51,52,65]. In contrast, VB inference can lead to a reasonable approximation of the exact posteriors, using less computation than MCMC and less effort to monitor the convergence [45,51,66,67]. In this paper, we present the derivation of the update rules of the parameters and variables using VB inference for both the BGiG and GiG models. (Portions of this derivation have been previously presented in [68,69]). Although these prior models have been widely used in various applications of compressive sensing, the study of the overall performance of these models under VB inference has yet to be thoroughly investigated. The preference for one model over the other becomes crucial when dealing with moderate or low sampling ratios, which we discuss in this paper. Here, we study the issues associated with each model via some motivational examples. Pre-/postprocessing approaches will then be discussed to tackle the issues. Finally, the overall performance of BGiG and GiG is compared.
The remainder of this work is organized as follows. In Section 2, we present a brief background on VB inference. We study Bernoulli–Gaussian-inverse Gamma modeling for CS using VB in Section 3. Some motivational examples are provided to show the issues with this approach. Section 4 represents Gaussian-inverse Gamma modeling, the associated update rules using VB inference and a motivational example of the issue that may occur using this approach. In Section 5, we study the improvement of the performances of the models after some pre-/postprocessing along with simulation results and comparisons. Section 6 concludes this work.

2. Variational Bayesian Inference

Variational Bayes (VB) is an effective approach to approximate intractable integrals that may arise in Bayesian inference. The main idea behind variational methods is to use a family of distributions over the latent variables with their own variational parameters. VB is a fast alternative to sampling methods such as Markov chain Monte Carlo (MCMC) and Sequential Monte Carlo (SMC) for performing approximate Bayesian inference [70,71]. For a probabilistic model with unknown parameters θ and hidden variables x , the posterior distribution of the unknowns, given a set of observations y , can be written as p ( x , θ | y ) = p ( x , θ , y ) / p ( y ) . Finding the exact posterior in closed form to perform the inference would be a challenge, as the marginal distribution p ( y ) = p ( y , x , θ ) d x d θ is often intractable. As an efficient approximation method for such inference problems, VB provides an analytical approximation to the posterior p ( x , θ | y ) . VB approximates the joint density p ( x , θ | y ) via a variational distribution Q x , θ ( x , θ ) , i.e., p ( x , θ | y ) Q x , θ ( x , θ ) . VB assumes that the distribution Q can be fully factorized with respect to the unknown parameters and hidden variables, i.e.,
Q x , θ ( x , θ ) = q x ( x ) q θ ( θ ) = i = 1 I q x ( x i ) j = 1 J q θ ( θ j ) ,
where I and J are the number of unknown parameters and hidden variables, respectively. This independence assumption in VB further simplifies the search for a closed-form solution to the approximation of the actual posterior. We desire to select the variational distribution Q x , θ ( x , θ ) as close as possible to p ( x , θ | y ) , where the closeness metric for distribution Q x , θ ( x , θ ) is formulated as minimizing the Kullback–Leibler (KL) divergence of the approximation Q x , θ ( x , θ ) and the true posterior p ( x , θ | y ) as
Q x , θ 🟉 ( x , θ ) = arg min Q x , θ ( x , θ ) KL Q x , θ ( x , θ ) | | p ( x , θ | y ) = arg min Q x , θ ( x , θ ) Q x , θ ( x , θ ) log Q x , θ ( x , θ ) p ( x , θ | y ) d x d θ .
The quantity log p ( y ) can be written as log p ( y ) = log { p ( x , θ , y ) d x d θ } . Then, defining
F Q x , θ ( x , θ ) = Q x , θ ( x , θ ) log p ( x , θ , y ) Q x , θ ( x , θ ) d x d θ .
It is straightforward to show that
log p ( y ) = F Q x , θ ( x , θ ) + KL Q x , θ ( x , θ ) , p ( x , θ | y ) .
Since (by Jensen’s inequality) K L ( Q x , θ ( x , θ ) 0 , log ( p ( y ) F ( Q x , θ ( x , θ ) . Since log ( p ( y ) is constant with respect to Q x , θ , minimizing the KL-divergence between the actual posterior distribution and the variational distribution is equivalent to maximizing the lower bound F ( · ) [66,67]. Since the term p ( y ) in p ( x , θ | y ) = p ( x , θ , y ) p ( y ) does not involve the variational distribution Q x , θ ( x , θ ) , this term can be ignored when maximizing F ( · ) . The lower bound F ( · ) on the model log-marginal likelihood can be iteratively optimized until the convergence by the following update rules [66,72].
VB-E step: 
q x [ t + 1 ] ( x ) exp { E q θ [ t ] [ log p ( x , y | θ ) ] }
VB-M step: 
q θ [ t + 1 ] ( θ ) p ( θ ) exp { E q x [ t + 1 ] [ log p ( x , y | θ ) ] }
This results in an iterative algorithm analogous to the expectation-maximization (EM) approach.

3. Bernoulli–Gaussian-Inverse Gamma Modeling and SBL(BGiG) Algorithm

In the inverse problem of CS defined in (1), the goal is to recover the sparse vector x s . In the Bernoulli–Gaussian-inverse Gamma model, the sparse solution is defined as
x s = ( s x ) ,
where s is a binary support vector indicating the non-zero locations in the solution, x represents values of the solution and ∘ is Hadamard (element-by-element) product [47]. We refer to the algorithm associated with this Bayesian modeling based on VB inference as SBL(BGiG). SBL using VB inference for the clustered pattern of sparse signals has already been investigated in the recent literature [45,50,51,58]. In this paper, however, we intend to focus on the ordinary SBL using VB inference modeling without promoting any structure on the supports other than sparsity itself. We show that when the sampling ratio is moderate or low (with respect to the sparsity level), the reconstruction performance becomes sensitive to selecting the support-related hyperparameters.
We define a set of priors as follows [47,68,69]. We model the elements of vector s as
s n Bernoulli ( γ n ) , γ n Beta ( α 0 , β 0 ) , n ,
where α 0 and β 0 are the support-related hyperparameters. Setting α 0 and β 0 to small values and with α 0 β 0 encourages s to be sparse on average. The prior on the solution value vector is defined as
x N ( 0 , τ 1 I N ) , τ Gamma ( a 0 , b 0 ) .
Here, τ is the precision value. Finally, the prior on the noise is
e N ( 0 , ϵ 1 I M ) , ϵ Gamma ( θ 0 , θ 1 ) ,
where θ 0 and θ 1 are set to small positive values.

3.1. Update Rules of SBL(BGiG) Using VB Inference

According to the VB algorithm defined in (2) and (3), the update rule of the variables and parameters of the BGiG model can be simplified as follows [68]. The details of these derivations appear in Appendix A.1.
  • Update rule for the support vector s
q ( s n | ) Bernoulli ( 1 1 + c n κ n ) , n = 1 , , N ,
where conditioning on − denotes conditioning on all relevant variables and observations. Therefore,
s ˜ n = 1 1 + c n κ n , n = 1 , , N ,
where
c n : = e ψ ( β 1 , n ) ψ ( α 1 , n ) , κ n : = e 1 2 ϵ ˜ a n 2 2 ( x n ˜ 2 + σ x n ˜ 2 ) 2 x n ˜ a n T y ˜ n , y ˜ m n : = y m l n N a m l s ˜ l x ˜ l .
Here, x ˜ : = < x > q x , ψ is the digamma function (the logarithmic derivative of the gamma function), and y ˜ n = [ y ˜ 1 n , , y ˜ M n ] T .
  • Update rule for the solution value matrix x
q ( x | ) N ( x ˜ , Σ x ˜ ) ,
where
Σ x ˜ = τ ˜ I N + ϵ ˜ Φ ˜ 1 and x ˜ = ϵ ˜ Σ x ˜ diag ( s ˜ ) A T y ,
and where diag ( s ) denotes a diagonal matrix with the components of s on its main diagonal, and
Φ ˜ : = ( A T A ) s ˜ s ˜ T + diag ( s ˜ ( 1 s ˜ ) ) .
  • Update rule for γ n
q ( γ n | ) Beta ( α 1 , n , β 1 , n ) , n = 1 , , N .
Therefore,
γ n ˜ = α 1 , n α 1 , n + β 1 , n , n = 1 , , N ,
where
α 1 , n : = α 0 + s ˜ n and β 1 , n : = β 0 + 1 s ˜ n .
  • Update rule for the solution precision τ
q ( τ | ) Gamma a 0 + N 2 , b 0 + 1 2 ( x ˜ 2 2 + Tr ( Σ x ˜ ) ) ,
where Σ x ˜ = diag ( σ x ˜ 1 2 , , σ x ˜ N 2 ) and Tr ( A ) is the trace of matrix A. Thus
τ ˜ = a 0 + N 2 b 0 + 1 2 x ˜ 2 2 + n = 1 N σ x ˜ n 2 .
  • Update rule for the noise precision ϵ
q ( ϵ | ) Gamma ( θ 0 + M 2 , θ 1 + 1 2 Ψ ˜ ) ,
where
Ψ ˜ : = y T y 2 ( x ˜ s ˜ ) T A T y + Tr ( x ˜ x ˜ T + Σ x ˜ ) Φ ˜ .
This yields to the following update rule for the precision of the noise component
ϵ ˜ = θ 0 + M 2 θ 1 + 1 2 Ψ ˜ .
The stopping criterion of the algorithm is made based on the log-marginalized likelihood. We define the stopping condition in terms of L : = log { p ( y | s , ϵ , τ ) } . The marginalized likelihood can be written as
p ( y | s , ϵ , τ ) = p ( y | x , s , ϵ ) p ( x | τ I N ) d x .
After some simplification, the negative log-likelihood is proportional to
L log | Σ 0 1 | + y T Σ 0 y ,
where
Σ 0 = ( ϵ ˜ 1 I M + τ ˜ 1 A S ˜ 2 A T ) 1
and S ˜ : = diag { s ˜ } . Therefore, the stopping condition can be made as
Δ L n [ t ] : = | Δ L [ t ] | / | L [ t 1 ] | T 0 ,
for some small value of threshold T 0 [50], where
L [ t ] : = log Σ 0 [ t ] y T Σ 0 [ t ] y .
and
Δ L [ t ] : = L [ t ] L [ t 1 ] = log | Σ 0 [ t ] Σ 0 [ t 1 ] | + y T ( Σ 0 [ t 1 ] Σ 0 [ t ] ) y .
Figure 1 illustrates the graphical Bayesian representation of the BGiG model, which is an undirected graph. The shaded node y shows the observations (measurements), and the small solid nodes represent the hyperparameters. Each unshaded node denotes a random variable (or a group of random variables).
The flowchart representation of the algorithm is shown in Figure 2 motivated by the graphical approach in [47,73]. According to the pseudocode in Algorithm 1 and the flowchart in Figure 2, first, the hyperparameters of the model are set. The support-related hyperparameters α 0 and β 0 are suggested to be set to small values with α 0 β 0 to encourage s to be sparse on the average. The hyperparameters a 0 and b 0 on the precision of the solution-value vector are also initialized and suggested to be small not to bias the estimation when the measurements are incorporated. The hyperparameters θ 0 and θ 1 on the precision of the noise are recommended to be of order 10 6 for high SNRs. For moderate and low SNRs, higher values are recommended. In the next step, all the main variables of the model are drawn i.i.d. from their corresponding prior distributions defined in (5)–(7). Then, the stopping condition is computed based on the log-marginalized likelihood in (19). In the main loop, all of the main variables of the model are updated via the expected values obtained from the VB inference. Specifically, we first update the support vector and the solution value components; then, the precisions of the solution vector and the noise are updated. Finally, the stopping criterion is computed through the measure of the log-marginalized likelihood of the observations. The pseudocode of the algorithm is provided below.
Algorithm 1: SBL(BGiG) Algorithm
x ^ s = x ˜ s ˜
x ˜ , s ˜ = SBL BGiG ( Y , A )
Set the hyperparameters, i.e., ( α 0 , β 0 ), ( a 0 , b 0 ), and ( θ 0 , θ 1 )
% Variables Initialization
Draw s ˜ and γ ˜ from (5)
Draw x ˜ and τ ˜ from (6)
Draw ϵ ˜ from (7)
t = 1     % Iterator
Compute L [ t ] from (19) and set L [ 0 ] = 0
% Main Loop for Estimations
While | L [ t ] L [ t 1 ] | | L [ t 1 ] | T 0 . For example T 0 = 10 6 .
       Compute s ˜ n from (8), n = 1 , , N                         % (Support vector component )
       Compute Σ x ˜ and x ˜ from (10)                                   % (Solution-value matrix component)
       Compute α 1 , n and β 1 , n from (13) n = 1 , , N        % (Parameters of the hyperprior γ )
       t Compute τ ˜ from (14)                                         % (Precision on the solution)
       Compute ϵ ˜ from (16)                                            % (Precision on the noise)
       Compute L [ t ] from (19) and then t = t + 1
End While

3.2. Issues with SBL(BGiG)

In this section, we show that the estimated solution using SBL(BGiG) algorithm is sensitive to support-related hyperparameters, i.e.,  α 0 and β 0 in (5). We provide an example under three cases to demonstrate this issue. We generated a random scenario, where the true solution x s R 100 has the sparsity level of k = 25 , that is, the true x (or s ) has k active elements. The active elements of s were drawn randomly. The nonzeros of x s , corresponding to the active locations of s , were drawn from N ( 0 , σ x 2 ) , with  σ x 2 = 1 . Each entry of the sensing matrix A was drawn i.i.d. from the Gaussian distribution N ( 0 , 1 ) , then normalized, so each column has the Euclidian norm of 1. The elements of measurement noise were drawn from N ( 0 , σ 2 ) with S N R = 25 dB, where S N R : = 20 log 10 ( σ x / σ ) . The hyperparameters of τ and ϵ were set to a 0 = b 0 = 10 3 and θ 0 = θ 1 = 10 6 , respectively. In Cases 1–3, we set the pair ( α 0 , β 0 ) with low emphasis on the prior ( 0.01 , 0.99 ) , moderate emphasis ( 0.1 , 0.9 ) and fairly high emphasis ( 1.4 , 2 ) , respectively.
From the top to the bottom row of Figure 3, Figure 4 and Figure 5, we illustrate the estimated results with the number of measurements set to 80, 60 and 40 (that is, the sample ratio λ is 0.80, 0.60, and 0.40), respectively. In each row of Figure 3, Figure 4 and Figure 5 from left to right, we show the comparison between the measurements y and the computed measurements based on y ^ = A ( s ˜ x ˜ ) , the true signal x s = s x and the reconstructed signal x ^ s = s ˜ x ˜ , the true support vector s and the estimated support vector s ˜ and the evolution of the estimated supports with respect to the iterations in the SBL(BGiG) algorithm.
According to Figure 3, the setting for ( α 0 , β 0 ) in Case 1 fails to provide perfect results even for high sampling ratios. Similarly, Figure 4 shows that the settings for ( α 0 , β 0 ) in Case 2 do not provide encouraging results even for high sampling ratios. Specifically, it turns out that Case 1 and Case 2 provide sparse solutions for the sampling ratios within the range [ 0 , 1 ] , where λ = 1 means M = N .
According to Figure 5, setting ( α 0 , β 0 ) to ( 1.4 , 2 ) seems to be a reasonable choice for high sampling ratios (over 70 % ), while it is not a good choice for the lower sampling ratios. This issue can be seen in the supports plot in the 2nd and 3rd row of Figure 5. One may argue that the estimated support vector s ^ can be filtered via some threshold value (such as 0.3 ) for λ = 0.6 . However, thresholding will adversely affect the detection rate, and setting the threshold depends on our understanding of the signal characteristics. Furthermore, we should account for the effect of the filtered supports since their corresponding estimated components in x ^ s contribute to fitting the model to the measurements.
In Table 1, we summarize the performance of the generated example for Cases 1–3, where P D , P F A and  N M S E denote the detection rate and false alarm rate in support recovery and the normalized mean-squared error between the true and the estimated sparse signal. This also shows that the algorithm fails to provide reasonable results for the sampling ratio of λ = 0.4 .
These experiments suggest that there is no fixed setting for ( α 0 , β 0 ) capable of performing reasonably well for all sampling ratios and thus, selecting the hyperparameters ( α 0 , β 0 ) should be made with care.
Continuing this examination, in Figure 6, Figure 7 and Figure 8, we illustrate the negative log-marginalized likelihood, the noise precision estimation and the estimated precision on the generated true solution in Cases 1–3, respectively. The horizontal axis shows the iterations until the stopping rule is met.
As expected, as the sampling ratio increases, the algorithm requires fewer iterations to meet its stopping condition. This can be seen on the negative log-marginalized likelihood plots in Figure 6, Figure 7 and Figure 8. In these experiments, the actual precision of the solution components was set to τ = 1 , and the actual noise precision was set to ϵ = 316.2 .
For Cases 1 and 2, according to Figure 6, Figure 7 and Figure 8, the estimated precisions on both the noise and solution components were far off from the actual ones even for λ = 0.8 . Thus, it resulted in poor performance in signal recovery for Cases 1 and 2 (see Figure 3 and Figure 4).
For Case 3, the estimated precisions on the noise and the solution components were acceptable for λ = 0.8 but far off from the actual ones for lower sampling ratios (see Figure 8). The main issue of the failures can be found in the update rule of the support learning vector s ˜ defined in (8). It is important to balance between the terms c n and κ n , where c n imposes the effect of hyper-prior on s accompanied by the current estimate of s n . In contrast, κ n imposes the contribution of the current estimates of noise precision, solution and other supports in fitting the model to the measurements. Therefore, if we impose a substantial weight on the sparsity via c n , the solution tends to neglect the effect of κ n and vice versa. This is why we had sparse (with poor performance) in Cases 1 and 2 for all the represented sampling ratios and nonsparse (with poor performance) for moderate and lower sampling ratios in Case 3. These results suggest that the algorithm and its update rules are sensitive to the selection of hyperparameters on the Gamma prior on the support vector s . The main issue can be seen in (9), where the selection of the hyperparameters α 0 and β 0 resulted in a large or small value in c n due to the digamma function.

4. Gaussian-Inverse Gamma Modeling and SBL(GiG) Algorithm

In this section, we consider the Gaussian-inverse Gamma (GiG) model. In this model, each component x n of the solution is modeled by zero-mean Gaussian with the precision τ n . The main difference between this model and the model defined in Section 3 is that the GiG model does not have the support vector s ; instead, different precisions are considered on the components of the solution vector x s in (1). A simpler version of GiG can also be used by defining the same precision τ for all the components of x s .
Here, we rather use different precisions to make the GiG model have almost the same complexity as the BGiG model in terms of the parameters to be learned. The set of priors in this model is defined as follows.
x n N ( 0 , τ n 1 ) , τ n Gamma ( a 0 , b 0 ) , n ,
where a 0 and b 0 denote the shape and rate of the Gamma distribution, respectively. The entries of the noise component e are defined the same as (7), i.e., 
e N ( 0 , ϵ 1 I M ) , ϵ Gamma ( θ 0 , θ 1 ) ,
where θ 0 and θ 1 are set to small positive values. The estimation of the parameters in this model is carried out using VB inference, as discussed below.

4.1. Update Rules of SBL(GiG) Using VB Inference

According to the VB algorithm described in (2) and (3), the update rule of the variables and parameters of the GiG model can be simplified as follows. The details of these derivations appear in Appendix A.2.
  • Update rule for the precision τ n on x n using VB
q ( τ n ) Gamma a 0 + 1 2 , b 0 + 1 2 ( x ˜ n 2 + σ x ˜ n 2 ) , n = 1 , , N .
Thus,
τ n ˜ = a 0 + 1 2 b 0 + 1 2 ( x ˜ n 2 + σ x ˜ n 2 ) , n = 1 , 2 , , N .
  • Update rule for the noise precision ϵ using VB
q ( ϵ ) Gamma ( θ 0 + M 2 , b 0 + 1 2 Ψ ˜ )
which yields
ϵ ˜ = θ 0 + M 2 θ 1 + 1 2 Ψ ˜ ,
where
Ψ ˜ : = y T y 2 x ˜ T A T y + Tr ( x ˜ x ˜ T + Σ x ˜ ) A T A .
  • Update rule for the solution vector x using VB
q x ( x ) N ( x ˜ , Σ x ˜ ) ,
where
Σ x ˜ : = ( T ˜ + ϵ ˜ A T A ) 1 and x ˜ : = ϵ ˜ Σ x ˜ A T y ,
and
T ˜ : = diag { [ τ ˜ 1 , , τ ˜ N ] } .
We set the stopping rule of the algorithm using the marginalized likelihood (evidence) defined as
p ( y | ϵ , τ ) = p ( y | x , ϵ , τ ) p ( x | τ ) d x .
After simplification and for the comparison purposes of L [ t ] with L [ t 1 ] in the updating process, we have
L [ t ] log | Σ 0 [ t ] | y T Σ 0 [ t ] y ,
where Σ 0 is defined as
Σ 0 : = ( ϵ ˜ 1 I M + T ˜ 1 A A T ) 1 .
Therefore, similar to SBL(BGiG), the stopping condition can be made as
Δ L n [ t ] : = | Δ L [ t ] | / | L [ t 1 ] | T 0 ,
for some small value of threshold T 0 .
Figure 9 illustrates the graphical Bayesian representation of the GiG model, which is an undirected graph. Similar to Figure 1, the shaded node y shows the observations, the small solid nodes represent the hyperparameters and the unshaded nodes denote the random variables.
The flowchart representation of the algorithm is shown in Figure 10. According to the pseudocode in Algorithm 2 and the flowchart in Figure 10, first, the hyperparameters of the model are set. The hyperparameters a 0 and b 0 on the precision of the solution-value vector are initialized and suggested to be small. Similar to SBL(BGiG), the hyperparameters θ 0 and θ 1 on the precision of the noise are recommended to be of order 10 6 for high SNRs. All the main variables of the model are drawn i.i.d. from their corresponding prior distributions defined in (22)–(26). Then, the stopping condition is computed based on (28). In the main loop, all the main variables of the model are updated via the expected values obtained from the VB inference through (22)–(26). The pseudocode of the algorithm is provided below.      
Algorithm 2: SBL(GiG) Algorithm
x ˜ s = SBL GiG ( Y , A )                 
Set the hyperparameters, i.e., ( a 0 , b 0 ) and ( θ 0 , θ 1 )
% Variables’ Initialization
Draw x ˜ s and ø ˜ from (21)
Draw ϵ ˜ from (7)
t = 1     % Iterator
Compute L ˜ [ t ] from (28) and (27), and set L ˜ [ 0 ] = 0
% Main Loop for Estimations
t = 1
While | L [ t ] L [ t 1 ] | | L [ t 1 ] | T 0 . For example T 0 = 10 6 .
       Compute Σ x ˜ and x ˜ s from (26)          % (Solution-value matrix component)
       Compute T ˜ from (22)                    % (Precisions on the solution)
       Compute ϵ ˜ from (23)                     % (Precision on the noise)
       Compute L [ t ] from (28) and (27), and then t = t + 1
End While

4.2. Issues with SBL(GiG)

An issue with the SBL(GiG) algorithm is that the solution becomes nonsparse since it does not incorporate a binary vector s (hard-thresholding or soft-thresholding if the expected value is used) as we had in SBL(BGiG). This may have no major effect on the signal reconstruction for high sampling ratios. However, the nonsparseness effect appears in low sampling ratios by misleading the algorithm to wrongly activate many components in the estimated signal yet providing a good fit of the model to the measurements. Here, we use the same example as we made for the SBL(BGiG) model with the same sensing matrix A, measurement vector y and noise e . Notice that in the SBL(BGiG) model, we considered the same precision τ on all the components of the solution value vector x support vector s . In contrast, the SBL(GiG) model does not have the support learning vector; instead, we assume that each component of the solution vector has different precision τ n . It turns out that SBL(GiG) is not very sensitive to the selection of the hyperparameters as the SBL(BGiG). Thus, here, we show the results for one case scenario for the hyperparameters. We use the same setting for the parameters of ϵ in the hyper prior as before, i.e., θ 0 = θ 1 = 10 6 , and the same parameters for all the precisions τ n of the solution component, i.e., a 0 = b 0 = 10 3 . In Figure 11 and Figure 12, we illustrate the results after applying the SBL(GiG) algorithm. In Figure 11, from left to right, we show the results for sampling ratios of λ = 0.8, 0.6, and 0.40, respectively. The first row shows the comparison of y with y ^ = A x ˜ s , the second row shows the true solution x s and the estimated solution x ˜ s , and the third row demonstrates the estimated precisions on the solution components. In Figure 12, we demonstrate the negative log-marginalized likelihood comparison and the estimated noise precision against the true noise precision for the sampling ratios of λ = 0.8, 0.6 and 0.4.
From the results shown in Figure 11 and Figure 12, we observe that the recovered signal tends to become nonsparse. This effect is illustrated in the second row of Figure 11. This can also be observed in the precision estimations of the solution components. More specifically, the true nonzero components in our simulations were drawn from a zero-mean Gaussian with the precision of τ n = 1 . Thus, the ideal precision estimation would be within the two classes of values of 1 and infinity or very large values. However, the estimated results in our simulation do not show such a classification. As the sampling ratio decreases, the solution estimate has poor performance, due not only to the reduction in the number of measurements but also the nonsparseness behavior.

5. Preprocessing versus Postprocessing and Simulations

In this section, we show that in order to improve the performance of Bernoulli–Gaussian-inverse Gamma modeling using the SBL(BGiG) algorithm, we need to perform a preprocessing step. The results in Section 4 suggest one can perform some postprocessing for the SBL(GiG) algorithm to improve the reconstruction performance. Below, we provide more details for each of these algorithms.

5.1. Pre-Processing for the SBL(BGiG) Algorithm

Based on the observations made on the performance of SBL(BGiG) in Section 3.2, we showed that the pair of hyperparameters ( α 0 , β 0 ) should be selected with care. In other words, obtaining good performance with this algorithm needs some preprocessing to assess an appropriate setting for the parameters. For a more rigorous study, here, we perform a grid search on the hyperparameters ( α 0 , β 0 ) to see whether we can find some common pattern in selecting these parameters for all sampling ratios. The grid search runs the algorithm for different values of α 0 and β 0 with the search range of [ 0.1 , 2 ] with the resolution of 0.1 . For each ( α 0 , β 0 ) within this range, we ran 200 random trials and then averaged the results. The settings of these trials are represented in Table 2.
We generated a random scenario, where the true solution x s R 100 has the sparsity level of k = 25 . The active elements of s were drawn randomly. The nonzeros of x s were drawn from N ( 0 , σ x 2 ) , with σ x 2 = 1 . Each entry of the sensing matrix A was drawn i.i.d. from the Gaussian distribution N ( 0 , 1 ) , then normalized. The elements of measurement noise were drawn from N ( 0 , σ 2 ) with S N R = 25 dB. The results were examined to see what values of ( α 0 , β 0 ) provided the highest performance in the detection rate vs. and false alarm rate. The simulation was executed for a range of sampling ratios in the range [ 0.05 , 1 ] with the step size of 0.05. The results are demonstrated in Figure 13. In this figure, we also provide the results of performing a random Sobol search for ( α 0 , β 0 ) . A Sobol sequence is a low discrepancy quasirandom sequence. The two right plots in Figure 13 show the results for the best setting of ( α 0 , β 0 ) .
It should be clear from Figure 13 that there is no fixed setting for these parameters in order to get the best performance for all sampling ratios. The two plots on the right of Figure 13 illustrate the performance based on the best values of these hyperparameters, which provided the best performance, i.e., tuned hyperparameters. We also examined the grid search results for the top 10 highest performances for each sampling ratio, where performance is in terms of P D P F A and the normalized mean-squared error (NMSE). In Figure 14a, we demonstrate the top 10 highest performances based on NMSE and P D P F A for different sampling ratios. In Figure 14b,c, we illustrate the values of ( α 0 , β 0 ) , which led to the performances shown in Figure 14a for different sampling ratios. Figure 15 details the top 10 values of ( α 0 , β 0 ) vs. sampling ratio.
According to Figure 14b,c, there is no specific pattern for these hyperparameters. Figure 15 also shows that hyperparameters need to be carefully selected.

5.2. Post-Processing for the SBL(GiG) Algorithm

Since the SBL(GiG) algorithm does not include the binary support vector s , as SBL(BGiG) possesses, the resulting solution tends to become nonsparse. This leads to a high detection rate for the location of active supports and a high false alarm rate. Thus, as the sampling ratio decreases, there is a high chance that this algorithm overwhelms the locations of the true solution. Therefore, SBL(GiG) requires some postprocessing to discard the components with low amplitudes. This problem becomes of great importance for applications where detecting the correct nonzero components is more crucial than the magnitudes of the nonzeros in the signal. This effect can be seen in Figure 16b. The curves with solid lines in this plot show the detection and false alarm rate in support recovery and the difference between the rates. This issue can be resolved by some postprocessing such as data-driven threshold tuning. That way, the amplitudes in the reconstructed signal with lower values than the threshold can be discarded. For this purpose, we set up 200 random trials, the same way as the one explained for SBL(BGiG), and then evaluate the performance in terms of NMSE by varying the threshold. Figure 16b shows the averaged results of 200 trials. The settings of these trials are represented in Table 3.
In Figure 16a, we observe that the postprocessing does not benefit us so much in terms of the reconstruction error for low and moderate sampling ratios. However, there is a threshold of around 0.25, for which the postprocessing step reduced the reconstruction error by approximately 3 dB. We set the threshold to 0.25 and ran 200 random trials by applying SBL(GiG) and evaluating the performance based on the detection and false alarm rate in support recovery. According to Figure 16b, the additional post-processing step provides reasonable performance.
Finally, in Figure 17, we compare the performance of the SBL(BGiG) algorithm (with performing the preprocessing step) with the SBL(GiG) algorithm (after performing postprocessing). We see that Bernoulli–Gaussian-inverse Gamma implemented via SBL(BGiG) provides better performance for low and high sampling ratios. In contrast, Gaussian-inverse Gamma modeling implemented via SBL(GiG) performs much better for the moderate sampling ratios.

6. Conclusions

We investigated solving the inverse problem of compressive sensing using VB inference for two sparse Bayesian models of Bernoulli–Gaussian-inverse Gamma (BGiG) and Gaussian-inverse Gamma (GiG). The issues of each approach were discussed and the performance between the two models was compared. Specifically, we showed the behavior of these models and algorithms when the sampling ratio is low and moderate as well as the importance of selecting the hyperparameters of BGiG model with care. We further provided some intuition for performing additional pre/post-processing steps, depending on the selected model for better performance.
Based on our study on the synthetic data and considering the overall performance of both algorithms and the complexity in additional pre-/postprocessing, we observed that for moderate sampling ratios, SBL(GiG) is performing better than SBL(BGiG) modeling when using VB for sparse signals with no specific pattern in the supports. In contrast, SBL(BGiG) provided better perfomance for low and high sampling ratios. Finally, a rigorous comparison is required to study in the future under real-world scenarios and various applications. The MATLAB codes for GiG and BGiG modeling are available at https://github.com/MoShekaramiz/Compressive-Sensing-GiG-versus-BGiG-Modeling.git, accessed on 15 December 2022.

Author Contributions

Methodology, M.S. and T.K.M.; Formal analysis, M.S. and T.K.M.; Investigation, M.S.; Resources, M.S.; Writing—original draft, M.S.; Writing—review & editing, M.S. and T.K.M.; Visualization, M.S.; Supervision, T.K.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

In this section, we provide details on deriving the update rules of the parameters and variables for both models and the associated algorithms.

Appendix A.1. Bernoulli–Gaussian-Inverse-Gamma Modeling and the SBL(BGiG)

  • Update rule for the precision τ of the solution value vector x
q ( τ ) p ( τ ; a 0 , b 0 ) e ( < log p ( x | τ I N ) > q x ) τ a 0 1 e b 0 τ e < log n = 1 N p ( x n ; τ 1 ) > q x τ a 0 1 e b 0 τ e < log { τ N 2 e { τ 2 x 2 2 } } > q x τ ( a 0 + N 2 ) 1 e ( b 0 + 1 2 < x 2 2 > q x ) τ ,
where
< x 2 2 > q x = < x T x > q x = Tr ( < x x T > q x ) = x ˜ 2 2 + n = 1 N σ x ˜ n 2 ,
and x ˜ : = < x > q x . Therefore,
q ( τ ) Gamma a 0 + N 2 , b 0 + 1 2 ( x ˜ 2 2 + n = 1 N σ x ˜ n 2 ) .
Finally, considering the point estimate on τ as the expected value of the Gamma distribution in q ( τ ) , the update rule for τ can be defined as
τ ˜ = a 0 + N 2 b 0 + 1 2 x ˜ 2 2 + n = 1 N σ x ˜ n 2 .
  • Update rule for the noise precision ϵ
q ( ϵ ) p ( ϵ ; θ 0 , θ 1 ) e < log p ( y | x , s , ϵ ) > q x q s ϵ θ 0 1 e θ 1 ϵ e < log ϵ M 2 e 1 2 ϵ y A ( s x ) 2 2 > q x q s ϵ ( θ 0 + M 2 ) 1 e ( θ 1 + 1 2 < y A ( s x ) 2 2 > q x q s ) ϵ ,
where
< y A ( s x ) 2 2 > q x q s = < y A S x ) 2 2 > q x q s = y T y 2 < x T S A T y > q x q s + < x T S A T A S x > q x q s = y T y 2 < x > q x T < S > q s A T y + < x T S A T A S x > q x q s = y T y 2 ( x ˜ s ˜ ) T A T y + < x T M s x > q x q s ,
where S = diag { s } , M s : = S A T A S , and
< x T M s x > q x q s = Tr ( < x x T > q x < M s > q s ) = Tr ( x ˜ x ˜ T + Σ x ˜ ) < M s > q s ,
where Σ x ˜ = diag { σ x 1 ˜ 2 , , σ x N ˜ 2 } , and
< M s > s = < S A T A S > q s = < ( A T A ) ( s s T ) > q s = ( A T A ) < ( s s T ) > q s = ( A T A ) s ˜ s ˜ T + diag s ˜ ( 1 s ˜ ) .
Therefore,
< x T M s x > q x q s = Tr ( x ˜ x ˜ T + Σ x ˜ ) ( A T A ) ( s ˜ s ˜ T + diag { s ˜ ( 1 s ˜ ) } ) .
As a result,
q ( ϵ ) Gamma θ 0 + M 2 , θ 1 + 1 2 Ψ ˜ ,
where
Ψ ˜ : = < y A ( s x ) 2 2 ) > q x q s = y T y 2 ( x ˜ s ˜ ) T A T y + Tr ( x ˜ x ˜ T + Σ x ˜ ) ( A T A ) ( s ˜ s ˜ T + diag { s ˜ ( 1 s ˜ ) } ) .
Finally, the update rule for the precision of the noise can be written as
ϵ ˜ = θ 0 + M 2 θ 1 + 1 2 Ψ ˜ ,
Remark A1.
Notice that Tr ( X T Y ) = i , j ( X Y ) i j = 1 T ( X Y ) 1 . Therefore,
Tr ( x ˜ x ˜ T + Σ x ˜ ) ( ( A T A ) ( s ˜ s ˜ T + diag { s ˜ ( 1 s ˜ ) } ) ) = 1 T ( x ˜ x ˜ T + Σ x ˜ ) ( A T A ) ( s ˜ s ˜ T + diag { s ˜ ( 1 s ˜ ) } ) ) 1 ,
where 1 = [ 1 , , 1 ] T . Thus, Ψ ˜ can be written as
Ψ ˜ : = y T y 2 ( x ˜ s ˜ ) T A T y + 1 T ( x ˜ x ˜ T + Σ x ˜ ) ( A T A ) ( s ˜ s ˜ T + diag { s ˜ ( 1 s ˜ ) } ) ) 1 .
  • Update rule for γ n
q ( γ n ) p ( γ n ; α 0 , β 0 ) e ( < log { p ( x , s , y | θ ) } > q x q s ) γ n α 0 1 ( 1 γ n ) β 0 1 e { < log { p ( s n | γ n ) } > q x q s } γ n α 0 1 ( 1 γ n ) β 0 1 e { < log { γ n s n ( 1 γ n ) 1 s n } > q s n } γ n α 0 1 ( 1 γ n ) β 0 1 e < s n > q s n log { γ n } e ( 1 < s n > q s n ) log { 1 γ n } γ n α 0 1 ( 1 γ n ) β 0 1 γ n < s n > q s n ( 1 γ n ) 1 < s n > q s n γ n ( α 0 + s ˜ n ) 1 ( 1 γ n ) β 0 s ˜ n .
Therefore,
q γ n ( γ n ) Beta ( α 1 , n , β 1 , n ) , n = 1 , , N ,
where α 1 , n : = α 0 + s ˜ n and β 1 , n : = β 0 + 1 s ˜ n . Finally, the update rule for γ n can be defined as
γ ˜ n = α 1 , n α 1 , n + β 1 , n .
  • Update rule for the solution vector x
q x ( x ) e { < log { p ( x , s , y | θ ) } > q x q s } e { < log { p ( x , s | θ ) p ( y | x , s , θ ) } > q θ q s } e { < log { p ( x | θ ) } > q θ } e { < log { p ( y | x , s , θ ) } > q θ q s } e { < log { p ( x | τ ) } > q τ } e { < log { p ( y | x , s , ϵ ) } > q ϵ q s } .
To update the elements of x , we have
p ( y | x , s , ϵ ) e { 1 2 ϵ y A ( s x ) 2 2 } .
Therefore,
< log { p ( y | x , s , ϵ ) } > q ϵ q s 1 2 < ϵ y A ( s x ) 2 2 > q ϵ q s 1 2 < ϵ > q ϵ < y A ( s x ) 2 2 > q s 1 2 ϵ ˜ < | y A ( s x ) 2 2 > q s
and
< y A ( s x ) 2 2 > q s = < Tr ( y y T + ( x s ) T A T A ( x s ) 2 ( x s ) T A T y ) > q s < Tr ( x s ) T A T A ( x s ) 2 ( x s ) T A T y > q s < Tr x T S T A T A S x 2 x T S A T y > q s Tr x T < S A T A S > q s x 2 x T S ˜ A T y .
This yields to
< y A ( s x ) 2 2 > q s x T < S A T A S > q s x 2 x T S ˜ A T y ,
which results in
< log p ( y | x , s , ϵ ) > q ϵ q s 1 2 ϵ ˜ ( x T < S A T A S > q s x 2 x T S ˜ A T y ) .
Thus, we can write q x ( x ) as
q x ( x ) e < log { p ( x | τ ) } > q τ e < log { p ( y | x , s , ϵ ) } > q ϵ q s e { 1 2 τ ˜ x T x } e { 1 2 ϵ ˜ ( x T < S A T A S > q s x 2 x T S ˜ A T y ) } e { 1 2 ( x T ( τ ˜ I N + ϵ ˜ > S A T A S > q s ) x 2 ϵ ˜ x T S ˜ A T y ) } .
Notice that S A T A S = ( A T A ) ( s s T ) . Since s n is drawn from a Bernoulli distribution, we have < s n 2 > q s = < s n > q s = s ˜ n , and
s ˜ s ˜ T = s ˜ 1 2 s ˜ 1 s ˜ 2 s ˜ 1 s ˜ n s ˜ n s ˜ 1 s ˜ n s ˜ 2 s ˜ n 2
Therefore,
< S T A T A S > q s = ( A T A ) ( s ˜ s ˜ T diag { s ˜ s ˜ } + diag { s ˜ } ) = ( A T A ) ( s ˜ s ˜ T + diag { s ˜ ( 1 s ˜ ) } ) ,
which yields to
q x ( x ) N ( x ˜ , Σ x ˜ ) ,
where
Σ x ˜ = τ ˜ I N + ϵ ˜ ( A T A ) ( s ˜ s ˜ T + diag { s ˜ ( 1 s ˜ ) } ) 1
and
x ˜ = ϵ ˜ Σ x ˜ S ˜ A T y ,
which x ˜ is the update rule for the solution value vector x .
  • Update rule for the support vector s
q s n ( s n ) e { < log { p ( x , s , y | θ ) } > q θ q x } e { < log { p ( x , s | θ ) p ( y | x , s , θ ) } > q θ q x } e { < log { p ( s n ; γ n ) } > q γ n } e { < log { p ( y | x , s , ϵ ) } > q s n q x q ϵ } e { < log { γ n s n ( 1 γ n ) 1 s n } > q γ n } e { < log { p ( y | x , s , ϵ ) } > q s n q x q ϵ } ,
where
e < log { γ n s n ( 1 γ n ) 1 s n } > q γ n = e s n < log { γ n } > q γ n e ( 1 s n ) < log { 1 γ n } > q γ n ,
for which
< log γ n > q γ n Beta ( α 1 , n , β 1 , n ) = ψ ( α 1 , n ) ψ ( α 1 , n + β 1 , n )
and
< log { 1 γ n } > q γ n Beta ( α 1 , n , β 1 , n ) = ψ ( β 1 , n ) ψ ( α 1 , n + β 1 , n ) ,
where ψ ( · ) is digamma function, the logarithmic derivative of the gamma function, i.e., ψ ( x ) = d d x log Γ ( x ) . Therefore,
e < log { γ n s n ( 1 γ n ) 1 s n } > q γ n = e s n ψ ( α 1 , n ) ψ ( α 1 , n + β 1 , n ) e ( 1 s n ) ψ ( β 1 , n ) ψ ( α 1 , n + β 1 , n ) .
Also,
e < log { p ( y | x , s , ϵ ) } > q s n q x q ϵ e 1 2 < ϵ y A ( s x ) 2 2 > q s n q x q ϵ e 1 2 ϵ ˜ < y A ( s x ) 2 2 > q s n q x e 1 2 ϵ ˜ < m = 1 M ( y m n = 1 N a m n s n x n ) 2 > q s n q x e 1 2 ϵ ˜ < ( y 1 l n N a 1 n s n x n ) a 1 n s n x n 2 + + ( y M l n N a M l s l x l ) a M n s n x n 2 > q s n q x ,
where y m n : = y m l n N a m l s l x l , m = 1 , 2 , , M . Therefore,
e < log { p ( y | x , s , ϵ ) } > q s n q x q ϵ e 1 2 ϵ ˜ < m = 1 M ( a m n s n x n y m n ) 2 > q s n q x e 1 2 ϵ m = 1 M a m n 2 s n 2 < x n 2 > q x 2 a m n s n < x n y m n > q s n q x e 1 2 ϵ ˜ a n 2 2 s n 2 < x n 2 > q x 2 m = 1 M a m n s n < x n y m n > q x , q s n e 1 2 ϵ ˜ a n 2 2 s n 2 ( x ˜ n 2 + σ x ˜ n 2 ) 2 s n x n ˜ m = 1 M a m n < y m n > q s n q x e 1 2 ϵ ˜ a n 2 2 ( x ˜ n 2 + σ x ˜ n 2 ) s n 2 2 s n x ˜ n a n T < y n > q s n q x ,
where y m n contains no x n component and
y n : = [ y 1 n , y 2 n , , y M n ] .
Thus,
< y m n > q s n q x = < y m l n N a m l s l x l > q s n q x = y m l n N a m l s ˜ l x ˜ l
which yields to
y ˜ m n : = < y m n > q s n q x y ˜ n = y l n N s ˜ l x ˜ l a l .
and thus
y ˜ n : = < y n > q s n q x .
Therefore,
e < log { p ( y | x , s , ϵ ) } > q s n q x q ϵ e 1 2 ϵ ˜ a n 2 2 ( x ˜ n 2 + σ x ˜ n 2 ) s n 2 2 ( x ˜ n a n T y ˜ n ) s n .
Finally,
q s n ( s n ) e s n ψ ( α 1 , n ) ψ ( α 1 , n + β 1 , n ) + ( 1 s n ) ψ ( β 1 , n ) ψ ( α 1 , n + β 1 , n ) 1 2 ϵ ˜ a n 2 2 ( x ˜ n 2 + σ x ˜ n 2 ) s n 2 2 x ˜ n a n T y ˜ n s n .
Since s n is an outcome of a Bernoulli random variable,
q s n ( s n = 0 ) e ψ ( β 1 , n ) ψ ( α 1 , n + β 1 , n )
and
q s n ( s n = 1 ) e { ψ ( α 1 , n ) ψ ( α 1 , n + β 1 , n ) 1 2 ϵ ˜ a n 2 2 ( x ˜ n 2 + σ x ˜ n 2 ) 2 x ˜ n a n T y ˜ n } .
Therefore,
q s n ( s n ) Bernoulli q s n ( s n = 1 ) q s n ( s n = 0 ) + q s n ( s n = 1 ) Bernoulli 1 1 + q s n ( s n = 0 ) q s n ( s n = 1 ) ,
which yields to
q s n ( s n ) Bernoulli 1 1 + e ψ ( β 1 , n ) ψ ( α 1 , n + β 1 , n ) e ψ ( α 1 , n ) + ψ ( α 1 , n + β 1 , n ) + 1 2 ϵ ˜ a n 2 2 ( x ˜ n 2 + σ x ˜ n 2 ) 2 x ˜ n a n T y ˜ n Bernoulli 1 1 + e ψ ( β 1 , n ) ψ ( α 1 , n ) + 1 2 ϵ ˜ a n 2 2 ( x ˜ n 2 + σ x ˜ n 2 ) 2 x ˜ n a n T y ˜ n .
The update rule for the component s n can then be written as
s ˜ n = 1 1 + e ψ ( β 1 , n ) ψ ( α 1 , n ) + 1 2 ϵ ˜ a n 2 2 ( x ˜ n 2 + σ x ˜ n 2 ) 2 x ˜ n a n T y ˜ n
or equivalently,
s ˜ n = 1 1 + c n κ n , n = 1 , , N ,
where
c n : = e { ψ ( β 1 , n ) ψ ( α 1 , n ) }
and
κ n : = e 1 2 ϵ ˜ a n 2 2 ( x ˜ n 2 + σ x ˜ n 2 2 ) 2 x ˜ n a n T y ˜ n .
  • Stopping rule
The stopping rule of the algorithm can be set based on the marginalized likelihood (evidence). We would rather follow the effect of s on the evidence because if s is learned, it would be easy to compute x s . Therefore, we marginalize the distribution on y and integrate x out. The details are described below.
p ( y | s , ϵ , τ ) = p ( y , x | s , ϵ , τ ) d x = p ( y | x , s , ϵ , τ ) p ( x | τ ) d x = 1 ( 2 π ϵ 1 ) M 2 e 1 2 ϵ y A ( s x ) 2 2 1 ( 2 π τ 1 ) N 2 e 1 2 τ x 2 2 d x = 1 ( 2 π ) M 2 ϵ M 2 τ N 2 1 ( 2 π ) N 2 e 1 2 ϵ y T y 2 ( s x ) T A T y + ( s x ) T A T A ( s x ) + τ x T x d x = 1 ( 2 π ) M 2 ϵ M 2 τ N 2 1 ( 2 π ) N 2 e 1 2 ϵ y T y 2 x T S A T y + x T S A T A S x + τ x T x d x = 1 ( 2 π ) M 2 ϵ M 2 τ N 2 e 1 2 ϵ y T y 1 ( 2 π ) N 2 e 1 2 x T ( ϵ S A T A S + τ I N ) x 2 ϵ x T S A T y d x
p ( y | s , ϵ , τ ) = 1 ( 2 π ) M 2 ϵ M 2 τ N 2 e 1 2 ϵ y T y | ( τ I N + ϵ S A T A S ) 1 | 1 2 1 ( 2 π ) N 2 1 | ( τ I N + ϵ S A T A S ) 1 | 1 2 × e 1 2 x T ( ϵ S A T A S + τ I N ) x 2 ϵ x T S A T y d x = 1 ( 2 π ) M 2 ϵ M 2 τ N 2 e 1 2 ϵ y T y | ( τ I N + ϵ S A T A S ) 1 | 1 2 1 ( 2 π ) N 2 1 | ( τ I N + ϵ S A T A S ) 1 | 1 2 × e 1 2 x ( τ I N + ϵ S A T A S ) 1 ϵ S A T y T ( τ I N + ϵ S A T A S ) x ( τ I N + ϵ S A T A S ) 1 ϵ S A T y ϵ 2 y T A S ( τ I N + ϵ S A T A S ) 1 S A T y d x ,
which results in
p ( y | s , ϵ , τ ) = 1 ( 2 π ) M 2 ϵ M 2 τ N 2 e 1 2 ϵ y T y | ( τ I N + ϵ S A T A S ) 1 | 1 2 e 1 2 ϵ 2 y T A S ( τ I N + ϵ S A T A S ) 1 S A T y .
Thus,
log p ( y | s , ϵ , τ ) = M 2 log { 2 π } + M 2 log ϵ + N 2 log τ 1 2 ϵ y T y + 1 2 log { | ( τ I N + ϵ S A T A S ) 1 | } + 1 2 ϵ 2 y T A S ( τ I N + ϵ S A T A S ) 1 S A T y
and
1 2 ϵ y T y + 1 2 ϵ 2 y T A S ( τ I N + ϵ S A T A S ) 1 S A T y = 1 2 y T I M ϵ A S ( τ I N + ϵ S A T A S ) 1 S A T y
Also,
N 2 log { τ } + 1 2 log { | ( τ I N + ϵ S A T A S ) 1 | } = 1 2 log { | ( τ I N ) ( τ I N + ϵ S A T A S ) 1 | } = 1 2 log { | ( τ 1 I N ) ( τ I N + ϵ S A T A S ) | } = 1 2 log { | I N + ϵ τ S A T A S | } = 1 2 log { | I M + ϵ τ A S 2 A T | } .
Thus,
L : = log p ( y | s , ϵ , τ ) = M 2 log { 2 π } + M 2 log { ϵ } 1 2 log | I M + ϵ τ A S 2 A T | 1 2 ϵ y T I M ϵ A S ( τ I N + ϵ S A T A S ) 1 S A T y .
For comparing the changes of L [ t ] with L [ t 1 ] in the updating process, we have
L M 2 log { ϵ } + 1 2 log { | ( I M + ϵ τ A S 2 A T ) 1 | } 1 2 ϵ y T I M ϵ A S ( τ I N + ϵ S A T A S ) 1 S A T y 1 2 log { | ϵ I M | } + log { | ( I M + ϵ τ A S 2 A T ) 1 | } 1 2 y T ( ϵ 1 I M + 1 τ A S 2 A T ) 1 y 1 2 log { | ϵ 1 I M | 1 | I M + ϵ τ A S 2 A T | 1 } 1 2 y T ( ϵ 1 I M + 1 τ A S 2 A T ) 1 y 1 2 log { | ϵ 1 I M ( I M + ϵ τ A S 2 A T ) 1 | } 1 2 y T ( ϵ 1 I M + 1 τ A S 2 A T ) 1 y log { | ϵ 1 I M + 1 τ A S 2 A T | 1 } y T ( ϵ 1 I M + 1 τ A S 2 A T ) 1 y .
Therefore,
L [ t ] log | Σ 0 [ t ] | y T Σ 0 [ t ] y ,
where
Σ 0 : = ( ϵ ˜ 1 I M + τ ˜ 1 A S ˜ 2 A T ) 1 ,
which yields to
L log { | Σ 0 1 | } + y T Σ 0 y .
This means that
p ( y | S , ϵ , τ ) = 1 ( 2 π ) M 2 1 | Σ 0 1 | 1 2 e { 1 2 y T Σ 0 y }
or equivalently,
p ( y | s , ϵ , τ ) N ( 0 , Σ 0 1 ) .
Therefore, the stopping criterion can be made based on
Δ L [ t ] : = L [ t ] L [ t 1 ] = log { Σ 0 [ t ] Σ 0 [ t 1 ] } + y T ( Σ 0 [ t 1 ] Σ 0 [ t ] ) y .

Appendix A.2. Gaussian-Inverse-Gamma Modeling and the SBL(GiG)

  • Update rule for the precision τ n of the nth component of the solution vector x
q ( τ n ) p ( τ n ; a 0 , b 0 ) e ( < log p ( x | T ) > q x n ) τ n a 0 1 e b 0 τ n e { < log n = 1 N p ( x n ; τ n 1 ) > q x n } τ n a 0 1 e b 0 τ n e { < log { τ n 1 2 e τ n 2 x n 2 } > q x n } τ n a 0 + 1 2 1 e b 0 τ n e { τ n 2 < x n 2 > q x n } τ n ( a 0 + 1 2 ) 1 e b 0 τ n e τ n 2 ( x ˜ n 2 + σ x ˜ n 2 ) τ n ( a 0 + 1 2 ) 1 e b 0 + 1 2 ( x ˜ n 2 + σ x ˜ n 2 ) τ n ,
where T : = diag { τ 1 , , τ N } . Therefore, we can model τ n as
q ( τ n ) Gamma a 0 + 1 2 , b 0 + 1 2 ( x ˜ n 2 + σ x ˜ n 2 ) .
The update rule for τ n can be then defined as follows.
τ ˜ n = a 0 + 1 2 b 0 + 1 2 ( x ˜ n 2 + σ x ˜ n 2 ) , n = 1 , 2 , , N
  • Update rule for the noise precision ϵ
q ( ϵ ) p ( ϵ ; θ 0 , θ 1 ) e { < log p ( y | x , ϵ ) > q x } ϵ θ 0 1 e θ 1 ϵ e { < log { ϵ M 2 e ( 1 2 ϵ y A x 2 2 ) } > q x } ϵ ( θ 0 + M 2 ) 1 e ϵ ( θ 1 + 1 2 < y A x 2 2 > q x ) ,
where
< y A x 2 2 > q x = y T y 2 < x > q x T A T y + < x T A T A x > q x = y T y 2 x ˜ T A T y + < x T A T A x > q x ,
and
< x T A T A x > q x = Tr ( < x T A T A x > q x ) = Tr ( < x x T > q x A T A ) = Tr ( x ˜ x ˜ T + Σ x ˜ ) A T A .
Therefore,
Ψ ˜ : = < y A x 2 2 > q x = y T y 2 x ˜ T A T y + Tr ( x ˜ x ˜ T + Σ x ˜ ) A T A .
Therefore, we can model ϵ ^ as
q ( ϵ ) Gamma θ 0 + M 2 , θ 1 + 1 2 Ψ ˜ .
Finally, the update rule for ϵ can be then written as
ϵ ˜ = θ 0 + M 2 θ 1 + 1 2 Ψ ˜ .
  • Update rule for the solution vector x
q x ( x ) e { < log { p ( x , y | θ ) } > q θ } e { < log { p ( x | θ ) p ( y | x , θ ) } > q θ } e { < log { p ( x | T ) } > q τ } e { < log { p ( y | x , ϵ ) } > q ϵ } e { < log { p ( x | T ) } > q τ } e { 1 2 x T T ˜ x } ,
where θ contains the information on the parameters T and ϵ , and T ˜ : = diag { τ ˜ 1 , , τ ˜ N } . To update the elements of x , we have
p ( y , x , ϵ ) ϵ M 2 e { 1 2 ϵ y A x 2 2 } e { 1 2 ϵ y A x 2 2 } .
Therefore,
< log { p ( y | x , ϵ ) } > q ϵ 1 2 < ϵ y A x 2 2 > q ϵ 1 2 < ϵ > q ϵ y A x 2 2 1 2 ϵ ˜ y A x 2 2 .
Thus, we can write q x ( x ) as
q x ( x ) e < log { p ( x | T ) } > q T e < log p ( y | x , θ ) > q θ e { 1 2 x T T ˜ x } e { 1 2 ϵ ˜ ( x T A T A x 2 x T A T y ) } e { 1 2 x T ( T ˜ + ϵ ˜ A T A ) x 2 ϵ ˜ x T A T y } .
Finally,
q x ( x ) N ( x ˜ , Σ x ˜ ) ,
where
Σ x ˜ : = ( T ˜ + ϵ ˜ A T A ) 1 and x ˜ : = ϵ ˜ Σ x ˜ A T y .
  • Stopping rule
We set the stopping rule of the algorithm based on the marginalized log-likelihood (evidence) defined as
p ( y | ϵ , T ) = p ( y , x | ϵ , T ) d x p ( y | x , ϵ , T ) p ( x | T ) d x = 1 ( 2 π ϵ 1 ) M 2 e 1 2 ϵ y A x 2 2 1 ( ( 2 π ) N | T 1 | ) 1 2 e 1 2 x T T x d x = 1 ( 2 π ) M 2 ϵ M 2 | T | 1 2 1 ( 2 π ) N 2 e 1 2 ϵ ( y T y 2 x T A T y + x T A T A x ) + x T T x d x = 1 ( 2 π ) M 2 ϵ M 2 | T | 1 2 1 ( 2 π ) N 2 e 1 2 ϵ ( y T y 2 x T A T y + x T A T A x ) + x T T x d x
= 1 ( 2 π ) M 2 ϵ M 2 | T | 1 2 e 1 2 ϵ y T y 1 ( 2 π ) N 2 e 1 2 x T ( ϵ A T A + T ) x 2 ϵ x T A T y d x = 1 ( 2 π ) M 2 ϵ M 2 | T | 1 2 e 1 2 ϵ y T y | ( T + ϵ A T A ) 1 | 1 2 1 ( 2 π ) N 2 1 | ( T + ϵ A T A ) 1 | 1 2 e 1 2 x T ( ϵ A T A + T ) x 2 ϵ x T A T y d x = 1 ( 2 π ) M 2 ϵ M 2 | T | 1 2 e 1 2 ϵ y T y | ( T + ϵ A T A ) 1 | 1 2 1 ( 2 π ) N 2 1 | ( T + ϵ A T A ) 1 | 1 2 × e 1 2 x ( T + ϵ A T A ) 1 ϵ A T y T ( T + ϵ A T A ) x ( T + ϵ A T A ) 1 ϵ A T y ϵ 2 y T A ( T + ϵ A T A ) 1 A T y d x .
Thus,
log p ( y | ϵ , T ) = M 2 log { 2 π } + M 2 log ϵ + 1 2 log | T | 1 2 ϵ y T y + 1 2 log { | ( T + ϵ A T A ) 1 | } + 1 2 ϵ 2 y T A ( T + ϵ A T A ) 1 A T y .
Notice that
1 2 ϵ y T y + 1 2 ϵ 2 y T A ( T + ϵ A T A ) 1 A T y = 1 2 y T ϵ ( I M ϵ A ( T + ϵ A T A ) 1 A T ) y
and,
1 2 log | T | + 1 2 log { | ( T + ϵ A T A ) 1 | } = 1 2 log | T | + log { | ( T + ϵ A T A ) 1 | } = 1 2 log { | T 1 ( T + ϵ A T A ) | } = 1 2 log { | I N + ϵ T 1 A T A | } = 1 2 log { | I M + ϵ A T 1 A T | } .
Thus,
L : = log p ( y | ϵ , T ) = M 2 log { 2 π } + M 2 log ϵ 1 2 log { | I M + ϵ A T 1 A T | } 1 2 ϵ y T I M ϵ A ( T + ϵ A T A ) 1 A T y .
For the comparing L [ t ] with L [ t 1 ] in the updating process, we have
L M 2 log ϵ + 1 2 log { | ( I M + ϵ A T 1 A T ) 1 | } 1 2 ϵ y T I M ϵ A ( T + ϵ A T A ) 1 A T y .
Therefore,
I M ϵ A ( T + ϵ A T ) 1 A T = ( I M + ϵ A T 1 A T ) 1 .
Thus,
L M 2 log ϵ + 1 2 log { | ( I M + + ϵ A T 1 A T ) 1 | } 1 2 ϵ y T ( I M + ϵ A T 1 A T ) 1 y 1 2 log { | ϵ 1 I M | 1 | I M + ϵ A T 1 A T | 1 } 1 2 y T ( ϵ 1 I M + A T 1 A T ) 1 y 1 2 log { | ϵ 1 I M ( I M + ϵ A T 1 A T ) 1 | } 1 2 y T ( ϵ 1 I M + A T 1 A T ) 1 y log { | ϵ 1 I M + A T 1 A T | 1 } y T ( ϵ 1 I M + A T 1 A T ) 1 y .
Therefore,
L [ t ] log | Σ 0 [ t ] | y T Σ 0 [ t ] y ,
where
Σ 0 : = ( ϵ ˜ 1 I M + A T ˜ 1 A T ) 1 .
This means that
p ( y | ϵ , T ) = 1 ( 2 π ) M 2 1 | Σ 0 1 | 1 2 e { 1 2 y T Σ 0 y }
or equivalently,
p ( y | ϵ , T ) N ( 0 , Σ 0 1 ) .
Thus, the stopping criterion can be made based on
Δ L [ t ] : = L [ t ] L [ t 1 ] = log | Σ 0 [ t ] Σ 0 [ t 1 ] | + y T ( Σ 0 [ t 1 ] Σ 0 [ t ] ) y .

References

  1. Candes, E.J.; Romberg, J.; Tao, T. Robust uncertainty principles: Exact signal reconstruction from highly incomplete frequency information. IEEE Trans. Inf. Theory 2006, 52, 489–509. [Google Scholar] [CrossRef] [Green Version]
  2. Donoho, D.L. Compressed sensing. IEEE Trans. Inf. Theory 2006, 52, 1289–1306. [Google Scholar] [CrossRef]
  3. Candes, E.J.; Wakin, M.B. An introduction to compressive sampling. IEEE Signal Process. Mag. 2008, 25, 21–30. [Google Scholar] [CrossRef]
  4. Duarte, M.; Davenport, M.; Takhar, D.; Laska, J.; Sun, T.; Kelly, K.; Baraniuk, R. Single-pixel imaging via compressive sampling. IEEE Signal Process. Mag. 2008, 25, 83–91. [Google Scholar] [CrossRef] [Green Version]
  5. Bajwa, W.; Haupt, J.; Sayeed, A.; Nowak, R. Compressed channel sensing: A new approach to estimating sparse multipath channels. Proc. IEEE 2010, 98, 1058–1076. [Google Scholar] [CrossRef]
  6. Lustig, M.; Donoho, D.; Pauly, J. Sparse MRI: The application of compressed sensing for rapid MR imaging. Magn. Reson. Med. 2007, 58, 1182–1195. [Google Scholar] [CrossRef]
  7. Kutynoik, G. Theory and applications of compressed sensing. GAMM-Mitteilungen 2013, 36, 79–101. [Google Scholar] [CrossRef] [Green Version]
  8. Chang, K.; Ding, P.; Li, B. Compressive sensing reconstruction of correlated images using joint regularization. IEEE Signal Process. Lett. 2016, 23, 449–453. [Google Scholar] [CrossRef]
  9. Wijewardhana, U.L.; Codreanu, M.; Latva-aho, M. Bayesian method for image recovery from block compressive sensing. In Proceedings of the 2016 50th Asilomar Conference on Signals, Systems and Computers, Pacific Grove, CA, USA, 6–9 November 2016; pp. 379–383. [Google Scholar]
  10. Qaisar, S.; Bilal, R.M.; Iqbal, W.; Naureen, M.; Lee, S. Compressive sensing: From theory to applications, a survey. Commun. Netw. J. 2013, 15, 443–456. [Google Scholar] [CrossRef]
  11. Mishali, M.; Eldar, Y.C. Blind multi-band signal reconstruction: Compressed sensing for analog signals. IEEE Trans. Signal Process. 2009, 57, 993–1009. [Google Scholar] [CrossRef] [Green Version]
  12. Mishali, M.; Eldar, Y.C. Xampling: Signal acquisition and processing in unions of subspaces. IEEE Trans. Signal Process. 2011, 59, 4719–4734. [Google Scholar] [CrossRef] [Green Version]
  13. Cohen, D.; Mishra, K.V.; Eldar, Y.C. Spectrum sharing Radar: Coexistence via xampling. IEEE Trans. Aerosp. Electron. Syst. 2017, 54, 1279–1296. [Google Scholar] [CrossRef] [Green Version]
  14. Aubry, A.; Carotenuto, V.; Maio, A.D.; Govoni, M.A.; Farina, A. Experimental analysis of block-sparsity-based spectrum sensing techniques for cognitive Radar. IEEE Trans. Aerosp. Electron. Syst. 2020, 57, 355–370. [Google Scholar] [CrossRef]
  15. Hwang, S.; Seo, J.; Park, J.; Kim, H.; Jeong, B.J. Compressive sensing-based Radar imaging and subcarrier allocation for joint MIMO OFDM Radar and communication system. Sensors 2021, 21, 2382. [Google Scholar] [CrossRef]
  16. Rani, M.; Dhok, S.B.; Deshmukh, R.B. A systematic review of compressive sensing: Concepts, implementations and applications. IEEE Access 2018, 6, 4875–4894. [Google Scholar] [CrossRef]
  17. Zhan, Z.; Li, Q.; Huang, J. Application of wavefield compressive sensing in surface wave tomography. Geophys. J. Int. 2018, 213, 1731–1743. [Google Scholar] [CrossRef] [Green Version]
  18. Da Poian, G.; Rozell, C.J.; Bernardini, R.; Rinaldo, R.; Clifford, G.D. Matched filtering for heart rate estimation on compressive sensing ECG measurements. IEEE Trans. Biomed. Eng. 2017, 65, 1349–1358. [Google Scholar] [CrossRef]
  19. Djelouat, H.; Zhai, X.; Disi, M.A.; Amira, A.; Bensaali, F. System-on-chip solution for patients biometric: A compressive sensing-based approach. IEEE Sens. J. 2018, 18, 9629–9639. [Google Scholar] [CrossRef] [Green Version]
  20. Zhang, P.; Wang, S.; Guo, K.; Wang, J. A secure data collection scheme based on compressive sensing in wireless sensor networks. Ad Hoc Netw. 2018, 70, 73–84. [Google Scholar] [CrossRef]
  21. Sharma, S.K.; Chatzinotas, S.; Ottersten, B. Compressive sparsity order estimation for wideband cognitive radio receiver. IEEE Trans. Signal Process. 2014, 62, 4984–4996. [Google Scholar] [CrossRef] [Green Version]
  22. Zhao, T.; Wang, Y. Statistical interpolation of spatially varying but sparsely measured 3D geo-data using compressive sensing and variational Bayesian inference. Math. Geosci. 2021, 53, 1171–1199. [Google Scholar] [CrossRef]
  23. Han, R.; Bai, L.; Zhang, W.; Liu, J.; Choi, J.; Zhang, W. Variational inference based sparse signal detection for next generation multiple access. IEEE J. Sel. Areas Commun. 2022, 40, 1114–1127. [Google Scholar] [CrossRef]
  24. Tang, V.H.; Bouzerdoum, A.; Phung, S.L. Variational Bayesian compressive multipolarization indoor Radar imaging. IEEE Trans. Geosci. Remote Sens. 2021, 59, 7459–7474. [Google Scholar] [CrossRef]
  25. Wan, Q.; Fang, J.; Huang, Y.; Duan, H.; Li, H. A Variational Bayesian inference-inspired unrolled deep network for MIMO detection. IEEE Trans. Signal Process. 2022, 70, 423–437. [Google Scholar] [CrossRef]
  26. Fang, J.; Shen, Y.; Li, H.; Wang, P. Pattern-coupled sparse Bayesian learning for recovery of block-sparse signals. IEEE Trans. Signal Process. 2015, 63, 360–372. [Google Scholar] [CrossRef] [Green Version]
  27. Shekaramiz, M.; Moon, T.K.; Gunther, J.H. Bayesian compressive sensing of sparse signals with unknown clustering patterns. Entropy 2019, 21, 247. [Google Scholar] [CrossRef] [Green Version]
  28. Wipf, D.P.; Rao, B.D. Sparse Bayesian learning for basis pursuit selection. IEEE Trans. Signal Process. 2004, 52, 2153–2164. [Google Scholar] [CrossRef]
  29. Lv, F.; Zhang, C.; Tang, Z.; Zhang, P. Block-sparse signal recovery based on adaptive matching pursuit via spike and slab prior. In Proceedings of the 2020 IEEE 11th Sensor Array and Multichannel Signal Processing Workshop (SAM), Hangzhou, China, 8–11 June 2020; pp. 1–5. [Google Scholar]
  30. Worley, B. Scalable mean-field sparse Bayesian learning. IEEE Trans. Signal Process. 2019, 67, 6314–6326. [Google Scholar] [CrossRef]
  31. Chen, P.; Zhao, J.; Bai, X. Block inverse-free sparse Bayesian learning for block sparse signal recovery. In Proceedings of the 2019 IEEE International Conference on Signal, Information and Data Processing (ICSIDP), Chongqing, China, 11–13 December 2019; pp. 1–4. [Google Scholar]
  32. Hilli, A.A.; Najafizadeh, L.; Petropulu, A. Weighted sparse Bayesian learning (WSBL) for basis selection in linear underdetermined systems. IEEE Trans. Veh. Technol. 2019, 68, 7353–7367. [Google Scholar] [CrossRef]
  33. Wang, D.; Zhang, Z. Variational Bayesian inference based robust multiple measurement sparse signal recovery. Digit. Signal Process. 2019, 89, 131–144. [Google Scholar] [CrossRef]
  34. Bayisa, F.L.; Zhou, Z.; Cronie, O.; Yu, J. Adaptive algorithm for sparse signal recovery. Digit. Signal Process. 2019, 87, 10–18. [Google Scholar] [CrossRef] [Green Version]
  35. Nayek, R.; Fuentes, R.; Worden, K.; Cross, E.J. On spike-and-slab priors for Bayesian equation discovery of nonlinear dynamical systems via sparse linear regression. Mech. Syst. Signal Process. 2021, 161, 107986. [Google Scholar] [CrossRef]
  36. Li, J.; Zhou, W.; Cheng, C. Adaptive support-driven Bayesian reweighted algorithm for sparse signal recovery. Signal Image Video Process. 2021, 15, 1295–1302. [Google Scholar] [CrossRef]
  37. Zong-Long, B.; Li-Ming, S.; Jin-Wei, S. Sparse Bayesian learning using adaptive LASSO priors. Acta Autom. Sin. 2021, 45, 1–16. [Google Scholar]
  38. Mallat, S.; Zhang, Z. Matching pursuits with time-frequency dictionaries. IEEE Trans. Signal Process. 1993, 41, 3397–3415. [Google Scholar] [CrossRef] [Green Version]
  39. Blumensath, T.; Davies, M.E. Iterative hard thresholding for compressive sensing. Appl. Comput. Harmon. Anal. 2009, 27, 265–274. [Google Scholar] [CrossRef] [Green Version]
  40. Stanković, L.; Daković, M.; Vujović, S. Adaptive variable step algorithm for missing samples recovery in sparse signals. IET Signal Process. 2014, 8, 246–256. [Google Scholar] [CrossRef] [Green Version]
  41. Chen, S.; Donoho, D. Basis pursuit. In Proceedings of the 1994 28th Asilomar Conference on Signals, Systems and Computers, Pacific Grove, CA, USA, 31 October–2 November 1994; pp. 41–44. [Google Scholar]
  42. Zhou, W.; Zhang, H.T.; Wang, J. An efficient sparse Bayesian learning algorithm based on Gaussian-scale mixtures. IEEE Trans. Neural Netw. Learn. Syst. 2021, 33, 3065–3078. [Google Scholar] [CrossRef]
  43. Sant, A.; Leinonen, M.; Rao, B.D. General total variation regularized sparse Bayesian learning for robust block-sparse signal recovery. In Proceedings of the ICASSP 2021—2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Toronto, ON, Canada, 6–11 June 2021; pp. 5604–5608. [Google Scholar]
  44. Liu, J.; Wu, Q.; Amin, M.G. Multi-Task Bayesian compressive sensing exploiting signal structures. Signal Process. 2021, 178, 107804. [Google Scholar] [CrossRef]
  45. He, L.; Chen, H.; Carin, L. Tree-structured compressive sensing with variational Bayesian analysis. IEEE Signal Process. Lett. 2010, 17, 233–236. [Google Scholar]
  46. Ji, S.; Xue, Y.; Carin, L. Bayesian compressive sensing. IEEE Trans. Signal Process. 2008, 56, 2346–2356. [Google Scholar] [CrossRef]
  47. Shekaramiz, M.; Moon, T.K.; Gunther, J.H. Hierarchical Bayesian approach for jointly-sparse solution of multiple-measurement vectors. In Proceedings of the 2014 48th Asilomar Conference on Signals, Systems and Computers, Pacific Grove, CA, USA, 2–5 November 2014; pp. 1962–1966. [Google Scholar]
  48. Oikonomou, V.P.; Nikolopoulos, S.; Kompatsiaris, I. A novel compressive sensing scheme under the variational Bayesian framework. In Proceedings of the 2019 27th European Signal Processing Conference (EUSIPCO), A Coruna, Spain, 2–6 September 2019; pp. 1–5. [Google Scholar]
  49. Wang, L.; Zhao, L.; Yu, L.; Wang, J.; Bi, G. Structured Bayesian learning for recovery of clustered sparse signal. Signal Process. 2020, 166, 107255. [Google Scholar] [CrossRef]
  50. Yu, L.; Wei, C.; Jia, J.; Sun, H. Compressive sensing for cluster structured sparse signals: Variational Bayes approach. IET Signal Process. 2016, 10, 770–779. [Google Scholar] [CrossRef] [Green Version]
  51. Babacan, S.D.; Nakajima, S.; Do, M.N. Bayesian group-sparse modeling and variational inference. IEEE Trans. Signal Process. 2014, 62, 2906–2921. [Google Scholar] [CrossRef]
  52. Yu, L.; Sun, H.; Barbot, J.P.; Zheng, G. Bayesian compressive sensing for cluster structured sparse signals. Signal Process. 2012, 92, 259–269. [Google Scholar] [CrossRef] [Green Version]
  53. Anderson, M.R.; Winther, O.; Hansen, L.K. Bayesian inference for structured spike and slab priors. In Proceedings of the Advances in Neural Information Processing Systems 27 (NIPS 2014), Montreal, QC, Canada, 8–13 December 2014; pp. 1745–1753. [Google Scholar]
  54. Babacan, S.; Molina, R.; Katsaggelos, A. Bayesian compressive sensing using Laplace priors. IEEE Trans. Image Process. 2010, 19, 53–63. [Google Scholar] [CrossRef]
  55. Hernandez-Lobato, D.; Hernandez-Lobato, J.M.; Dupont, P. Generalized spike-and-slab priors for Bayesian group feature selection using expectation propagation. J. Mach. Learn. Res. 2013, 14, 1891–1945. [Google Scholar]
  56. Ji, S.; Dunson, D.; Carin, L. Multitask compressive sensing. IEEE Trans. Signal Process. 2009, 57, 92–106. [Google Scholar] [CrossRef]
  57. Shekaramiz, M.; Moon, T.K.; Gunther, J.H. Sparse Bayesian learning using variational Bayes inference based on a greedy criterion. In Proceedings of the 2017 51st Asilomar Conference on Signals, Systems, and Computers, Pacific Grove, CA, USA, 29 October–1 November 2017; pp. 858–862. [Google Scholar]
  58. Wu, Q.; Fang, S. Structured Bayesian compressive sensing with spatial location dependence via variational Bayesian inference. Digit. Signal Process. 2017, 71, 95–107. [Google Scholar] [CrossRef]
  59. Wipf, D.P.; Rao, B.D. An empirical Bayesian strategy for solving the simultaneous sparse approximation problem. IEEE Trans. Signal Process. 2007, 55, 3704–3716. [Google Scholar] [CrossRef]
  60. Tibshirani, R.; Saunders, M.; Rosset, S.; Zhu, J.; Knight, K. Sparsity and smoothness via the fused LASSO. J. R. Stat. Soc. Ser. B 2005, 67, 91–108. [Google Scholar] [CrossRef] [Green Version]
  61. Blumensath, T.; Davies, M.E. Normalized iterative hard thresholding: Guaranteed stability and performance. IEEE J. Sel. Top. Signal Process. 2010, 4, 298–309. [Google Scholar] [CrossRef] [Green Version]
  62. Qin, L.; Tan, J.; Wang, Z.; Wang, G.; Guo, X. Exploiting the tree-structured compressive sensing of Wavelet coefficients via block sparse Bayesian learning. Electron. Lett. 2018, 54, 975–976. [Google Scholar] [CrossRef]
  63. Ambat, S.K.; Chatterjee, S.; Hari, K.V. Fusion of greedy pursuits for compressed sensing signal reconstruction. In Proceedings of the 2012 Proceedings of the 20th European Signal Processing Conference (EUSIPCO), Bucharest, Romania, 27–31 August 2012; pp. 1434–1438. [Google Scholar]
  64. Cao, Z.; Dai, J.; Xu, W.; Chang, C. Fast variational Bayesian inference for temporally correlated sparse signal recovery. IEEE Sigal Process. Lett. 2021, 28, 214–218. [Google Scholar] [CrossRef]
  65. Gelman, A.; Rubin, D.B. Inference from iterative simulation using multiple sequences. Stat. Sci. 1992, 7, 457–511. [Google Scholar] [CrossRef]
  66. Beal, M. Variational Algorithms for Approximate Bayesian Inference. Ph.D. Dissertation, University College London, London, UK, 2003. [Google Scholar]
  67. Tzikas, D.G.; Likas, A.C.; Galatsanos, N.P. The variational approximation for Bayesian inference. IEEE Signal Process. Mag. 2008, 25, 131–142. [Google Scholar] [CrossRef]
  68. Shekaramiz, M.; Moon, T.K. Compressive sensing via variational Bayesian inference. In Proceedings of the 2020 Intermountain Engineering, Technology and Computing (IETC), Orem, UT, USA, 2–3 October 2020; pp. 1–6. [Google Scholar]
  69. Shekaramiz, M.; Moon, T.K. Sparse Bayesian learning via variational Bayes fused With orthogonal matching pursuit. In Proceedings of the 2022 Intermountain Engineering, Technology and Computing (IETC), Orem, UT, USA, 13–14 May 2022; pp. 1–5. [Google Scholar]
  70. You, C.; Ormerod, J.T.; Mueller, S. On variational Bayes estimation and variational information criteria for linear regression models. Aust. N. Z. J. Stat. 2014, 56, 73–87. [Google Scholar] [CrossRef]
  71. Tran, M.N.; Nguyen, T.N.; Dao, V.H. A practical tutorial on variational Bayes. arXiv 2021, arXiv:2103.01327. [Google Scholar]
  72. Fox, C.; Roberts, S. A tutorial on variational Bayesian inference. Artif. Intell. Rev. 2011, 38, 85–95. [Google Scholar] [CrossRef]
  73. Manipur, I.; Manzo, M.; Granata, I.; Giordano, M.; Maddalena, L.; Guarracino, M.R. Netpro2vec: A graph embedding framework for biomedical applications. IEEE/ACM Trans. Comput. Biol. Bioinform. 2021, 19, 729–740. [Google Scholar] [CrossRef]
Figure 1. Graphical Bayesian representation of the BGiG model.
Figure 1. Graphical Bayesian representation of the BGiG model.
Entropy 25 00511 g001
Figure 2. Flowchart of SBL(BGiG) algorithm.
Figure 2. Flowchart of SBL(BGiG) algorithm.
Entropy 25 00511 g002
Figure 3. Case 1: ( α 0 , β 0 ) = ( 0.01 , 0.99 ) . From top to bottom, the rows show the results of SBL(BGiG) for the sampling ratio λ = 0.80 , 0.60 , 0.40 , respectively.
Figure 3. Case 1: ( α 0 , β 0 ) = ( 0.01 , 0.99 ) . From top to bottom, the rows show the results of SBL(BGiG) for the sampling ratio λ = 0.80 , 0.60 , 0.40 , respectively.
Entropy 25 00511 g003
Figure 4. Case 2: ( α 0 , β 0 ) = ( 0.1 , 0.9 ) . From top to bottom, the rows show the results of SBL(BGiG) for the sampling ratio λ = 0.80 , 0.60 , 0.40 , respectively.
Figure 4. Case 2: ( α 0 , β 0 ) = ( 0.1 , 0.9 ) . From top to bottom, the rows show the results of SBL(BGiG) for the sampling ratio λ = 0.80 , 0.60 , 0.40 , respectively.
Entropy 25 00511 g004
Figure 5. Case 3: ( α 0 , β 0 ) = ( 1.4 , 2 ) . From top to bottom, the rows show the results of SBL(BGiG) for the sampling ratio λ = 0.80 , 0.60 , 0.40 , respectively.
Figure 5. Case 3: ( α 0 , β 0 ) = ( 1.4 , 2 ) . From top to bottom, the rows show the results of SBL(BGiG) for the sampling ratio λ = 0.80 , 0.60 , 0.40 , respectively.
Entropy 25 00511 g005
Figure 6. Case 1: Performance evaluation of SBL(BGiG).
Figure 6. Case 1: Performance evaluation of SBL(BGiG).
Entropy 25 00511 g006
Figure 7. Case 2: Performance evaluation of SBL(BGiG).
Figure 7. Case 2: Performance evaluation of SBL(BGiG).
Entropy 25 00511 g007
Figure 8. Case 3: Performance evaluation of SBL(BGiG).
Figure 8. Case 3: Performance evaluation of SBL(BGiG).
Entropy 25 00511 g008
Figure 9. Graphical Bayesian representation of the GiG model.
Figure 9. Graphical Bayesian representation of the GiG model.
Entropy 25 00511 g009
Figure 10. Flowchart of SBL(GiG) algorithm.
Figure 10. Flowchart of SBL(GiG) algorithm.
Entropy 25 00511 g010
Figure 11. From left to right, we show the results for sampling ratios of λ = 0.8, 0.6 and 0.40, respectively. The first row shows the comparison of y with y ^ = A x s ˜ , the second row shows the true solution x s and the estimated solution x ˜ s , and the third row demonstrates the estimated precisions on the solution components.
Figure 11. From left to right, we show the results for sampling ratios of λ = 0.8, 0.6 and 0.40, respectively. The first row shows the comparison of y with y ^ = A x s ˜ , the second row shows the true solution x s and the estimated solution x ˜ s , and the third row demonstrates the estimated precisions on the solution components.
Entropy 25 00511 g011
Figure 12. The behavior of negative marginalized log-likelihood and the precision on the noise using SBL(GiG) for the sampling ratios of 0.4, 0.6 and 0.80.
Figure 12. The behavior of negative marginalized log-likelihood and the precision on the noise using SBL(GiG) for the sampling ratios of 0.4, 0.6 and 0.80.
Entropy 25 00511 g012
Figure 13. Performance evaluation of SBL(BGiG) using grid and random Sobol search.
Figure 13. Performance evaluation of SBL(BGiG) using grid and random Sobol search.
Entropy 25 00511 g013aEntropy 25 00511 g013b
Figure 14. (a) Overall performance (b) Top 10 ( α 0 , β 0 ) with lowest NMSE (c) Top 10 ( α 0 , β 0 ) with highest P D P F A .
Figure 14. (a) Overall performance (b) Top 10 ( α 0 , β 0 ) with lowest NMSE (c) Top 10 ( α 0 , β 0 ) with highest P D P F A .
Entropy 25 00511 g014
Figure 15. (a) Top 10 ( α 0 , β 0 ) with lowest NMSE vs. sampling ratio (b) Top 10 ( α 0 , β 0 ) with highest P D P F A vs. sampling ratio.
Figure 15. (a) Top 10 ( α 0 , β 0 ) with lowest NMSE vs. sampling ratio (b) Top 10 ( α 0 , β 0 ) with highest P D P F A vs. sampling ratio.
Entropy 25 00511 g015
Figure 16. Performance of SBL(GiG). (a) NMSE of SBL(GiG) vs. threshold. (b) Performance of SBL(GiG) before and after postprocessing.
Figure 16. Performance of SBL(GiG). (a) NMSE of SBL(GiG) vs. threshold. (b) Performance of SBL(GiG) before and after postprocessing.
Entropy 25 00511 g016
Figure 17. Performance of SBL(BGiG) and SBL(GiG) after preprocessing and postprocessing, respectively.
Figure 17. Performance of SBL(BGiG) and SBL(GiG) after preprocessing and postprocessing, respectively.
Entropy 25 00511 g017
Table 1. Performance results of SBL(BGiG) for Cases 1–3.
Table 1. Performance results of SBL(BGiG) for Cases 1–3.
Case 1: ( α 0 = 0.01 , β 0 = 0.99 ) Case 2: ( α 0 = 0.1 , β 0 = 0.9 ) Case 3: ( α 0 = 1.4 , β 0 = 2.0 )
λ P D P FA NMSE (dB) λ P D P FA NMSE (dB) λ P D P FA NMSE (dB)
0.80.200−2.3670.80.240−3.1090.80.720−16.264
0.60.080−1.3260.60.160−2.1970.610−5.226
0.40.080−1.1810.40.080−1.1810.411−0.088
Table 2. Settings for preprocessing analysis and simulations on SBL(BGiG).
Table 2. Settings for preprocessing analysis and simulations on SBL(BGiG).
α 0 β 0 a 0 b 0 θ 0 θ 1 Sparsity γ N
[ 0.1 , 2 ] [ 0.1 , 2 ] 10 3 10 3 10 6 10 6 25(5)100
s τ x x s M ϵ e A y
(5)(6)(6) x s = x s 5 : N 316(7) [ A ] m n N ( 0 , 1 ) A x s + e
Table 3. Settings for preprocessing analysis and simulations on SBL(GiG).
Table 3. Settings for preprocessing analysis and simulations on SBL(GiG).
a 0 b 0 θ 0 θ 1 SparsityN
10 3 10 3 10 6 10 6 25100
τ n x s M e A y
(22)(25) 5 : N (23) [ A ] m n N ( 0 , 1 ) A x s + e
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Shekaramiz, M.; Moon, T.K. Compressive Sensing via Variational Bayesian Inference under Two Widely Used Priors: Modeling, Comparison and Discussion. Entropy 2023, 25, 511. https://doi.org/10.3390/e25030511

AMA Style

Shekaramiz M, Moon TK. Compressive Sensing via Variational Bayesian Inference under Two Widely Used Priors: Modeling, Comparison and Discussion. Entropy. 2023; 25(3):511. https://doi.org/10.3390/e25030511

Chicago/Turabian Style

Shekaramiz, Mohammad, and Todd K. Moon. 2023. "Compressive Sensing via Variational Bayesian Inference under Two Widely Used Priors: Modeling, Comparison and Discussion" Entropy 25, no. 3: 511. https://doi.org/10.3390/e25030511

APA Style

Shekaramiz, M., & Moon, T. K. (2023). Compressive Sensing via Variational Bayesian Inference under Two Widely Used Priors: Modeling, Comparison and Discussion. Entropy, 25(3), 511. https://doi.org/10.3390/e25030511

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop