Next Article in Journal
Signal Processing for Transient Flow Rate Determination: An Analytical Soft Sensor Using Two Pressure Signals
Previous Article in Journal
A 2.4 GHz IEEE 802.15.4 Multi-Hop Network for Mountainous Forest and Watercourse Environments: Sensor Node Deployment and Performance Evaluation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

RIP Sensing Matrices Construction for Sparsifying Dictionaries with Application to MRI Imaging

1
Institute of Information Science, Academia Sinica, Taipei 11529, Taiwan
2
Department of Mathematics, National University of Singapore, Singapore 119077, Singapore
*
Author to whom correspondence should be addressed.
Signals 2024, 5(4), 794-811; https://doi.org/10.3390/signals5040044
Submission received: 16 September 2024 / Revised: 30 October 2024 / Accepted: 5 November 2024 / Published: 2 December 2024

Abstract

:
Practical applications of compressed sensing often restrict the choice of its two main ingredients. They may (i) prescribe the use of particular redundant dictionaries for certain classes of signals to become sparsely represented or (ii) dictate specific measurement mechanisms which exploit certain physical principles. On the problem of RIP measurement matrix design in compressed sensing with redundant dictionaries, we give a simple construction to derive sensing matrices whose compositions with a prescribed dictionary have with high probability the RIP in the k log ( n / k ) regime. Our construction thus provides recovery guarantees usually only attainable for sensing matrices from random ensembles with sparsifying orthonormal bases. Moreover, we use the dictionary factorization idea that our construction rests on in the application of magnetic resonance imaging, in which also the sensing matrix is prescribed by quantum mechanical principles. We propose a recovery algorithm based on transforming the acquired measurements such that the compressed sensing theory for RIP embeddings can be utilized to recover wavelet coefficients of the target image, and show its performance on examples from the fastMRI dataset.

1. Introduction

Compressed sensing provides a framework under which sparse or compressible signals can be stably reconstructed from far fewer linear measurements than their ambient dimension [1,2]. The number of required measurements depends on the signal complexity in terms of sparsity, and properties of the sensing matrix with respect to sparse vectors, such as the restricted isometry property (RIP) [3]. A sensing matrix S, which embeds high-dimensional signals into a lower-dimensional measurement space, is said to have the RIP of order k if there exists a constant δ k [ 0 , 1 ) , such that for any k-sparse vector x R n :
( 1 δ k ) x 2 2 S x 2 2 ( 1 + δ k ) x 2 2 .
Improving upon [4], Candès [5] showed that k-sparse and compressible vectors can be stably recovered from observations y = S x + η with measurement error η bounded by ϵ , given that S has the RIP with δ 2 k < 2 1 (which has been improved to 0.4652 [6]) via the sparsity promoting convex program:
minimize x x 1 subject to y S x 2 ϵ .
For certain λ > 0 , (2) can be equivalently reformulated to the problem of minimizing over x the unconstrained objective 1 2 y S x 2 2 + λ x 1 consisting of a data-fidelity and regularization term, which can be solved using iterative methods [7].
While it is NP-hard to verify the RIP for a given matrix, random matrices from sub-Gaussian or Bernoulli ensembles do possess the RIP with high probability—also when composed with orthonormal bases [8]. In practice, however, many classes of signals are sparse only with respect to a redundant dictionary or tight frame that is non-orthonormal (e.g., Gabor, curvelet, wavelet, or data-driven learned dictionaries). In this case, the sensing matrix S in (1) has to be replaced by the composition S D , which no longer possesses the RIP. This prevents the RIP as recovery guaranteeing tool in many CS applications.
Contributions: We first present a method to derive a sensing matrix S R m × l for a given sparsifying dictionary D R l × n , such that with high probability, S D has the RIP for m on the order of k log ( n / k ) , where k is the sparsity level of a coefficient vector. We use a sufficient condition for the RIP of random matrices that satisfy a concentration of measure inequality [8]. Starting from any random matrix A of the dictionary dimension for which a random row selection E A R m × n satisfies this concentration inequality, a tailored sensing matrix can be derived from any factorization D = G A H , in which G R l × l is invertible and H R n × n is orthonormal: for the sensing matrix S : = E G 1 , the composition S D has then the desired RIP with high probability. We show in Section 3 that the required factorization exists whenever D and A have equal rank and detail constructions. A particular implication is that one can, thus, obtain the same RIP-based recovery guarantees for sensing matrix compositions with general over-complete dictionaries as for Gaussian or Bernoulli ensembles with orthonormal bases.
We utilize dictionary factorization in the context of CS MRI applications to showcase the viability of our method for a significant purpose. The primary aim of accelerating MRI scans is to lessen the time patients spend in the machines, thereby reducing the discomfort that can result from remaining motionless. The speed at which MRI images are accelerated is affected by the number of rows in an over-complete dictionary. As a result, recovery algorithms aim to use fewer rows while ensuring high-quality image reconstruction. In CS MRI, sensing mechanisms exploit quantum mechanical principles and existing hardware constraints restrict sensing matrix design beyond non-uniformly subsampling the measured (complex-valued) Fourier spatial frequency coefficients of the target image x ˜ = D x , which we suppose is synthesized from (real-valued) sparse coefficients x with respect to a dictionary D. The sensing matrix S = R F is thus modeled as the product of a discrete Fourier transform F and a row subsampling R . The CS recovery of x can then be formulated as the 1 -synthesis problem:
minimize x x 1 subject to y R F D x 2 ϵ
in which R F possesses the RIP  [9,10,11], while, in general, R F D does not, such that the RIP recovery guarantees do not directly apply. Using the dictionary factorization idea, we attempt to neutralize the effect of F in order to obtain an RIP synthesis problem. We do so by choosing (real) factors G such that (both real and imaginary parts of) R F G optimally match R , so that the factorizations G A H that optimally approximate D allow, by virtue of R F D R F G A H R A H , to replace the constraint in (3) by y R A H x 2 ϵ . Both the real and the imaginary parts of R A H are incoherent sampling matrices for sparse signals that possess the RIP. We provide a review of CS application in Section 2 and detail the application of CS factorization method to MRI in Section 4, and report in Section 5 numerical experiments comparing our approach to total-variation based CS MRI [12].
Related work: If the sparsifying dictionary in the general 1 -synthesis approach is a tight frame, i.e., x = D x ˜ , recovery guarantees can be derived by considering the 1 -analysis approach of minimizing over x ˜ the objective D x ˜ 1 subject to the constraint y S x ˜ 2 ϵ . Recovery guarantees for the 1 -analysis approach are connected to the notion of D-RIP [13], which requires the sensing matrix to satisfy the RIP inequality for all images of k-sparse vectors under a tight frame D. Any RIP matrix satisfies the D-RIP when multiplied by a random sign matrix [14]. Unless D is orthonormal, the geometric structures, properties and empirical performances of the 1 -analysis and 1 -synthesis approaches in general differ [15].
A weaker condition than the RIP that can facilitate recovery guarantees for S D with general dictionary D is the mutual incoherence between the sensing matrix and the dictionary [2,16]. However, sparsity ranges for signals to be recovered, as well as necessary number of measurements are more restrictive than for RIP guarantees, and overcomplete dictionaries with highly coherent columns in general lead to large coherence of their product with sensing matrices. Finally, the nullspace property of a sensing matrix gives a necessary and sufficient condition for stable recovery of sparse signals using the convex optimization [17].

2. Compressed Sensing Applications

Compressed Sensing (CS) has found numerous applications across various fields. One significant area is addressing the future data storage challenges [18]. For example, advancements in communication and signal processing are producing vast amounts of data; 5G networks can reach peak data rates of about 300 Mb per second, while an image of a black hole needs several petabytes for storage. In medical signal processing, the volume of required data is also substantial. CS can enhance wireless communication capabilities in cognitive radio networks and enable spatial compression and projection in wireless sensor networks [19]. Additionally, modern imaging techniques rely on arrays of millions of detector pixels. In contrast, CS is foundational to single-pixel cameras, which utilize mask patterns to filter scenes and capture transmitted intensity measurements with a one-pixel detector. This technology is well-suited for imaging beyond the visible spectrum. It allows for precise timing or depth resolution, making it ideal for applications such as infrared imaging and 3D situational awareness in autonomous vehicles [20].
An important CS application, in which the sensing mechanisms is dictated by hardware constraints, is magnetic resonance imaging (MRI). In MRI a scanner collects Fourier domain measurements of a target medical image. Accelerating the speed of MRI data acquisition by reducing the number of required measurements remains of great interest to the medical community. Inference of the true underlying spatial image can be achieved via several CS strategies when non-uniformly under-sampling below the requirements of the Shannon–Nyquist theory within the maximum frequency spectrum, by incorporating the a priori knowledge about sparsity of medical images in a dictionary transform domain or of its spatial gradients [21]. While deep learning approaches have been a recent center of attention to recover data under-sampled in this way [22], recovery guarantees and interpretability set the CS approach apart from such machine learning techniques [23]. Though state-of-the-art machine learning models for MRI have been validated for clinical interchangeability in four-fold data acquisition acceleration [24], they rely on empirical studies and present challenges concerning, for instance, reconstruction hallucinations that can be problematic for interpreting radiologists [25], unknown training data biases [26], or robustness to distribution shifts between training and application data (e.g., scanning technology [27], target anatomy [28], or acceleration factors [29]). Significant issues concerning the stability of distribution shifts are highlighted in [30]. The authors stress the need to apply a method across different datasets featuring anatomical regions and images collected through different scanning protocols. This approach is crucial for showcasing the method’s versatility and evaluating the model’s resilience to distribution shifts. In contrast, the sparsity-driven CS approach can be fine-tuned to perform close to deep learning methods for MRI acceleration, via unrolling algorithms which consist of a small fraction of the parameters employed by deep learning approaches [23].

3. Sensing Matrix Construction

Signal complexity, in terms of sparsity, determines the amount of possible under-sampling in CS. Constructions utilizing randomness can produce RIP matrices for which the number of required measurements m scales linearly with the sparsity level k of the vector to be recovered. Such matrices can be derived from distributions for which the following concentration of measure inequality, resembling (1), holds, such as for subgaussian or Bernoulli ensembles.
Theorem 1
([8]). Let 0 < δ k < 1 and A R m × n be an iid random matrix. If E A x 2 2 = x 2 2 for all x R n , and for any ϵ ( 0 , 1 ) the concentration inequality P ( A x 2 2 x 2 2 | ϵ x 2 2 ) 2 e l c ( ϵ ) holds for some c ( ϵ ) > 0 and all x R n , then there exists c 1 , c 2 > 0 (depending only on δ k ) such that, whenever k c 1 m log ( n / k ) , the RIP (1) holds for A with probability at least 1 2 e c 2 m .
Random matrices satisfying the concentration inequality are universal with respect to orthonormal bases [8], i.e., the same RIP conclusions hold for their products with unitary matrices. We use this universality to derive sensing matrices for any sparsity-inducing dictionary by random row selection of an invertible transform adapted to the dictionary.
Theorem 2.
Let D R l × n ( l n ) be a dictionary, A E l × n and E E m × l matrices such that E A satisfies the assumptions of Theorem 1. Suppose the dictionary allows a factorization D = G A H for some invertible G and orthonormal H. Then S : = E G 1 R m × l is a sensing matrix for D such that, with probability as in Theorem 1, S D has the RIP whenever m k log ( n / k ) .
In this construction S D equals E A H , which implies the claim due to the above mentioned universality. The existence of a factorization D = G A H as assumed in Theorem 2 requires A and D to have equal rank. We next show that the latter also suffices. Given a full-rank sparsifying dictionary, the sensing matrix construction of Theorem 2 can therefore (with probability one) be carried out, starting from a Gaussian matrix A of the same dimensions as the dictionary, and a random row selection E .
Proposition 1.
Let A , D R l × n with l n . Then, the following statements are equivalent:
(i) 
There exists an invertible G R l × l and orthonormal H R n × n such that G D = A H ;
(ii) 
There exists an invertible G R l × l with G D D G = A A ;
(iii) 
A and D have equal rank.
Proof. 
We show that (iii) implies (ii), and in turn (i). Assuming (iii), A A and D D have equal rank. Their respective spectral decompositions Q A Σ A Q A and Q D Σ D Q D can, thus, be assumed to have zeros in the same positions of the diagonal matrices Σ A , Σ D . Replacing those zero diagonal entries by ones to result in Σ A , Σ D , the matrix G : = Q A ( Σ A Σ D 1 ) 1 / 2 Q D is invertible and:
G D D G = G Q D Σ D Q D G = Q A ( Σ A Σ D 1 ) 1 / 2 Σ D ( Σ A Σ D 1 ) 1 / 2 Q A = A A .
Defining H : = A + G D + N A N D , with A + denoting the pseudo-inverse of A, and N A , respectively N D , having pairwise orthonormal columns spanning the nullspaces of A, respectively D, implies:
A H = A A + G D + A N A N D = A A + G D = G D
since A A + is the identity on the range of A, which contains the range of G. Moreover:
H H = A + G D D G ( A + ) + N A N A = A + A A ( A + ) + N A N A = A + A + N A N A ,
which is the identity. Finally, note that (i) implies (iii).    □
An alternative factorization can be derived from orthonormal bases for the ranges and nullspaces of prescribed equal rank A , D R l × n as follows. Let [ U A N A ] , [ U D N D ] be orthonormal matrices, where the columns of U A , U D span the ranges, and the columns of N A , N D the nullspaces, of A , D . Extend D U D , A U A to invertible square matrices D U D ˜ , A U A ˜ , appending columns. Then:
G : = D U D ˜ A U A ˜ 1 and H : = U A U D + N A N D
are invertible, respectively, orthonormal, and G A H D = ( G A U A D U D ) U D is zero since G A U A = D U D .
We next detail a factorization for the practically important case of a prescribed tight frame dictionary D. We may then assume D D to be the identity.
Theorem 3.
Suppose A , D R l × n have full rank l n , and D is a tight frame. Then D = G A H with invertible, respectively orthonormal:
G : = O ( A A ) 1 / 2 and H : = A G D + N A N D ,
where O R l × l is any orthonormal matrix, and the columns of N A , N D R n × n l are any orthonormal bases of the nullspaces of A, respectively D. In particular, if A is also a tight frame, then G is orthonormal.
Proof. 
Since G A A G is the identity and A N A the zero operator, a direct calculation shows G A H = D . While G is by definition invertible, D D being the identity implies that H H = A ( A A ) 1 A + N A N A is the identity.    □

4. Application to Accelerated MRI Imaging

The goal of fast MRI is the recovery of an image Z R n 1 × n 2 from under-sampled Fourier (k-space) measurements. The discrete Fourier transform of Z is the complex matrix:
Z ^ : = F 1 Z F 2 C n 1 × n 2 ,
where F 1 C n 1 × n 1 , F 2 C n 2 × n 2 are (symmetric and unitary) discrete Fourier transform matrices. Data acquired via accelerated MRI can be modeled as:
Y : = R Z ^ C n 1 × n 2 ,
where R is a { 0 , 1 } -entry n 1 × n 1 diagonal matrix that selects rows of Y corresponding to some under-sampling scheme [31]. The image can be assumed to have a sparse representation:
Z = D 1 X D 2
in the transform domain of wavelet dictionaries D i R n i × N i ( N i n i , i = 1 , 2 ) that implement a separable bivariate wavelet transform. The coefficient matrix X R N 1 × N 2 then consists of low-pass channel approximation coefficients, and sparse high-pass channel coefficients that capture edges and singular points in the target image. In this formulation, the CS-MRI problem is then the inverse problem of approximating sparse coefficients X that solve:
Y = R F 1 D 1 X D 2 F 2 ,
that is, to reconstruct Z (via (6)) from the measurements Y which are in general corrupted by noise.
Being tied to the measurements (7) by physical principles, we first transform the acquired data. Starting from a sensing factorization D 2 = G 2 A H 2 according to Proposition 1, we consider the transformed observation:
Y ˜ : = Y F 2 ( G 2 ) 1 = R F 1 D 1 X ( A 2 H 2 ) .
We next construct for D 1 two (real) factorizations that are adapted to the real and imaginary parts of F 1 . First, we let:
G R : = argmin G R n 1 × n 1 invertible R ( Re ( F 1 ) G I ) F 2
and, second, we let:
H R : = argmin H R n 2 × n 2 unitary D 1 G R A H F 2 .
Analogously, we define G I and H I via solving (8) for the imaginary part of F 1 , before solving (9) for G I . We then let H : = H R + i H I . For details on numerical solutions of (8) and (9) via the augmented Lagrangian approach we refer to the Appendix A.
Our motivation for desiring in (8) small differences between R and the real (respectively imaginary) parts of R F 1 G R (respectively R F 1 G I ) is that then (9) implies that the following difference is small:
Φ ( X ) : = 1 2 Y ˜ R A H X ( A H 2 ) F 2 ,
such that we can attempt to approximate sparse coefficients X by solving the 1 -regularized problem:
minimize X R N 1 × N 2 Φ ( X ) + γ X H 1 ,
where X H denotes the highpass coefficients of X and γ > 0 .

5. Experimental Results

5.1. Sensing Matrix of Sparsifying Dictionaries

Experiments were conducted to assess the sensing matrix derived based on the dictionary factorizations proposed in Section 3 for CDF 9/7 wavelets and dictionaries derived via K-SVD, for which synthetically generated univariate sparse signals are then shown to be recoverable with the guarantees of Theorem 2.
The entries of A R l × n are i.i.d. Bernoulli random variables (where ± 1 n with equal probability) and Gaussian random numbers (where N ( 0 , n 1 ) ). D and A have equal ranks. Experiments were performed on sparsifying dictionaries of R 128 × 1024 , including D w a v e l e t and D K S V D , which were derived using the K-SVD algorithm, based on the set of training vectors obtained from the 256 × 256 gray-scale test image “Boat” (parameters of the K-SVD process were set at K = 1024 , L = 64 , n u m I t e r a t i o n = 50 , I n i t i a l i z a t i o n M e t h o d = G i v e n M a t r i x , and e r r o r G o a l = 10 8 ). The image was divided into overlapping patches of 16 × 16 pixels with stride 2 (in each dimension) and overlaps of 4 pixels (in each dimension), which resulted in 3721 training patches. As the stride is 2 in each dimension, each patch formed a vector of 128 pixels. After the mean of each patch was normalized to zero, the resulting patches were transformed into vectors (via the vec operation) to form the set of training vectors. The dictionary D w a v e l e t was derived from the CDF 9-7 wavelet. The first 640 columns (i.e., each level has 128 columns and there are 5 levels) in D w a v e l e t were obtained from the first 5 levels (the support of a wavelet at level 6 is larger than the size of an image patch), whereas the remainder were generated randomly. The column norms of the sparsifying dictionaries were normalized. Figure 1 and Figure 2 present the sparsifying dictionaries and corresponding matrices G, derived for various A in accordance with (4). The subgraph associated with G in Figure 2 corresponds to the sensing matrix of D. It is worth noting that the lack of structure in G does not contradict some research in compressed sensing literature that focuses on designing structured sensing matrices. The structured sensing matrix approach aims to efficiently manage an extensive system’s computational resources, such as storage and computational efficiency. In contrast, our research prioritizes creating a sensing matrix that complements a given dictionary, with fewer rows than columns, to maintain the RIP. Exploring the possibility of constructing a structured sensing matrix while retaining the RIP for a given dictionary is an interesting question for future investigation.
Under Theorem 2, the following two optimizations derive the same solution, if E G 1 D = A H :
min x z E G 1 D x 2 2 x 0 k ,
and
min x z E A x 2 2 x 0 k .
Note that z in (11) and (12) are low-dimensional observations of x obtained from E A x and E in (11) and (12) are the same. To verify that (11) yields performance similar (in terms of probability) to the the sparse recovery obtained using (12), experiments were performed on the sparsifying dictionaries ( D w a v e l e t and D K S V D ) and using the compressive sampling matched pursuit (CoSaMP) algorithm [32] for the recovery of sparse vectors. Figure 3 and Figure 4 illustrate CS recovery performance using plots indicating the probability of successfully recovering a sparse vector versus the CS ratio. The figures compare the performance of our approach (11) versus the benchmark (12) using Gaussian and Bernoulli sensing matrices at various levels of sparsity. The horizontal and vertical axes, respectively, indicate the CS ratio (i.e., m n × 100 % , where m is the number of rows of E and n = 1024 ) and the probability of successfully recovering a sparse vector. We claim that the true sparse vector x can be recovered as long as estimate x ^ satisfies x ^ x 1 < 1024 × 10 2 .

5.2. MRI Image Recovery

Significant research effort based on the sparse representation has been directed towards finding ways to accelerate MR imaging. At the heart of these sparse reconstructions assumes MR images can be under-sampled in such a way that data collection time can be dramatically reduced while maintaining image quality. Here, we report experimental results that demonstrate our proposed MRI recovery algorithm on data from the publicly available fastMRI dataset [31]. Our aim is not to develop an algorithm to achieve the state-of-the-art performance for MRI image recovery (baseline deep neural network methods such as variational networks and U-nets [33,34]). The experiment is designed to demonstrate that using the proposed matrix factorization method can be a promising approach for recovering under-sampled MRI images. In this application, one of the factors in our proposed decomposition is adapted to the acceleration mask (restricted by physically prescribed sensing mechanism), which otherwise does not allow for direct RIP-based recovery guarantees in the 1 -synthesis approach to CS. A recent survey of applying the compressed sensing technique to the MRI image recovery [35] presents a variety of techniques to improve the quality of recovered images. In this study, we compared the performance of our method to that based on the TV approach, which promotes sparsity of spacial gradients by approximates a solution to (see [36,37,38]):
argmin Z R n 1 × n 2 Y R F 1 Z F 2 F 2 + λ i , j Z ( i , j ) 2 .
Under-sampling masks (restrained by the instrumental conditions) for 4-fold and 8-fold acceleration were applied to full k-space data according to (5). The masks sample 25%, respectively 12.5%, of horizontal scan lines including a fully sampled center region—corresponding to low spatial frequencies—containing 8%, respectively 4%, of scan lines, as shown in Figure 5. We selected CDF 9/7 wavelets with three levels as the sparsifying dictionaries. The lowpass wavelet band and the central areas of the acceleration masks may not be perfectly aligned, leading to potential artifacts in the reconstruction as a result of cross-band interference between the lowpass and highpass wavelet bands.
In the provided images Figure 6, Figure 7, Figure 8 and Figure 9, we can see a visual comparison between the reconstruction outcomes generated by our algorithm and the total-variation (TV) method. These comparisons are based on 4-fold and 8-fold under-sampled Fourier measurements of knee and brain scans from the fastMRI dataset. Parameters were chosen for best performance from a set of values as ρ = 0.02 , ν = 0.00016 , μ = 0.00024 , as well as λ = 5 × 10 5 in (13), and γ = 0.0035 in (10). MRI data: https://fastmri.med.nyu.edu/ (accessed date: 1 July 2022). The parameters were established through empirical testing and adjustment (to obtain the source codes for our algorithm, can contact [email protected] or [email protected]). Images (displayed with 90 degree rotation and cut to quadratic center region) taken from coronal proton density knee dataset (image size 372 × 640, z-slice 20 of file1000031 (knee A) and file1000071 (knee B)), and brain dataset (image size 320 × 640, z-slice 2 of file_brain_AXT1_201_6002786 (brain A) and file_brain_AXT1_201_6002740 (brain B)). The parameters were established through empirical testing and adjustment. Table 1 summarizes a comparison of PSNR and SSIM values for the knee and brain images depicted in the figures. Our average PSNR gains for 4-fold under-sampled data are 0.5 dB on the knee and 0.9 dB on the brain images, while our average gains for the 8-fold under-sampled images are about 2 dB. Our SSIM values are superior to that of the TV-method in all experiments.

6. Conclusions

This article describes a novel approach to CS involving the construction a sensing matrix for a prescribed sparsifying dictionary, such that their composition has the RIP. This property theoretically guarantees the accurate recovery of a fixed number of sparse coefficients. While previous theoretical work has demonstrated RIP assurances only for orthonormal matrices, our study extends these insights to include general over-complete dictionaries. We assert that our sensing matrix can recover sparse coefficients from an over-complete dictionary as effectively as a random sensing matrix for an orthonormal matrix. This advancement is significant, as many practical applications that employ compressive sensing rely on over-complete dictionaries. We further apply the factorization approach to accelerated MR imaging—deriving an embedding for under-sampled k-space data that facilitates RIP recovery guarantees. In this application one of the factors in our proposed decomposition is adapted to the physically prescribed sensing mechanism, which otherwise does not allow for direct RIP based recovery guarantees in the 1 -synthesis approach to CS.
In addition to the achievements mentioned, it is notable to recognize that the analysis could be further improved if the factorization structure in G has a specific arrangement. A structured sensing matrix can significantly enhance storage and computational efficiency in large systems. A reviewer proposes a promising avenue for future exploration for MRI acceleration. Traditional compressive sensing (CS) recovery techniques and modern neural network methods generate different artifacts. While the artifacts from conventional algorithms can usually be traced back to the algorithm and the regularization penalty, those produced by neural networks tend to be more unpredictable. Different neural networks tend to produce distinct model-shifting artifacts. Additionally, the unpredictable nature of the artifacts generated by fixed neural networks presents a significant challenge. Experienced clinicians can quickly identify the artifacts associated with compressive sensing (CS) recovery methods, which helps reduce false positives. However, neural network methods offer a significant advantage, achieving better recovery accuracy and inference speed than traditional techniques. In line with the reviewer’s suggestions, we recommend developing a hybrid approach that combines the strengths and weaknesses of both neural networks and CS techniques to optimize overall performance.

Author Contributions

Conceptualization, W.-L.H.; Methodology, J.H., W.-L.H. and A.H.; Software, J.H.; Validation, J.H., W.-L.H. and A.H.; Formal analysis, J.H., W.-L.H. and A.H.; Data curation, J.H.; Writing—original draft, W.-L.H.; Writing—review & editing, A.H.; Visualization, J.H.; Supervision, W.-L.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

All MRI data in Figure 6(a0), Figure 7(a0), Figure 8(a0) and Figure 9(a0) have been made publicly available at the fastMRI dataset and can be accessed at https://fastmri.med.nyu.edu/ (accessed date: 1 July 2022).

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

To solve (8) numerically via the Lagrangian approach, we consider the augmented Lagrangian:
L 1 ( G , G ˜ ) : = 1 2 R ( Re ( F 1 ) G I ) F 2 + Tr λ 1 ( G ˜ G I ) + Tr λ 2 ( G G ˜ I ) + ρ 2 G ˜ G I F 2 + G G ˜ I F 2 ,
with fixed ρ > 0 . We initialize G R as in (4), G ˜ as G R 1 , and λ 1 = λ 2 as zero matrices, and iterate the updates until convergence of λ 1 and λ 2 :
(i)
G R argmin G L 1 ( G , G ˜ ) ;
(ii)
G ˜ argmin G ˜ L 1 ( G R , G ˜ ) ;
(iii)
λ 1 λ 1 + ρ ( G ˜ G R I ) and λ 2 λ 2 + ρ ( G R G ˜ I ) .
where steps (i) and (ii) can be approximated by solving the Sylvester equations G L 1 = 0 and G ˜ L 1 = 0 numerically (e.g., via matlab function sylvester), where:
G L 1 = Re ( F 1 ) R ( Re ( F 1 ) G I ) + G ˜ λ 1 + λ 2 G ˜ + ρ ( G ˜ ( G ˜ G I ) + ( G G ˜ I ) G ˜ ) = ( Re ( F 1 ) R Re ( F 1 ) + ρ G ˜ G ˜ ) G + G ( ρ G ˜ G ˜ ) + ( G ˜ λ 1 + λ 2 G ˜ Re ( F 1 ) R 2 ρ G ˜ ) ,
and
G ˜ L 1 = G λ 2 + λ 1 G + ρ ( ( G ˜ G I ) G + G ( G G ˜ I ) ) = ρ ( G G G ˜ + G ˜ G G ) + G λ 2 + λ 1 G 2 ρ G .
Problem (9) is solved analogously, now considering:
L 2 ( H , H ˜ ) : = 1 2 D 1 G R A H F 2 + ν 2 H H ˜ F 2 + Tr λ 3 ( H ˜ H I ) + Tr λ 4 ( H H ˜ I ) + μ 2 H ˜ H I F 2 + H H ˜ I F 2 ,
initializing H R as in (4), H ˜ as H R and λ 3 = λ 4 as zero matrices. The iteration steps are now:
(i’)
H R argmin H L 2 ( H , H ˜ ) ;
(ii’)
H ˜ argmin H ˜ L 2 ( H R , H ˜ ) ;
(iii’)
λ 3 λ 3 + μ ( H ˜ H R I ) and λ 4 λ 4 + μ ( H R H ˜ I ) ,
and the derivatives concerning (i’) and (ii’) are:
H L 2 = A G ( D G A H ) + ν H ˜ ( H H ˜ ) + H ˜ λ 3 + λ 4 H ˜ + μ ( H ˜ ( H ˜ H I ) ) + μ ( ( H H ˜ I ) H ˜ ) = ( A G G A + ν H ˜ + μ H ˜ H ˜ ) H + H ( μ H ˜ H ˜ ) + A G D + H ˜ λ 3 + λ 4 H ˜ 2 μ H ˜ ν H ˜ H ˜ ,
and
H ˜ L 2 = ν H ( H ˜ H ) + λ 3 H + H λ 4 + μ ( ( H ˜ H I ) H ) + μ ( H ( H H ˜ I ) ) = ( ν H + μ H H ) H ˜ + H ˜ ( μ H H ) + ( λ 3 H + H λ 4 2 μ H ν H H ) .

References

  1. Candès, E.J.; Romberg, J.; Tao, T. Robust uncertainty principles: Exact signal reconstruction from highly incomplete frequency information. IEEE Trans. Inf. Theory 2006, 52, 489–509. [Google Scholar] [CrossRef]
  2. Donoho, D.L. Compressed sensing. IEEE Trans. Inf. Theory 2006, 52, 1289–1306. [Google Scholar] [CrossRef]
  3. Candès, E.J.; Tao, T. Decoding by linear programming. IEEE Trans. Inf. Theory 2005, 51, 4203–4215. [Google Scholar] [CrossRef]
  4. Candès, E.J.; Romberg, J.K.; Tao, T. Stable signal recovery from incomplete and inaccurate measurements. Commun. Pure Appl. Math. 2006, 59, 1207–1223. [Google Scholar] [CrossRef]
  5. Candès, E.J. The restricted isometry property and its implications for compressed sensing. Comptes Rendus Math. 2008, 346, 589–592. [Google Scholar] [CrossRef]
  6. Foucart, S. A note on guaranteed sparse recovery via 1-minimization. Appl. Comput. Harmon. Anal. 2010, 29, 97–103. [Google Scholar] [CrossRef]
  7. Tropp, J. Just relax: Convex programming methods for identifying sparse signals in noise. IEEE Trans. Inf. Theory 2006, 52, 1030–1051. [Google Scholar] [CrossRef]
  8. Baraniuk, R.; Davenport, M.; DeVore, R.; Wakin, M. A simple proof of the restricted isometry property for random matrices. Constr. Approx. 2008, 28, 253–263. [Google Scholar] [CrossRef]
  9. Candès, E.J.; Tao, T. Near optimal signal recovery from random projections: Universal encoding strategies? IEEE Trans. Inf. Theory 2006, 52, 5406–5425. [Google Scholar] [CrossRef]
  10. Rudelson, M.; Vershynin, R. On sparse reconstruction from Fourier and Gaussian measurements. Commun. Pure Appl. Math. 2008, 61, 1025–1045. [Google Scholar] [CrossRef]
  11. Bourgain, J. An improved estimate in the restricted isometry problem. In Geometric Aspects of Functional Analysis: Israel Seminar (GAFA) 2011–2013; Springer: Berlin, Germany, 2014; pp. 65–70. [Google Scholar]
  12. Lustig, M.; Donoho, D.L.; Santos, J.M.; Pauly, J.M. Compressed sensing MRI. IEEE Signal Process. Mag. 2008, 25, 72–82. [Google Scholar] [CrossRef]
  13. Candès, E.J.; Eldar, Y.C.; Needell, D.; Randall, P. Compressed sensing with coherent and redundant dictionaries. Appl. Comput. Harmon. Anal. 2011, 31, 59–73. [Google Scholar] [CrossRef]
  14. Krahmer, F.; Ward, R. New and improved Johnson–Lindenstrauss embeddings via the restricted isometry property. SIAM J. Math. Anal. 2011, 43, 1269–1281. [Google Scholar] [CrossRef]
  15. Elad, M.; Milanfar, P.; Rubinstein, R. Analysis versus synthesis in signal priors. Inverse Probl. 2007, 23, 947–968. [Google Scholar] [CrossRef]
  16. Tropp, J.A.; Gilbert, A.C. Signal recovery from random measurements via orthogonal matching pursuit. IEEE Trans. Inf. Theory 2007, 53, 4655–4666. [Google Scholar] [CrossRef]
  17. Cohen, A.; Dahmen, W.; DeVore, R. Compressed sensing and best k-term approximation. J. Amer. Math. Soc. 2009, 22, 211–231. [Google Scholar] [CrossRef]
  18. Upadhyaya, V.; Salim, M. Compressive sensing: Methods, techniques, and applications. In IOP Conference Series: Materials Science and Engineering; IOP Publishing: Bristol, UK, 2021; Volume 1099, p. 012012. [Google Scholar]
  19. Sankararajan, R.; Rajendran, H.; Sukumaran, A.N. Compressive Sensing for Wireless Communication: Challenges and Opportunities; River Publishers: Nordjylland, Denmark, 2022. [Google Scholar]
  20. Gibson, G.M.; Johnson, S.D.; Padgett, M.J. Single-pixel imaging 12 years on: A review. Opt. Express 2020, 28, 28190–28208. [Google Scholar] [CrossRef] [PubMed]
  21. Lustig, M.; Donoho, D.; Pauly, J.M. Sparse MRI: The application of compressed sensing for rapid MR imaging. Magn. Reson. Med. 2007, 58, 1182–1195. [Google Scholar] [CrossRef]
  22. Pal, A.; Rathi, Y. A review and experimental evaluation of deep learning methods for MRI reconstruction. J. Mach. Learn. Biomed. Imag. 2022. [Google Scholar] [CrossRef]
  23. Gu, H.; Yaman, B.; Moeller, S.; Ellermann, J.; Ugurbil, K.; Akçakaya, M. Revisiting 1-wavelet compressed-sensing MRI in the era of deep learning. Proc. Natl. Acad. Sci. USA 2022, 119, e2201062119. [Google Scholar] [CrossRef]
  24. Recht, M.; Zbontar, J.; Sodickson, D.; Knoll, F.; Yakubova, N.; Sriram, A.; Murrell, T.; Defazio, A.; Rabbat, M.; Rybak, L.; et al. Using deep learning to accelerate knee MRI at 3T: Results of an interchangeability study. Amer. J. Roentgenol. 2020, 215, 1421–1429. [Google Scholar] [CrossRef] [PubMed]
  25. Knoll, F.; Zbontar, J.; Sriram, A.; Muckley, M.; Bruno, M.; Defazio, A.; Parente, M.; Geras, K.; Katsnelson, J.; Chandarana, H.; et al. fastMRI: A publicly available raw k-space and DICOM dataset of knee images for accelerated MR image reconstruction using machine learning. Radiol. Artif. Intell. 2020, 2, e190007. [Google Scholar] [CrossRef] [PubMed]
  26. Shimron, E.; Tamir, J.I.; Wang, K.; Lustig, M. Subtle inverse crimes: Naïvely training machine learning algorithms could lead to overly-optimistic results. arXiv 2021, arXiv:2109.08237. [Google Scholar]
  27. Liu, X.; Wang, J.; Peng, C.; Chandra, S.S.; Liu, F.; Zhou, S.K. Undersampled MRI reconstruction with side information-guided normalisation. arXiv 2022, arXiv:2203.03196. [Google Scholar]
  28. Liu, X.; Wang, J.; Liu, F.; Zhou, S.K. Universal undersampled MRI reconstruction. In Proceedings of the Medical Image Computing and Computer Assisted Intervention—MICCAI 2021: 24th Int. Conf., Strasbourg, France, 27 September–1 October 2021, Proc., VI; Springer: Berlin/Heidelberg, Germany, 2021; pp. 211–221. [Google Scholar]
  29. Taori, R.; Dave, A.; Shankar, V.; Carlini, N.; Recht, B.; Schmidt, L. Measuring robustness to natural distribution shifts in image classification. Adv. Neural Inf. Process. Syst. 2020, 33, 18583–18599. [Google Scholar]
  30. Rasool, M.A.; Ahmed, S.; Sabina, U.; Whangbo, T.K. Konet: Towards a weighted ensemble learning model for knee osteoporosis classification. IEEE Access 2024, 12, 5731–5742. [Google Scholar] [CrossRef]
  31. Zbontar, J.; Knoll, F.; Sriram, A.; Murrell, T.; Huang, Z.; Muckley, M.J.; Defazio, A.; Stern, R.; Johnson, P.; Bruno, M.; et al. fastMRI: An open dataset and benchmarks for accelerated MRI. arXiv 2018, arXiv:1811.08839. [Google Scholar]
  32. Needell, D.; Tropp, J.A. CoSaMP: Iterative signal recovery from incomplete and inaccurate samples. Appl. Comput. Harmon. Anal. 2009, 26, 301–321. [Google Scholar] [CrossRef]
  33. Knoll, F.; Murrell, T.; Sriram, A.; Yakubova, N.; Zbontar, J.; Rabbat, M.; Defazio, A.; Muckley, M.; Sodickson, D.; Zitnick, C.; et al. Advancing machine learning for MR image reconstruction with an open competition: Overview of the 2019 fastMRI challenge. Magn. Reson. Med. 2020, 84, 3054–3070. [Google Scholar] [CrossRef]
  34. Muckley, M.J.; Riemenschneider, B.; Radmanesh, A.; Kim, S.; Jeong, G.; Ko, J.; Jun, Y.; Shin, H.; Hwang, D.; Mostapha, M.; et al. Results of the 2020 fastMRI challenge for machine learning MR image reconstruction. IEEE Trans. Med. Imag. 2021, 40, 2306–2317. [Google Scholar] [CrossRef]
  35. Ye, J.C. Compressed sensing mri: A review from signal processing perspective. BMC Biomed. Eng. 2019, 1, 1–17. [Google Scholar] [CrossRef] [PubMed]
  36. Chambolle, A. An algorithm for total variation minimization and applications. J. Math. Imaging Vis. 2004, 20, 89–97. [Google Scholar]
  37. Landi, G.; Piccolomini, E.L.; Zama, F. A total variation-based reconstruction method for dynamic mri. Comput. Math. Methods Med. 2008, 9, 69–80. [Google Scholar] [CrossRef]
  38. Panić, M.; Jakovetić, D.; Vukobratović, D.; Crnojević, V.; Pižurica, A. Mri reconstruction using markov random field and total variation as composite prior. Sensors 2020, 20, 3185. [Google Scholar] [CrossRef]
Figure 1. Visualization of sparsifying dictionaries (of size 128 × 1024 ) with integer values ranging from 0 to 255.
Figure 1. Visualization of sparsifying dictionaries (of size 128 × 1024 ) with integer values ranging from 0 to 255.
Signals 05 00044 g001
Figure 2. Matrix G R 128 × 128 of several sparsifying dictionaries D for various A. Recall that D = G A H and the sensing matrix of D is E G 1 . In (a1,b1), A is a Gaussian random matrix. In (a2,b2), A is a Bernoulli random matrix. In (a1,a2), D = D K S V D . In (b1,b2), D = D w a v e l e t .
Figure 2. Matrix G R 128 × 128 of several sparsifying dictionaries D for various A. Recall that D = G A H and the sensing matrix of D is E G 1 . In (a1,b1), A is a Gaussian random matrix. In (a2,b2), A is a Bernoulli random matrix. In (a1,a2), D = D K S V D . In (b1,b2), D = D w a v e l e t .
Signals 05 00044 g002
Figure 3. Comparisons of CS recovery performance (i.e., the probability of sparse vector recovery versus CS ratio) using sparsifying dictionary D K S V D . Red and blue curves were, respectively, obtained using the benchmark (12) and our approach (11). Sparse vectors x were randomly generated and each point on the curve is the average of 2000 probability measurements. The positions of non-zero coefficients of x are uniformly distributed and the values of the non-zero coefficients of x are uniformly distributed in [ 1 , 1 ] . In (a1,a2), sparsity level is 10; in (b1,b2), sparsity level is 12; and in (c1,c2), sparsity level is 14.
Figure 3. Comparisons of CS recovery performance (i.e., the probability of sparse vector recovery versus CS ratio) using sparsifying dictionary D K S V D . Red and blue curves were, respectively, obtained using the benchmark (12) and our approach (11). Sparse vectors x were randomly generated and each point on the curve is the average of 2000 probability measurements. The positions of non-zero coefficients of x are uniformly distributed and the values of the non-zero coefficients of x are uniformly distributed in [ 1 , 1 ] . In (a1,a2), sparsity level is 10; in (b1,b2), sparsity level is 12; and in (c1,c2), sparsity level is 14.
Signals 05 00044 g003
Figure 4. Comparisons of CS recovery performance (i.e., the probability of sparse vector recovery versus CS ratio) using sparsifying dictionary D w a v e l e t (CDF 9/7). Red and blue curves were, respectively, obtained using the benchmark (12) and our approach (11). Sparse vectors x were randomly generated and each point on the curve is the average of 2000 probability measurements. The positions of non-zero coefficients of x are uniformly distributed and the values of the non-zero coefficients of x are uniformly distributed in [ 1 , 1 ] . In (a1,a2), sparsity level is 10; in (b1,b2), sparsity level is 12; and in (c1,c2), sparsity level is 14.
Figure 4. Comparisons of CS recovery performance (i.e., the probability of sparse vector recovery versus CS ratio) using sparsifying dictionary D w a v e l e t (CDF 9/7). Red and blue curves were, respectively, obtained using the benchmark (12) and our approach (11). Sparse vectors x were randomly generated and each point on the curve is the average of 2000 probability measurements. The positions of non-zero coefficients of x are uniformly distributed and the values of the non-zero coefficients of x are uniformly distributed in [ 1 , 1 ] . In (a1,a2), sparsity level is 10; in (b1,b2), sparsity level is 12; and in (c1,c2), sparsity level is 14.
Signals 05 00044 g004
Figure 5. Scan line locations for 4-fold (left) and 8-fold (right) under-sampling, including fully sampled low-frequency region.
Figure 5. Scan line locations for 4-fold (left) and 8-fold (right) under-sampling, including fully sampled low-frequency region.
Signals 05 00044 g005
Figure 6. (a0) FFT reconstruction of fully sampled MRI k-space data (file 1000031.h5: the 20th in a set of 40 datasets, each comprising 372 × 640 data points, representing coronal proton density (PD) knee images). (a1) Image of the proposed algorithm with 4-fold sampling. (a2) Image of the proposed algorithm with 8-fold sampling. (b1) Image of the TV algorithm with 4-fold sampling. (b2) Image of the TV algorithm with 8-fold sampling. The vertical band patterns observed in the images are artifacts resulting from under-sampling. The artifact in (a2) is less noticeable compared to (b2).
Figure 6. (a0) FFT reconstruction of fully sampled MRI k-space data (file 1000031.h5: the 20th in a set of 40 datasets, each comprising 372 × 640 data points, representing coronal proton density (PD) knee images). (a1) Image of the proposed algorithm with 4-fold sampling. (a2) Image of the proposed algorithm with 8-fold sampling. (b1) Image of the TV algorithm with 4-fold sampling. (b2) Image of the TV algorithm with 8-fold sampling. The vertical band patterns observed in the images are artifacts resulting from under-sampling. The artifact in (a2) is less noticeable compared to (b2).
Signals 05 00044 g006
Figure 7. (a0) FFT reconstruction of fully sampled MRI k-space data (file 1000071.h5: the 20th in a set of 35 datasets, each comprising 372 × 640 data points, representing coronal proton density (PD) knee images). (a1) Image of the proposed algorithm with 4-fold sampling. (a2) Image of the proposed algorithm with 8-fold sampling. (b1) Image of the TV algorithm with 4-fold sampling. (b2) Image of the TV algorithm with 8-fold sampling.
Figure 7. (a0) FFT reconstruction of fully sampled MRI k-space data (file 1000071.h5: the 20th in a set of 35 datasets, each comprising 372 × 640 data points, representing coronal proton density (PD) knee images). (a1) Image of the proposed algorithm with 4-fold sampling. (a2) Image of the proposed algorithm with 8-fold sampling. (b1) Image of the TV algorithm with 4-fold sampling. (b2) Image of the TV algorithm with 8-fold sampling.
Signals 05 00044 g007
Figure 8. (a0) The fully sampled MRI k-space data (file_brain_AXT1_201_6002786.h5) involves the second set of 16 brain datasets, each containing 320 × 640 data points. (a1) Image displaying the proposed algorithm with 4-fold sampling. (a2) Image displaying the proposed algorithm with 8-fold sampling. (b1) Image showing the TV algorithm with 4-fold sampling. (b2) Image showing the TV algorithm with 8-fold sampling. The structure is more discernible in (a2) and (a1) than in (b2) and (b1), respectively.
Figure 8. (a0) The fully sampled MRI k-space data (file_brain_AXT1_201_6002786.h5) involves the second set of 16 brain datasets, each containing 320 × 640 data points. (a1) Image displaying the proposed algorithm with 4-fold sampling. (a2) Image displaying the proposed algorithm with 8-fold sampling. (b1) Image showing the TV algorithm with 4-fold sampling. (b2) Image showing the TV algorithm with 8-fold sampling. The structure is more discernible in (a2) and (a1) than in (b2) and (b1), respectively.
Signals 05 00044 g008
Figure 9. (a0) The fully sampled MRI k-space data (file_brain_AXT1_201_6002740.h5) involves the second set of 16 brain datasets, each containing 320 × 640 data points. (a1) Image displaying the proposed algorithm with 4-fold sampling. (a2) Image displaying the proposed algorithm with 8-fold sampling. (b1) Image showing the TV algorithm with 4-fold sampling. (b2) Image showing the TV algorithm with 8-fold sampling. The structure is more discernible in (a2) and (a1) than in (b2) and (b1), respectively.
Figure 9. (a0) The fully sampled MRI k-space data (file_brain_AXT1_201_6002740.h5) involves the second set of 16 brain datasets, each containing 320 × 640 data points. (a1) Image displaying the proposed algorithm with 4-fold sampling. (a2) Image displaying the proposed algorithm with 8-fold sampling. (b1) Image showing the TV algorithm with 4-fold sampling. (b2) Image showing the TV algorithm with 8-fold sampling. The structure is more discernible in (a2) and (a1) than in (b2) and (b1), respectively.
Signals 05 00044 g009
Table 1. Performance comparison for images in Figure 6, Figure 7, Figure 8 and Figure 9, for 4-fold and 8-fold acceleration.
Table 1. Performance comparison for images in Figure 6, Figure 7, Figure 8 and Figure 9, for 4-fold and 8-fold acceleration.
ProposedTV Method
PSNRSSIMPSNRSSIM
Knee A28.526.40.6570.57028.224.50.6150.538
Knee B28.426.10.6350.53527.824.10.6250.541
Brain A27.724.10.6420.52426.722.30.6370.493
Brain B27.724.60.6680.49426.822.70.6440.474
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ho, J.; Hwang, W.-L.; Heinecke, A. RIP Sensing Matrices Construction for Sparsifying Dictionaries with Application to MRI Imaging. Signals 2024, 5, 794-811. https://doi.org/10.3390/signals5040044

AMA Style

Ho J, Hwang W-L, Heinecke A. RIP Sensing Matrices Construction for Sparsifying Dictionaries with Application to MRI Imaging. Signals. 2024; 5(4):794-811. https://doi.org/10.3390/signals5040044

Chicago/Turabian Style

Ho, Jinn, Wen-Liang Hwang, and Andreas Heinecke. 2024. "RIP Sensing Matrices Construction for Sparsifying Dictionaries with Application to MRI Imaging" Signals 5, no. 4: 794-811. https://doi.org/10.3390/signals5040044

APA Style

Ho, J., Hwang, W.-L., & Heinecke, A. (2024). RIP Sensing Matrices Construction for Sparsifying Dictionaries with Application to MRI Imaging. Signals, 5(4), 794-811. https://doi.org/10.3390/signals5040044

Article Metrics

Back to TopTop