Next Article in Journal
An Ultra-Broadband Conductor-Backed Coplanar Waveguide with Sine Edges
Previous Article in Journal
Vertical Microfluidic Trapping System for Capturing and Simultaneous Electrochemical Detection of Cells
Previous Article in Special Issue
Degradation Type-Aware Image Restoration for Effective Object Detection in Adverse Weather
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Isotropic Brain MRI Reconstruction from Orthogonal Scans Using 3D Convolutional Neural Network

1
School of Big Data and Artificial Intelligence, Chengdu Technological University, Chengdu 611730, China
2
Digital Twin Laboratory, Chengdu Technological University, Chengdu 611730, China
*
Authors to whom correspondence should be addressed.
Sensors 2024, 24(20), 6639; https://doi.org/10.3390/s24206639
Submission received: 9 August 2024 / Revised: 4 September 2024 / Accepted: 10 October 2024 / Published: 15 October 2024

Abstract

:
As an alternative to true isotropic 3D imaging, image super-resolution (SR) has been applied to reconstruct an isotropic 3D volume from multiple anisotropic scans. However, traditional SR methods struggle with inadequate performance, prolonged processing times, and the necessity for manual feature extraction. Motivated by the exceptional representational ability and automatic feature extraction of convolutional neural networks (CNNs), in this work, we present an end-to-end isotropic MRI reconstruction strategy based on deep learning. The proposed method is based on 3D convolutional neural networks (3D CNNs), which can effectively capture the 3D structural features of MRI volumes and accurately predict potential structure. In addition, the proposed method takes multiple orthogonal scans as input and thus enables the model to use more complementary information from different dimensions for precise inference. Experimental results show that the proposed algorithm achieves promising performance in terms of both quantitative and qualitative assessments. In addition, it can process a 3D volume with a size of 256 × 256 × 256 in less than 1 min with the support of an NVIDIA GeForce GTX 1080Ti GPU, which suggests that it is not only a quantitatively superior method but also a practical one.

1. Introduction

Magnetic resonance imaging (MRI) is a crucial and versatile medical imaging modality broadly used in clinical diagnosis and image-guided therapeutics. In principle, it utilizes the magnetic resonance phenomenon to collect data in the frequency domain and then generates images through inverse Fourier Transformation in situations where many sensors are involved in the process of generating MR images, including magnetic field generation, pulse signal excitation and reception, etc. [1,2]. In many MRI experiments, a basic consideration is how to achieve a good equilibrium between spatial resolution, signal-to-noise ratio (SNR), and acquisition time [3]. For instance, to reduce motion artifacts and improve the SNR of MR slices, many MRI scans are performed with relatively few slices and rather large slice thicknesses [4]. Consequently, most 3D MR volumes are collected as tomographic sets of image slices, which have higher resolution in imaging planes and lower resolution along the slice-select direction, causing anisotropic spatial resolution in 3D space [5]. This could be problematic for downstream tasks, such as computer-aided diagnosis [6], quantitative analysis, and visualization since the image will be missing more high-frequency information in the through-plane direction [4]. Therefore, it is much desired and beneficial to produce MRI data with 3D isotropic spatial resolution in practical applications.
An intuitive manner to acquire 3D isotropic and high spatial resolution MRI volumes is to improve the hardware devices of MRI scanners, e.g., higher magnetic fields, stronger and faster gradients, etc. However, these solutions are often expensive, require hardware upgrades, and are still subject to various complex factors, e.g., physical constraints, sensor performance, and system noise [4,5]. Conversely, to keep a similar SNR, reducing the voxel size to produce isotropic resolution requires averaging multiple acquisitions, thus increasing the imaging time and inflexibility in routine practice [7].
Another alternative is to enhance the resolution of magnetic resonance (MR) images along the through-plane direction using image post-processing techniques. One popular technique is termed super-resolution (SR) imaging, which aims at recovering a high-resolution (HR) image from one or more low-resolution (LR) images [8]. As a classical problem, image SR is still an active yet challenging hotspot in both the natural and medical image-processing communities [9]. In the literature, a variety of SR methods have been studied, such as interpolation methods [10], edge-based methods [11], modeling and reconstruction methods [12], and example-based learning methods [13], as well as dictionary learning methods [14,15]. In terms of isotropic MRI reconstruction, the methods in [4,16] adopted 3D block-based self-similarity learning and sparse representation to reconstruct HR volumes with isotropic resolution, which are essentially shallow learning techniques. These methods are inherently limited in SR performance because (1) finite additional information is used for solving the severely ill-posed inverse problem, and (2) the representational capacity of these models is inadequate for accurate SR inference [8,17].
In recent years, deep-learning-based [18] SR algorithms have been broadly and actively studied, and significantly facilitated the rapid development of various benchmark SR tasks. Owing to the powerful capacity of deep models to capture hierarchical features that map from LR space to HR space [8,19], recent methods based on deep learning [18] techniques have achieved great improvements in SR performance.
For natural images, a pioneering approach is SRCNN [20], which applies a three-layer convolutional neural network (CNN) [21] to super-resolve a single input image. Subsequently, a host of CNN-based SR models have been reported and some strategies have also been developed to improve the performance of deep models, such as increasing the model scale (including network parameter, depth as well as width) [22,23], residual learning [24,25], directly mapping from input LR images [26,27], adversarial training [28,29], etc. Other representative SR models include FC2N [30], SAN [31], SwinIR [32], ELAN [33], SRFormer [34], etc. In medical image processing, improving image quality and mitigating image degradation, e.g., image enhancement [35] and artifact removal [36], has been proposed to help promote the performance of subsequent processing and analysis. There are also some efforts to utilize deep learning technology to deal with medical image SR tasks, such as CSN [17], SERAN [37], DisC-Diff [38], Dual-ArbNet [39], McMRSR [40], etc. In the context of MRI isotropic reconstruction, however, there are relatively few works based on deep learning. One of the latest and possibly the most relevant works on this topic proposed to super-resolve brain MR images through the use of 3D CNNs [41], but it mainly focused on general image SR tasks, rather than isotropic MRI reconstruction.
Although significant progress has been made in deep learning in recent years, with advanced model architectures such as attention mechanisms [31,42,43] and Transformers [44,45,46] enhancing the performance of related tasks, these advanced structures are not very suitable for isotropic resolution reconstruction of MR images, as clinical applications have high requirements for fast inference. For 3D MR volumes, this issue is more prominent. Therefore, more advanced but complex architectures might not be suitable for isotropic MRI reconstruction in 3D scenarios.
In this work, we propose an isotropic reconstruction super-resolution network (isoSRN) to solve the problem of isotropic MRI reconstruction. For local feature fusion, our isoSRN extends weighted channel concatenation [30] and wide activation [47] to 3D space to capture sufficient 3D structural information, thereby contributing to accurate nonlinear SR inference. As with many other SR models [17,22,32,34], we introduce residual global skip connection (RGSC) to ease the training difficulty of the deep models. In addition, our isoSRN takes 3D orthogonal scans as input to make full use of orthogonal supplementary information, as in [4]. However, we utilize this information for deep feature learning instead of traditional sparse representation. The overall structure of our isoSRN is illustrated in Figure 1. With the strong representational capacity of 3D CNN models, our isoSRN can recover isotropic HR volumes more accurately. Extensive experiments show that our model is noticeably superior to other methods both quantitatively and qualitatively. The main contributions of this paper can be summarized as follows:
We present a simple yet efficient 3D model for isotropic MRI reconstruction that utilizes multiple orthogonal LR volumes with anisotropic resolution to generate an isotropic HR volume.
The proposed model is built upon simply extending wide activation [47] and weighted channel concatenation [30] to 3D space, which can promote 3D feature learning while maintaining efficient inference of the model.
The proposed model is evaluated on several simulated and real MRI datasets, and it shows significant superiority to other compared methods in terms of both quantitative evaluations and qualitative analyses.
The rest of this work is arranged as follows. We first present the SR problem formulation in Section 2, and the details of the proposed model in Section 3. Then, the experimental results and analyses are given in Section 4. Finally, we conclude and discuss the whole paper in Section 6.

2. Problem Formulation

Image SR is usually formalized as an inverse problem with an ill-posed nature, which reconstructs an HR MR volume from one or more LR MR volumes according to the MRI imaging model. Given V LR observations x 1 , x 2 , , x V R h × w × d and their HR volume y R H × W × D , the imaging model (or HR volume degradation model) can be usually formulated as:
x v = D v M v y + n v , v = 1 , 2 , , V ,
where M v represents the joint degradation operations on HR volume y to generate the v-th LR volume x v , such as blurring and geometric transformations, etc., and D v is a downsampling operation that reduces the shape of HR volume to that of the LR volume x v . n v usually denotes the Rician noise [48]. This formulation describes the task as a multiple-image SR (MISR) problem and it degrades to a single-image SR (SISR) problem when V = 1 . In the context of unsupervised learning, the SR image can be solved by optimizing the following loss:
y * = arg min y v = 1 V | | x v D v M v y | | 2 2 ,
where y * is the prediction of HR image y . Directly optimizing Equation (2) typically leads to unstable solutions that require the utilization of appropriate regularization terms to well pose and stabilize the solution. Therefore, the general form of the target for image SR is usually expressed as:
y * = arg min y v = 1 V | | x v D v M v y | | 2 2 + γ ϕ ( y ) ,
where ϕ ( y ) is the regularization term associated with an image prior, and γ is a non-negative trade-off parameter that governs the compromise between the regularization term ϕ ( y ) and the fidelity term. However, the use of regularization still suffers from unstable solutions as it modifies implicitly the acceptable solution space without any guarantee of recovering realistic HR volumes [41], and how to choose a suitable regularization term and optimal trade-off parameter γ is practically difficult.
Compared with unsupervised learning, supervised learning techniques are superior in that they can reconstruct novel details that are not available in LR images. In a supervised context, the HR image y can be recovered with the following formulation, in which regularization is implicitly contained:
y * = arg min y v = 1 V | | y H v U v x v | | 2 2 ,
where H v and U v formalize the process of HR reconstruction and upsampling for x v . For easier HR reconstruction and SR inference, we convert the MISR problem in Equation (4) to an SISR problem as follows: (i) upsample each LR volume x v by, e.g., interpolation: x v u = U v x v ; (ii) fuse these upsampled volumes by simple element-wise average:
x = 1 V v = 1 V x v u = 1 V v = 1 V U v x v .
As for HR reconstruction operation H v , we assume that it behaves in the same manner for each x v (or the fused LR volume x ) after the upsampling operation U v . Therefore, let H v = H and then, Equation (4) can be rewritten as:
y * = arg min y | | y H x | | 2 2 ,
which evolves into the optimization objective of a typical SISR problem in the context of supervised learning. In the case of deep learning, H is usually modeled as an artificial neural network (ANN), for instance, a typical CNN architecture for complex nonlinear SR inference.
However, as far as image SR is concerned, it has been shown that L 1 loss has better convergence than L 2 loss, which is more beneficial to MR image SR tasks [17,22]. We, therefore, optimize a L 1 loss to solve Equation (6). Given a training dataset D consisting of | D | paired HR volumes y and the corresponding LR volumes x , the L 1 loss can is defined as:
L 1 ( θ ) = 1 | D | i = 1 | D | y ( i ) F x ( i ) ; θ 1 ,
where F ( · ) = H represents the mapping function of the CNN structure from x to y , and θ denotes the parameter set of the network. y ^ = F ( x ; θ ) is the estimate of sample label y .

3. Isotropic Super-Resolution Network

3.1. 3D Feature Extraction

Most current models based on CNNs adopt convolutions with zero-padding to keep the spatial size of the output features unchanged, such as [17,20,22,30,34,43,49], etc. In this paper, all the convolutional layers we discuss are with zero-filling and single stride.
In principle, each dimension of the LR volume x v may have anisotropic scaling factors to achieve isotropic reconstruction as the resolutions of all dimensions can be different from each other, while in practice, x v usually has the same high spatial resolution in the imaging plane and different low resolution in the through-plane direction. Therefore, the isotropic reconstruction of MR volumes is essentially a 1D SR problem that can even be solved through 1D CNN models. In this case, assume that v t R L × C t denotes the intermediate feature tensor of the t-th layer with length L and channel number C t , then, the feature tensor at the ( t + 1 ) -th layer is computed by:
v m t + 1 ( x ) = σ [ b m t + n = 1 C t k = 1 K l w n m t ( k ) v n t ( x ˜ ) ] ,
where m = 1 , , C t + 1 indexes over feature channels of v t + 1 and n indexes over those of v t . K l indicates the length of 1D convolutional kernel w t R K l × C t × C t + 1 . The parentheses are employed to indicate the spatial position of feature tensors and x ˜ = x + k K l / 2 . σ ( · ) represents a nonlinear function, e.g., a ReLU, and b t is the biases at the t-th layer whose length follows the number of output channels C t + 1 . Therefore, only the information along the dimension to be upscaled is adopted to infer the feature map of the next layer, as demonstrated in Figure 2a. To use more information for inference, it is intuitive to extend the 1D convolution layer to a 3D convolution layer. Let v t R H × W × D × C t represent the feature map at the t-th layer with C t channels, then, the feature map at the ( t + 1 ) -th layer is computed as follows in the 3D case:
v m t + 1 ( x , y , z ) = σ b m t + ψ m t ( x , y , z ) .
Here ψ m t ( x , y , z ) is given by:
ψ m t ( x , y , z ) = n = 1 C t i = 1 K h j = 1 K w k = 1 K d w n m t ( i , j , k ) v n t ( x ˜ , y ˜ , z ˜ ) ,
where m and n have the same meanings as in the 1D case, and w t R K h × K w × K d × C t × C t + 1 denotes the 3D convolutional kernel of the t-th layer with spatial size of K h   ×   K w   ×   K d , and:
x ˜ = x + i K h / 2 , y ˜ = y + j K w / 2 , z ˜ = z + k K d / 2 .
Therefore, in the case of 3D, in addition to exploring features with multiple channels, the network will also make full use of information in the 3D space to perform inference. Moreover, 3D convolutional kernels increase network parameters, obviously enlarging the representational capacity of the model.

3.2. Joint Linear Attention

Attention mechanisms are widely used to process the different components of an input signal distinctively. They decide the allocation of processing resources according to the information amount and importance of these signal components. Attention has been adopted extensively in deep learning and it shows the potential to improve model performance in many tasks such as image classification [50,51], object localisation [52], image restoration [49], etc. It is typically implemented by integrating a nonlinear function (e.g., a ReLU or sigmoid) with sequential operations [50], in the form of self-attention.
Unlike the above works, Zhao et al. [30] presented a novel joint linear attention mechanism for linear and nonlinear features in the network, which is proven to be conducive to the full mining of model representational capacity. However, they mainly aim at natural image SR tasks in 2D. In this paper, we extend the joint linear attention to 3D space and use it for the isotropic reconstruction of MRI volumes. Let [ ] denote the operation of channel concatenation, then a channel concatenation block in 3D space (3D CB) can be formulated as the following:
x i = L [ π i x i 1 , λ i H ( x i 1 ) ] ,
where L ( · ) corresponds to the 1×1×1 convolutional layer at the end of a CB, as shown in Figure 1. Furthermore, H ( · ) represents the function of the nonlinear mapping branch, and π i and λ i represent the weighting factors of the identity branch and nonlinear mapping branch, respectively. Let t = i 1 and u t = H ( x i 1 ) , and we remove the spatial dimensions of the 1 × 1 × 1 convolutional kernel w t R 1 × 1 × 1 × 2 C t × C t + 1 for ease of representation. Then, we can follow the convention in Section 3.1 to reformulate the 3D weighted channel concatenation:
x m t + 1 ( x , y , z ) = n = 1 C t π t + 1 w n m t x n t ( x , y , z ) + n = 1 C t λ t + 1 w ( n + C t ) m t u n t ( x , y , z ) .
Since the overall process fuses identity mapping and nonlinear mapping, and no activation is attached to the 1   ×   1   ×   1 convolutional layer, it can be viewed as a joint linear attention of the linear and nonlinear features. In addition, owing to the learnability of w , λ and π , when (i) w n n t = w ( n + C t ) n t = 1 for n = 1 , 2 , , C t and the other elements in w t are 0, and (ii) π t + 1 = λ t + 1 = 1 , then the weighted channel concatenation actually degrades to the residual connection. In this case, if π and λ are learnable in the 2D case, it degrades to adaptive residual learning [30].

3.3. Network Architecture

The overall network architecture is shown in Figure 1. Similar to many previous SR networks, such as [17,22,30] etc., the proposed isoSRN is modularized and consists of three phases: feature extraction, non-linear mapping, and image reconstruction.
Feature extraction is achieved by a 3   ×   3   ×   3 convolutional layer, which is used to simulate the dense patch extraction in many conventional methods such as sparse representation and dictionary learning. To protect low-level visual features, it is usually embedded in the network without a nonlinear activation function. The nonlinear inference part of the proposed isoSRN model consists of a group of cascaded 3D CB blocks, whose internal structure is shown in Figure 1. For the nonlinear branch of a CB block, the sequential operations of Conv-ReLU-Conv are adopted with wide activation [47]. Suppose there are n CB blocks in the network and the mapping function of the i-th block is denoted as B i ( · ) , then the entire nonlinear mapping process of our isoSRN can be iteratively formulated as:
x n = B n ( x n 1 ) = B n ( B n 1 ( ( B 1 ( x 0 ) ) ) ) ,
where x 0 denotes the extracted feature by the first 3   ×   3   ×   3 convolutional layer, as illustrated in Figure 1, and x n implies the deep features of the network. Next, the shallow features x 0 and deep features x n are integrated by a commonly-used residual connection: x 0 + x n . Image reconstruction is implemented by two 3   ×   3   ×   3 convolutional layers. The first convolutional layer is used for further fusion of shallow and deep features and the second one is employed to map these features into HR image space, as shown in Figure 1. We also demonstrate the architecture of our model in Algorithm 1.   
Algorithm 1: The isoSRN Model for Isotropic MRI Reconstruction
Sensors 24 06639 i001

3.4. Network Scale

Network scale usually refers to the depth, width, and number of parameters of the network. In general, a larger network scale indicates a stronger representational ability of the model, as well as better performance. In particular, the depth and number of parameters have a significant impact on the performance of the model. The depth of a deep network is typically defined as the longest path from the input to output [17]. As for the proposed isoSRN, the depth can be formulated according to Figure 1:
D = 3 n + 4 ,
where n denotes the number of 3D-CBs. We set n = 16 in our implementation, and D therefore is 52. It can be seen that D is independent of the scaling factor since the nonlinear inference is in HR image space. This increases the computational effort of the model, but the advantage is that it is easy to implement and can deal with arbitrary scaling factors, including fractional factors. Conversely, the proposed isoSRN takes ≈ 3.63 M learnable parameters according to Figure 1 and the description in Section 4.1. This is a moderate amount of parameters, which is also amenable to practical deployment.

4. Experimental Results

4.1. Datasets and Implementation Details

We randomly chose 155 pair structural MR volumes from the HCP dataset (https://www.humanconnectome.org/, accessed on 5 September 2022) [53] (including both T1 and T2 data). These volumes were collected with 0.7 mm × 0.7 mm × 0.7 mm isotropic resolution and matrix size of 260   ×   311   ×   260 . We divided these 155 volumes into 100 training samples, 50 testing samples, and 5 fast validation samples. To verify the ability of the proposed isoSRN model to process other MR data, we collected three other datasets: Sim-H, Sim-P, and Set7, which contain two, four, and seven volumes, respectively. The detailed information of these datasets is shown in Table 1. Note that Sim-H and Sim-P are generated from simulated data from BrainWeb (https://brainweb.bic.mni.mcgill.ca/brainweb/, accessed on 12 August 2022) [54], while Set7 consists of in vivo data acquired using a 3T GE scanner with a T1-3D FSPGR sequence (TR/TE = 5.936 ms/1.956 ms and flip angle [FA] = 9, matrix size = 256 × 256, and field of view [FOV] = 25.6 × 25.6 cm2, slice thickness = 1 mm).
The kernel size of the proposed isoSRN follows the annotation of Figure 1, and we set the number of 3D CB blocks to 16. The number of feature maps is set to 32, which is magnified by a factor of 4 in the wide activation [47]. The learnable weighting factors λ i and π i are initialized to 1.0 before model training. We extract 24   ×   24   ×   24 cubes from LR volumes with their corresponding HR cubes from HR volumes to train the model. Data augmentation is completed by flipping up and down, left and right, and back and forth. We set the batch size to 8 for fast training. The L 1 loss function in Equation (7) is minimized using the Adam optimizer [55] by setting β 1 = 0.9 , β 2 = 0.999 and ϵ = 10 8 . The learning rate is initialized as 2   ×   10 4 and halved at every 10 5 iterations, with 4   ×   10 5 iterations in total.

4.2. Training Example Generation

Given a HR volume y with isotropic resolution, we generate the corresponding LR volumes x v , v s . = 1 , , V , according to the procedure shown in Figure 3. We utilize three orthogonal scans for isotropic MRI reconstruction in this paper, therefore, V = 3 here. Firstly, we downsample the isotropic HR volume y along three orthogonal directions to generate three LR volumes with anisotropic resolution, which simulates three orthogonal scans. For downsampling, we apply spline interpolation to fuse multiple slices into one single slice (weighted average). This simulates the partial volume effect (PVE) that increases as the slice thickness increases.
Subsequently, we upscale these anisotropic volumes to the expected size by the same interpolation. This implements the upsampling operation U v in Equation (5). Lastly, these upsampled LR volumes x v u are fused into a single LR volume x by simple element-wise averaging corresponding to the operation “ M ” in Figure 3 and Equation (5). In this manner, the LR volume x and HR volume y constitute a pair of samples for model training.

4.3. Evaluations of the Proposed Method

In this subsection, we evaluate the proposed isoSRN model on both simulated and clinical MR volumes (T1w and T2w) to verify its effectiveness in terms of various scenarios including scaling factor, noise level, pathology, and the number of orthogonal scans, as well as testing on in vivo MR data. We compare our isoSRN model with two traditional methods and a CNN-based method, namely, CubeAvg [4], NLM [56], and SRCNN-3D [20], respectively, for a comprehensive evaluation.
For a quantitative evaluation, we employ peak signal-to-noise ratio (PSNR), structural similarity index measurement (SSIM) [57], and image sharpness [58] as evaluation metrics. We also utilize the geometric self-ensemble [22] to further boost model performance, which is represented as isoSRN+.

4.3.1. Slice Thickness

In this work, anisotropic LR images are generated by fusing multiple adjacent slices into a single slice in a weighted-average manner. This process is used to simulate the PVE in anisotropic acquisition. Typically, it gets stronger as the slice thickness increases [4], which corresponds to the scaling factor of image SR. To verify the effectiveness of the proposed isoSRN model under different MR image types and scaling factors, we experiment on both T1 and T2 data with six scaling factors: ×2∼×7. This indicates that we are going to reconstruct isotropic HR volumes with voxel size 0.7 mm × 0.7 mm × 0.7 mm from three orthogonal anisotropic scans with slice thicknesses of 1.4 mm, 2.1 mm, 2.8 mm, 3.5 mm, 4.2 mm, and 4.9 mm, respectively.
Table 2 exhibits the quantitative comparison between these methods in terms of PSNR and SSIM. It can be seen that our isoSRN model outperforms other methods by a large margin for all scaling factors. For instance, compared with CubeAvg [4], the PSNR value of isoSRN+ on T1w MR data for SR × 2 is 11.68 dB higher. For all the compared methods, we can observe that PSNR/SSIM values decrease as the scaling factor increases. Nevertheless, for SR × 7 on T2w data, our isoSRN+ still obtains PSNR/SSIM gains of 7.67 dB/0.0851.
Figure 4 displays three orthogonal slices of the reconstructed T1w MR volume in the HCPtest dataset when the slice thickness is 3.5 mm, i.e., SR × 5. For display purposes, the anisotropic scans are resampled to the same size as the HR isotropic volume via spline interpolation. As can be seen, the slices generated by resampling are heavily blurred and many details are lost in the 2nd∼4th columns in Figure 4, due to the PVE. However, our isoSRN and isoSRN+ can produce pleasing visual results with isotropic and high resolution.

4.3.2. Noise Power

In MRI, raw data is intrinsically complex-valued and usually corrupted with zero mean Gaussian noise with equal variance [59]. After inverse Fourier transformation, MR magnitude data has been shown to be Rician distributed [58]. Since Rician noise is commonly found in MR images, it is important to study the influence of noise on model performance. To this end, we simply add (Note that Rician noise is not additive but data-dependent [58], here “add” just means to make the MR magnitude data be Rician-distributed). Rician noise with σ = 0 , 5 , 10 , 15 , 20 , and 25 to anisotropic scans x v , and train the models with noise-free labels. Therefore, the proposed isoSRN model still works in an end-to-end manner without additional denoising steps when dealing with noisy samples.
Table 3 shows the reconstruction accuracy of the compared methods on the simulated dataset Sim-H, which contains two volumes representing T1w and T2w data of a healthy subject in the BrainWeb dataset. It can be observed that the proposed isoSRN (or isoSRN+) model greatly surpasses the traditional methods at all noise levels. For instance, in the case of Rician noise level σ = 5 and slice thickness = 3.5 mm (SR × 5), the isoSRN+ achieves a significant PSNR/SSIM improvement of 10.79 dB/0.093 relative to CubeAvg. Even with a large Rician noise power, e.g., σ = 20 or 25, the superiority of the proposed isoSRN model over other methods is still remarkable. Another observation in Table 3 is that the reconstruction accuracy of all methods decreases as the noise power increases. This is easy to understand because the increased noise power makes image degradation more serious and the reconstruction problem of MRI images more difficult.

4.3.3. Pathology

To study the reconstruction performance of the proposed isoSRN model on pathologic scans, we collected a multiple sclerosis (MS) dataset Sim-P extracted from the BrainWeb dataset, which contains four volumes with normal, mild, moderate, and severe MS, respectively.
The quantitative performance of the compared methods on this dataset is shown in Table 4. Note that these results are obtained by testing the model trained with HCP data directly on the Sim-P dataset. We can observe that our isoSRN+ model still outperforms traditional methods significantly, e.g., for SR × 2 on T1w data, the PSNR of the isoSRN+ is 12.65 dB higher than that of the CubeAvg [4], which is a large increment. Even with a large slice thickness of 4.9 mm, the increment still reaches 10.60 dB. Similar results can also be observed in the T2w data.
Figure 5 shows the reconstruction results of the proposed model over the simulated severe MS T1w volume, accompanied by the anisotropic scans. The red arrows indicate the locations of multiple sclerosis in different views. Similar to other structures in the image, the sclerosis can easily become blurred and obscured in the thick-slice scans, making it difficult to distinguish from other structures. However, in the results using our models, these lesions are satisfactorily recovered with an appearance close to the ground truth, as shown in the last two columns of Figure 5.

4.3.4. The Number of Input Scans

Similar to [4,16], in this section, we also investigate the influence of the number of input scans on the performance of the proposed method. To this end, we train the proposed model with different combinations of three orthogonal scans. When the number of input scans is equal to 1, the input x of the model is equivalent to the simple spline interpolation of the original LR scans. Once the models are well-trained, they are tested on two simulated datasets, Sim-H and Sim-P (T1w), with scaling factor SR × 3. Figure 6a shows the validation curves when training the models with different numbers of input scans. It can be seen that the model converges rapidly and stably in all cases. An obvious observation is that model performance remains basically at the same level when the input scan number is the same (1 or 2), but it is improved significantly when the number of input scans increases. We can also make the same observation about the testing results shown in Figure 6b,c.
It is worth noting that, unlike traditional optimization-based methods (e.g., [4,16], the time efficiency of our method is not affected by the number of inputs due to the end-to-end feature. For an input with a size of 260   ×   311   ×   260 , the running time of the isoSRN model is about 40 s with the support of a single NVIDIA GeForce GTX 1080 Ti GPU. However, the method described in [4] takes more than 10 min to perform a complete reconstruction from three orthogonal scans.

4.4. Comparison with Other Methods

To fully compare the proposed method with other advanced methods, we introduce two other comparative models: ReCNN [41] and VDSR3D [23]. The former is a residual network with 10 Conv+ReLU units that uses a framework similar to ours. The latter is an extension of VDSR [23] based on the framework of the proposed solution. We introduce these methods for comprehensive comparison since they are easy to reproduce and have a similar scale of parameters to our isoSRN.

4.4.1. Evaluation on In Vivo Data

To verify the generalization of our isoSRN to other data, we also present the comparative results on an in vivo dataset, Set7, as shown in Table 5. Although Set7 is derived from imaging hardware, the subjects, parameters, and environment are totally different from the HCP training samples [53], our isoSRN still works well and surpasses the compared methods by a large margin. For instance, the proposed isoSRN and isoSRN+ produce 10.11 dB and 10.23 dB PSNR gains relative to the baseline CubeAvg [4] for SR × 2. Even for large scaling factors, e.g., SR × 7, the performance increment is still up to 7.44 dB and 7.56 dB, respectively. This fully illustrates the good adaptability of the proposed method to in vivo data.
Figure 7 displays the visual comparison between these methods on an in vivo volume from Set7, for SR × 7. We can clearly see the remarkable visual superiority of deep learning methods to traditional methods. Meanwhile, Figure 8 shows the residuals between the results of the compared methods and the ground truth, where we can observe that our proposed models present a better approximation to the ground truth. This conclusion is also demonstrated by the quantitative results annotated below the clipped images in Figure 7.

4.4.2. Running Time

Our solution for isotropic resolution MRI reconstruction is an end-to-end mapping with three orthogonal anisotropic scans as inputs. Therefore, the running time of the entire model involves two parts: fusion of isotropic LR volume x from three anisotropic volumes { x v } v = 1 3 and SR inference of isotropic volumes y . Furthermore, the first part actually corresponds to the baseline CubeAvg [4]. Figure 9 compares the tradeoff of these methods on HCPtest (T1) for three scaling factors. The results are collected with an Omnisky workstation equipped with 64 GB memory, two Intel Xeon E5-2630 CPUs (2.20 GHz), and four NVIDIA GeForce GTX 1080 Ti GPUs. Note that we only consider the running time of SR inference for NLM [56], SRCNN3D [20], ReCNN [41], VDSR3D [23], and our isoSRN.
It can be seen that although our isoSRN runs slightly slower than other deep-learning-based methods, it is noticeably faster than the traditional method NLM [56]. Because our isoSRN performs significantly better than other compared methods (see Table 2, Table 3, Table 4 and Table 5), it provides a better compromise between model performance and running efficiency. Moreover, we can also see that as the input gets larger, the efficiency advantage of end-to-end mapping over traditional methods becomes more obvious. It can be also seen that the running efficiency of SR inference is independent of SR scaling factors.

5. Discussion and Future Work

5.1. Comparative Methods

In the proposed method, we need to acquire three orthogonal LR scans and upsample them with spline interpolation; thus, it is hard to make a fair comparison between our isoSRN model and LRTV [60], which recovers isotropic HR volumes directly from 3D LR volumes. Furthermore, we cannot compare our approach with [4,16] because the authors did not release the source code of their methods. Intuitively, however, the proposed isoSRN model can perform better than [4,16] in that it performs end-to-end mappings with some operations in the pipeline of [4,16] implicitly included in the optimization, e.g., patch extraction and image recovery. In addition, the techniques utilized in [4,16] can be substantially viewed as shallow learning models [20] that have limited representational capacity. Conversely, due to iterative optimization in the implementation, the methods described in [4,16] should be much slower than our model.

5.2. Multiple and Fractional Scales

In real MRI scenes, the ratio of in-plane resolution to that of slice-select direction can be arbitrary, even fractional. How then does a single CNN model deal with this case? In fact, the CNN can be “taught” to do this by simply fusing the corresponding training samples into the training set [9,23]. In the pipeline of our isoSRN model, this is easy to implement because we upsample LR volumes with spline interpolation before feeding them into the network and conduct nonlinear inference in the HR image space. In image SR, performing nonlinear inference in LR feature space helps to improve the training and inferring efficiency, but it is inconvenient in the case of multiple and fractional scaling factors. In this regard, the element-wise average (i.e., “ M ” in Figure 3 and Equation (5)) used to fuse multiple orthogonal scans is more friendly to practical applications due to its simplicity and easy implementation.

5.3. Generalization to Other Data

As described in Section 4.1, the proposed model is trained with 100 HCP samples [53], while tested on four datasets as shown in Table 1. Except in the case of HCPtest [53], the datasets have very different imaging conditions from those of the training data. In Section 4, however, we can observe a consistent improvement in the performance of the proposed isoSRN, which demonstrates its favorable generalization to different types of datasets. This also reveals the great representational capacity of deep models, and the ease of use in practical applications when considering the convenience of automatic feature extraction.

5.4. Extension to Real-World Scenarios

Three orthogonal scans of the subject are required to obtain three anisotropic volumes x v ( v = 1 ,   2 ,   3 ) when deploying our framework in practical applications. The critical issue is how to fuse x v accurately to generate LR volume x with isotropic resolution for SR inference. However, due to the discrepancies in brightness, contrast, and displacement between { x v } v = 1 3 , it is a challenging problem in itself to accurately fuse these LR volumes. An intuitive solution is to calibrate and correct these LR volumes before element-wise averaging but possibly with low accuracy. Another manner may be building a multi-branch network that takes anisotropic LR volumes { x v } v = 1 3 as inputs and generates the fused LR volume x , or directly produces HR volume y . In future work, we will delve into these scenarios and drive the practical deployment of the solution.

6. Conclusions

This paper presents an end-to-end method based on deep 3D CNNs for reconstructing an HR volume with isotropic resolution from multiple anisotropic LR acquisitions. With the effective characterization of structural features in 3D space by deep 3D CNNs and the complementary information provided by orthogonal scans, the proposed isoSRN can surpass traditional methods by a large margin, as shown by our qualitative and quantitative experiments. Moreover, because it works in an end-to-end manner and does not require manual feature extraction, it is more practical and clinically flexible than traditional methods like NLM [56], sparse representation, and dictionary learning [4,16].
In this study, we presented a preliminary study of the application of CNNs in MRI fusion and isotropic reconstruction. Similar methods can be extended to other MRI applications, such as fMRI, dynamic cine MRI, etc., through high-dimensional convolutional models to capture spatio-temporal features, promoting subsequent diagnosis and computer-aided analysis of these MRI data.

Author Contributions

Conceptualization, J.T. and H.Z.; methodology, J.T. and C.X.; software, J.T.; validation, J.T., C.X. and H.Z.; formal analysis, H.Z.; investigation, J.T.; resources, J.T. and C.X.; data curation, J.T.; writing—original draft preparation, J.T.; writing—review and editing, H.Z.; visualization, J.T.; supervision, H.Z.; project administration, C.X.; funding acquisition, J.T., C.X. and H.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported partially by the Chengdu Technological University Research Fund Program under grant No. 2023ZR015 and the Laboratory Open Fund Project under grant No. 2022ZH02806, and in part by the Sichuan Science and Technology Program under grant No. 2022YFG0326.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. He, Z.; He, W.; Wu, J.; Xu, Z. The novel design of a single-sided MRI probe for assessing burn depth. Sensors 2017, 17, 526. [Google Scholar] [CrossRef] [PubMed]
  2. Achar, S.; Hwang, D.; Finkenstaedt, T.; Malis, V.; Bae, W.C. Deep-learning-aided evaluation of spondylolysis imaged with ultrashort echo time magnetic resonance imaging. Sensors 2023, 23, 8001. [Google Scholar] [CrossRef] [PubMed]
  3. Plenge, E.; Poot, D.H.; Bernsen, M.; Kotek, G.; Houston, G.; Wielopolski, P.; van der Weerd, L.; Niessen, W.J.; Meijering, E. Super-Resolut. Methods MRI: Can They Improv. Trade-Off Resolut. Signal- Ratio, Acquis. Time? Magn. Reson. Med. 2012, 68, 1983–1993. [Google Scholar] [CrossRef]
  4. Jia, Y.; Gholipour, A.; He, Z.; Warfield, S.K. A new sparse representation framework for reconstruction of an isotropic high spatial resolution MR volume from orthogonal anisotropic resolution scans. IEEE TMI 2017, 36, 1182–1193. [Google Scholar] [CrossRef] [PubMed]
  5. Aganj, I.; Lenglet, C.; Yacoub, E.; Sapiro, G.; Harel, N. A 3D wavelet fusion approach for the reconstruction of isotropic-resolution MR images from orthogonal anisotropic-resolution scans. Magn. Reson. Med. 2012, 67, 1167–1172. [Google Scholar] [CrossRef]
  6. Mandal, P.K.; Mahto, R.V. Deep multi-branch CNN architecture for early Alzheimer’s detection from brain MRIs. Sensors 2023, 23, 8192. [Google Scholar] [CrossRef]
  7. Remedios, S.W.; Han, S.; Xue, Y.; Carass, A.; Tran, T.D.; Pham, D.L.; Prince, J.L. Deep filter bank regression for super-resolution of anisotropic MR brain images. In Proceedings of the Medical Image Computing and Computer Assisted Intervention—MICCAI 2022, Singapore, 18–22 September 2022; pp. 613–622. [Google Scholar]
  8. Yang, W.; Zhang, X.; Tian, Y.; Wang, W.; Xue, J.H. Deep learning for single image super-resolution: A brief review. arXiv 2018, arXiv:1808.03344. [Google Scholar]
  9. Zhang, K.; Zuo, W.; Zhang, L. Learning a single convolutional super-resolution network for multiple degradations. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 3262–3271. [Google Scholar]
  10. Keys, R.G. Cubic convolution interpolation for digital image processing. IEEE Trans. Acoust. Speech Signal Process. 1981, 29, 1153–1160. [Google Scholar] [CrossRef]
  11. Freedman, G.; Fattal, R. Image and video upscaling from local self-examples. TOG 2011, 30, 12:1–12:11. [Google Scholar] [CrossRef]
  12. Irani, M.; Peleg, S. Image sequence enhancement using multiple motions analysis. In Proceedings of the CVPR, Champaign, IL, USA, 15–18 June 1992; pp. 216–221. [Google Scholar]
  13. Timofte, R.; Rothe, R.; Gool, L.V. Seven ways to improve example-based single image super resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 1865–1873. [Google Scholar]
  14. Yang, J.; Wright, J.; Huang, T.; Ma, Y. Image super-resolution as sparse representation of raw image patches. In Proceedings of the 2008 IEEE Conference on Computer Vision and Pattern Recognition, Anchorage, AK, USA, 23–28 June 2008; pp. 1–8. [Google Scholar]
  15. Yang, J.; Wang, Z.; Lin, Z.; Cohen, S.; Huang, T.S. Coupled dictionary training for image super-resolution. IEEE Trans. Image Process. 2012, 21, 3467–3478. [Google Scholar] [CrossRef]
  16. Bustin, A.; Voilliot, D.; Menini, A.; Felblinger, J.; de Chillou, C.; Burschka, D.; Bonnemains, L.; Odille, F. Isotropic Reconstr. MR Images Using 3D Patch-Based Self-Similarity Learn. IEEE Trans. Med. Imaging 2018, 37, 1932–1942. [Google Scholar] [CrossRef] [PubMed]
  17. Zhao, X.; Zhang, Y.; Zhang, T.; Zou, X. Channel splitting network for single MR image super-resolution. IEEE Trans. Image Process. 2019, 28, 5649–5662. [Google Scholar] [CrossRef]
  18. LeCun, Y.; Bengio, Y.; Hinton, G. Deep Learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  19. Wang, Z.; Chen, J.; Hoi, S.C.H. Deep learning for image super-resolution: A survey. arXiv 2019, arXiv:1902.06068. [Google Scholar]
  20. Dong, C.; Loy, C.C.; He, K.; Tang, X. Image super-resolution using deep convolutional networks. TPAMI 2016, 38, 295–307. [Google Scholar] [CrossRef]
  21. LeCun, Y.; Boser, B.E.; Denker, J.S.; Howard, R.E.; Hubbard, W.; Jackel, L.D. Backpropagation Applied to Handwritten Zip Code Recognition. Neural Comput. 1989, 1, 541–551. [Google Scholar] [CrossRef]
  22. Lim, B.; Son, S.; Kim, H.; Nah, S.; Lee, K.M. Enhanced deep residual networks for single image super-resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 1132–1140. [Google Scholar]
  23. Kim, J.; Kwon Lee, J.; Mu Lee, K. Accurate image super-resolution using very deep convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 1646–1654. [Google Scholar]
  24. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  25. He, K.; Zhang, X.; Ren, S.; Sun, J. Identity mappings in deep residual networks. In Proceedings of the ECCV, Amsterdam, The Netherlands, 11–14 October 2016; pp. 630–645. [Google Scholar]
  26. Dong, C.; Loy, C.C.; Tang, X. Accelerating the super-resolution convolutional neural network. In Proceedings of the ECCV, Amsterdam, The Netherlands, 11–14 October 2016; pp. 391–407. [Google Scholar]
  27. Shi, W.; Caballero, J.; Huszár, F.; Totz, J.; Aitken, A.P.; Bishop, R.; Rueckert, D.; Wang, Z. Real-time single image and video super-resolution using an efficient sub-pixel convolutional neural network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 1874–1883. [Google Scholar]
  28. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. In Proceedings of the Annual Conference on Neural Information Processing Systems 2014, Montreal, QC, Canada, 8–13 December 2014; pp. 2672–2680. [Google Scholar]
  29. Ledig, C.; Theis, L.; Huszár, F.; Caballero, J.; Cunningham, A.; Acosta, A.; Aitken, A.; Tejani, A.; Totz, J.; Wang, Z.; et al. Photo-realistic single image super-resolution using a generative adversarial network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 105–114. [Google Scholar]
  30. Zhao, X.; Liao, Y.; Li, Y.; Zhang, T.; Zou, X. FC2N: Fully channel-concatenated network for single image super-resolution. arXiv 2019, arXiv:1907.03221. [Google Scholar]
  31. Dai, T.; Cai, J.; Zhang, Y.; Xia, S.; Zhang, L. Second-order attention network for single image super-resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 11065–11074. [Google Scholar]
  32. Liang, J.; Cao, J.; Sun, G.; Zhang, K.; Van Gool, L.; Timofte, R. SwinIR: Image restoration using swin transformer. In Proceedings of the CVPR Workshop, Virtual, 25 June 2021; pp. 1833–1844. [Google Scholar]
  33. Zhou, L.; Cai, H.; Gu, J.; Li, Z.; Liu, Y.; Chen, X.; Qiao, Y.; Dong, C. Efficient image super-resolution using vast-receptive-field attention. In Proceedings of the ECCV Workshops, New Orleans, LA, USA, 19–20 June 2022; pp. 256–272. [Google Scholar]
  34. Zhou, Y.; Li, Z.; Guo, C.L.; Bai, S.; Cheng, M.M.; Hou, Q. Srformer: Permuted self-attention for single image super-resolution. In Proceedings of the 2023 International Conference on Computer Vision, Paris, France, 2–6 October 2023; pp. 12780–12791. [Google Scholar]
  35. Ullah, F.; Ansari, S.U.; Hanif, M.; Ayari, M.A.; Chowdhury, M.E.H.; Khandakar, A.A.; Khan, M.S. Brain MR image enhancement for tumor segmentation using 3D U-Net. Sensors 2021, 21, 7528. [Google Scholar] [CrossRef] [PubMed]
  36. Duong, S.T.; Phung, S.L.; Bouzerdoum, A.; Ang, S.P.; Schira, M.M. Correcting susceptibility artifacts of MRI sensors in brain scanning: A 3D anatomy-guided deep learning approach. Sensors 2021, 21, 2314. [Google Scholar] [CrossRef]
  37. Zhang, Y.; Li, K.; Li, K.; Fu, Y. MR image super-resolution with squeeze and excitation reasoning attention network. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2021), Nashville, TN, USA, 20–25 June 2021; pp. 13425–13434. [Google Scholar]
  38. Mao, Y.; Jiang, L.; Chen, X.; Li, C. Disc-diff: Disentangled conditional diffusion model for multi-contrast mri super-resolution. In Proceedings of the Medical Image Computing and Computer Assisted Intervention—MICCAI 2023, Vancouver, BC, Canada, 8–12 October 2023; Springer: Cham, Switzerland, 2023; pp. 387–397. [Google Scholar]
  39. Zhang, J.; Chi, Y.; Lyu, J.; Yang, W.; Tian, Y. Dual arbitrary scale super-resolution for multi-contrast MRI. In Proceedings of the Medical Image Computing and Computer Assisted Intervention—MICCAI 2023, Vancouver, BC, Canada, 8–12 October 2023; pp. 282–292. [Google Scholar]
  40. Li, G.; Lv, J.; Tian, Y.; Dou, Q.; Wang, C.; Xu, C.; Qin, J. Transformer-empowered multi-scale contextual matching and aggregation for multi-contrast MRI super-resolution. In Proceedings of the 2022 IEEE CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 20636–20645. [Google Scholar]
  41. Pham, C.H.; Tor-Díeza, C.; Meunierb, H.; Bednarek, N.; Fablet, R.; Passat, N.; Rousseau, F. Multiscale Brain MRI Super-Resolut. Using Deep 3D Convolutional Networks. Comput. Med Imaging Graph. 2019, 77, 101647. [Google Scholar] [CrossRef]
  42. Niu, B.; Wen, W.; Ren, W.; Zhang, X.; Yang, L.; Wang, S.; Zhang, K.; Cao, X.; Shen, H. Single image super-resolution via a holistic attention network. In Proceedings of the ECCV, Glasgow, UK, 23–28 August 2020; Volume 12357, pp. 191–207. [Google Scholar]
  43. Mei, Y.; Fan, Y.; Zhou, Y. Image super-resolution with non-local sparse attention. In Proceedings of the 2021 IEEE CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 3517–3526. [Google Scholar]
  44. Huang, P.; Li, C.; He, P.; Xiao, H.; Ping, Y.; Feng, P.; Tian, S.; Chen, H.; Mercaldo, F.; Santone, A.; et al. MamlFormer: Priori-experience guiding transformer network via manifold adversarial multi-modal learning for laryngeal histopathological grading. Inf. Fusion 2024, 108, 102333. [Google Scholar] [CrossRef]
  45. Huang, P.; Xiao, H.; He, P.; Li, C.; Guo, X.; Tian, S.; Feng, P.; Chen, H.; Sun, Y.; Mercaldo, F.; et al. LA-ViT: A Network with Transformers Constrained by Learned-Parameter-Free Attention for Interpretable Grading in a New Laryngeal Histopathology Image Dataset. IEEE J. Biomed. Health Inform. 2024, 28, 3557–3570. [Google Scholar] [CrossRef]
  46. Pan, H.; Peng, H.; Xing, Y.; Jiayang, L.; Hualiang, X.; Sukun, T.; Peng, F. Breast tumor grading network based on adaptive fusion and microscopic imaging. Opto-Electron. Eng. 2023, 50, 220158. [Google Scholar]
  47. Yu, J.; Fan, Y.; Yang, J.; Xu, N.; Wang, Z.; Wang, X.; Huang, T. Wide activation for efficient and accurate image super-resolution. arXiv 2018, arXiv:1808.08718. [Google Scholar]
  48. Oktay, O.; Bai, W.; Lee, M.; Guerrero, R.; Kamnitsas, K.; Caballero, J.; de Marvao, A.; Cook, S.; O’Regan, D.; Rueckert, D. Multi-input cardiac image super-resolution using convolutional neural networks. In Proceedings of the 19th International Conference on Medical Image Computing and Computer-Assisted Intervention, MICCAI 2016, Athens, Greece, 17–21 October 2016; pp. 246–254. [Google Scholar]
  49. Chen, X.; Wang, X.; Zhou, J.; Qiao, Y.; Dong, C. Activating more pixels in image super-resolution transformer. In Proceedings of the 2023 Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 22367–22377. [Google Scholar]
  50. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the 2018 Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  51. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  52. Han, D.; Pan, X.; Han, Y.; Song, S.; Huang, G. Flatten Transformer: Vision Transformer using focused linear attention. In Proceedings of the 2023 International Conference on Computer Vision, Paris, France, 2–6 October 2023; pp. 5961–5971. [Google Scholar]
  53. Glasser, M.F.; Sotiropoulos, S.N.; Wilson, J.A.; Coalson, T.S.; Fischl, B.; Andersson, J.L.; Xu, J.; Jbabdi, S.; Webster, M.; Polimeni, J.R.; et al. The minimal preprocessing pipelines for the Human Connectome Project. NeuroImage 2013, 80, 105–124. [Google Scholar] [CrossRef] [PubMed]
  54. Cocosco, C.A.; Kollokian, V.; Kwan, R.K.S.; Pike, G.B.; Evans, A.C. Brainweb: Online interface to a 3D MRI simulated brain database. NeuroImage 1997, 5, S425. [Google Scholar]
  55. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. In Proceedings of the ICLR, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  56. Manjón, J.V.; Coupé, P.; Buades, A.; Fonov, V.; Collins, D.L. Non-local MRI upsampling. Med. Image Anal. 2010, 14, 784–792. [Google Scholar] [CrossRef]
  57. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed]
  58. He, L.; Greenshields, I.R. A nonlocal maximum likelihood estimation method for rician noise reduction in MR images. IEEE Trans. Med. Imaging 2009, 28, 165–172. [Google Scholar] [PubMed]
  59. Gudbjartsson, H.; Patz, S. The Rician distribution of noisy MRI data. Magn. Reson. Med. 2010, 34, 910–914. [Google Scholar] [CrossRef] [PubMed]
  60. Shi, F.; Cheng, J.; Wang, L.; Yap, P.T.; Shen, D. LRTV: MR image super-resolution with low-rank and total variation regularizations. IEEE Trans. Med Imaging 2015, 34, 2459–2466. [Google Scholar] [CrossRef] [PubMed]
Figure 1. The overall network structure of the proposed 3D CNN model for isotropic MRI reconstruction. The feature extraction contains only one 3D conv layer and the image recovery part consists of two 3D conv layers. The nonlinear inference part is composed of several 3D concat blocks (3D-CBs), each of which is built with two 3D conv layers with one ReLU layer in the middle. Note that our 3D-CB is different from the residual block in [22].
Figure 1. The overall network structure of the proposed 3D CNN model for isotropic MRI reconstruction. The feature extraction contains only one 3D conv layer and the image recovery part consists of two 3D conv layers. The nonlinear inference part is composed of several 3D concat blocks (3D-CBs), each of which is built with two 3D conv layers with one ReLU layer in the middle. Note that our 3D-CB is different from the residual block in [22].
Sensors 24 06639 g001
Figure 2. For isotropic MRI restoration, convolutional operations covering more dimensions can extract richer and more useful information. For instance, the information for inferring the feature map of the next layer, in the 1D case, is only from the dimension that requires to be scaled. However, in the 3D case (our work), the information in a 3D cube can be used to compute the feature map of the next layer.
Figure 2. For isotropic MRI restoration, convolutional operations covering more dimensions can extract richer and more useful information. For instance, the information for inferring the feature map of the next layer, in the 1D case, is only from the dimension that requires to be scaled. However, in the 3D case (our work), the information in a 3D cube can be used to compute the feature map of the next layer.
Sensors 24 06639 g002
Figure 3. Sample generation. y represents a 3D HR volume with isotropic resolution and x denotes the corresponding LR volume with isotropic resolution. M is an operation of element-wise averaging. [ d ] r and [ d ] r stand for r-fold downsampling and upsampling along the dimension d , which could be A (Axial), S (Sagittal), or C (Coronal).
Figure 3. Sample generation. y represents a 3D HR volume with isotropic resolution and x denotes the corresponding LR volume with isotropic resolution. M is an operation of element-wise averaging. [ d ] r and [ d ] r stand for r-fold downsampling and upsampling along the dimension d , which could be A (Axial), S (Sagittal), or C (Coronal).
Sensors 24 06639 g003
Figure 4. The visualization of the proposed isoSRN on a T1 volume in HCPtest (SR × 5, slice thickness = 3.5 mm). Top to bottom: Axial, Coronal, and Sagittal views. The 2nd to 4th columns are resampled scans through spline interpolation. The best result is marked in red, and the second-best is marked in blue.
Figure 4. The visualization of the proposed isoSRN on a T1 volume in HCPtest (SR × 5, slice thickness = 3.5 mm). Top to bottom: Axial, Coronal, and Sagittal views. The 2nd to 4th columns are resampled scans through spline interpolation. The best result is marked in red, and the second-best is marked in blue.
Sensors 24 06639 g004
Figure 5. Top to bottom: Axial, Coronal, and Sagittal views of a simulated multiple sclerosis T1 volume in Sim-P (SR × 4, slice thickness = 2.8 mm). The best result is marked in red, and the second-best is marked in blue.
Figure 5. Top to bottom: Axial, Coronal, and Sagittal views of a simulated multiple sclerosis T1 volume in Sim-P (SR × 4, slice thickness = 2.8 mm). The best result is marked in red, and the second-best is marked in blue.
Sensors 24 06639 g005
Figure 6. The influence of the number of input anisotropic LR scans (T1, SR × 3). When there is only one input scan, e.g., A, C, or S, CubeAvg is equivalent to simple spline interpolation. A: Axial scan; C: Coronal scan; S: Sagittal scan.
Figure 6. The influence of the number of input anisotropic LR scans (T1, SR × 3). When there is only one input scan, e.g., A, C, or S, CubeAvg is equivalent to simple spline interpolation. A: Axial scan; C: Coronal scan; S: Sagittal scan.
Sensors 24 06639 g006
Figure 7. Visual comparison between the compared methods on the in vivo dataset Set7 for SR × 7. The best result is marked in red, and the second-best is marked in blue. Top to bottom: Sagittal, Axial, and Coronal views.
Figure 7. Visual comparison between the compared methods on the in vivo dataset Set7 for SR × 7. The best result is marked in red, and the second-best is marked in blue. Top to bottom: Sagittal, Axial, and Coronal views.
Sensors 24 06639 g007
Figure 8. Visualization of the residuals between reconstruction results of different methods and the ground truth, corresponding to Figure 7. Higher pixel intensity indicates a larger difference between the reconstruction and ground truth (SR × 7). Top to bottom: Sagittal, Axial, and Coronal views.
Figure 8. Visualization of the residuals between reconstruction results of different methods and the ground truth, corresponding to Figure 7. Higher pixel intensity indicates a larger difference between the reconstruction and ground truth (SR × 7). Top to bottom: Sagittal, Axial, and Coronal views.
Sensors 24 06639 g008
Figure 9. Inference time versus model performance on HCPtest dataset (T1). Note that except CubeAvg [4], we only collect the inference time for the mapping from x to y and exclude the time for fusing x v ( v = 1 ,   2 ,   3 ) to x .
Figure 9. Inference time versus model performance on HCPtest dataset (T1). Note that except CubeAvg [4], we only collect the inference time for the mapping from x to y and exclude the time for fusing x v ( v = 1 ,   2 ,   3 ) to x .
Sensors 24 06639 g009
Table 1. Details of the testing datasets used in this work. Note that these data are collected with isotropic resolution. # denotes the volume number of a dataset.
Table 1. Details of the testing datasets used in this work. Note that these data are collected with isotropic resolution. # denotes the volume number of a dataset.
DatasetsModeDims# VolumesVoxel SizeSource
HCPtest [53]T1/T2260 × 311 × 260500.7 mmHCP
Sim-H [54]T1/T2217 × 181 × 18121.0 mmBrainweb
Sim-P [54]T1/T2217 × 181 × 18141.0 mmBrainweb
Set7 (Collected)T1256 × 256 × 15471.0 mmin vivo
Table 2. Performance comparison on HCPtest dataset in terms of different scaling factors (×2∼×7). The largest values are marked in red, and the second-largest are marked in blue (PSNR (dB)/SSIM). Both T1- and T2-weighted MR volumes are included here.
Table 2. Performance comparison on HCPtest dataset in terms of different scaling factors (×2∼×7). The largest values are marked in red, and the second-largest are marked in blue (PSNR (dB)/SSIM). Both T1- and T2-weighted MR volumes are included here.
MethodsTypeSR × 2SR × 3SR × 4SR × 5SR × 6SR × 7
CubeAvg [4]T145.61/0.993740.19/0.980337.45/0.964835.66/0.948234.22/0.929233.29/0.9130
NLM [56]46.44/0.994940.95/0.983438.42/0.971636.35/0.955934.85/0.932933.72/0.9220
SRCNN3D [20]53.10/0.998645.66/0.992842.20/0.985039.87/0.975438.14/0.964636.86/0.9538
isoSRN [Ours]57.19/0.999349.35/0.996546.23/0.993344.33/0.990142.84/0.986641.78/0.9834
isoSRN+ [Ours]57.29/0.999349.46/0.996646.34/0.993444.44/0.990342.95/0.986941.92/0.9838
CubeAvg [4]T241.01/0.992035.58/0.973633.10/0.954331.53/0.934530.25/0.912129.41/0.8932
NLM [56]41.77/0.993536.36/0.978433.99/0.963232.11/0.943730.77/0.923529.77/0.9036
SRCNN3D [20]49.29/0.998540.91/0.990937.23/0.980034.86/0.966233.12/0.950731.94/0.9360
isoSRN [Ours]55.02/0.999445.51/0.996141.91/0.991939.77/0.987538.13/0.982536.91/0.9776
isoSRN+ [Ours]55.19/0.999445.66/0.996242.05/0.992139.92/0.987838.28/0.982937.08/0.9783
Table 3. Performance comparison on the Sim-H dataset in terms of Rician noise power ( σ = 0 25 ). The largest values are marked in red, and the second-largest are marked in blue (PSNR (dB)/SSIM).
Table 3. Performance comparison on the Sim-H dataset in terms of Rician noise power ( σ = 0 25 ). The largest values are marked in red, and the second-largest are marked in blue (PSNR (dB)/SSIM).
Rician NoiseScale σ = 0 σ = 5 σ = 10 σ = 15 σ = 20 σ = 25
CubeAvg [4]SR × 526.50/0.920626.45/0.897926.30/0.854626.07/0.818425.77/0.790225.43/0.7668
NLM [56]27.21/0.935227.17/0.914227.05/0.873926.86/0.841326.62/0.816626.33/0.7966
isoSRN [Ours]37.80/0.992336.86/0.989936.18/0.988235.67/0.985635.21/0.982834.81/0.9810
isoSRN+ [Ours]38.02/0.992937.24/0.990936.65/0.989436.15/0.987335.70/0.984135.29/0.9829
CubeAvg [4]SR × 723.68/0.850023.59/0.829423.52/0.789923.41/0.756723.26/0.730723.08/0.7089
NLM [56]24.07/0.866824.05/0.847724.00/0.812723.92/0.783923.81/0.761523.67/0.7434
isoSRN [Ours]33.43/0.978932.75/0.976032.34/0.972932.05/0.968831.81/0.966631.51/0.9627
isoSRN+ [Ours]33.68/0.980933.24/0.978832.83/0.975532.60/0.972232.33/0.970132.05/0.9662
Table 4. Quantitative performance of the compared methods on the simulated pathologic dataset Sim-P, which contains four MR volumes with four different degrees of multiple sclerosis and two image types T1w and T2w. The largest values are marked in red, and the second-largest are marked in blue (PSNR (dB)/SSIM).
Table 4. Quantitative performance of the compared methods on the simulated pathologic dataset Sim-P, which contains four MR volumes with four different degrees of multiple sclerosis and two image types T1w and T2w. The largest values are marked in red, and the second-largest are marked in blue (PSNR (dB)/SSIM).
MethodsTypeSR × 2SR × 3SR × 4SR × 5SR × 6SR × 7
CubeAvg [4]T144.10/0.997737.46/0.990133.44/0.976030.65/0.955728.75/0.932927.17/0.9043
NLM [56]44.39/0.998038.04/0.992134.32/0.981631.52/0.964229.55/0.944427.73/0.9151
SRCNN3D [20]51.23/0.999543.47/0.997038.31/0.990335.11/0.979932.71/0.965030.56/0.9421
isoSRN [Ours]56.58/0.999848.93/0.999044.79/0.997741.91/0.995339.71/0.992137.47/0.9868
isoSRN+ [Ours]56.75/0.999849.14/0.999145.00/0.997842.12/0.995639.96/0.992537.77/0.9878
CubeAvg [4]T244.10/0.997737.46/0.990133.44/0.976030.65/0.955728.75/0.932927.17/0.9043
NLM [56]44.39/0.998038.04/0.992134.32/0.981631.52/0.964229.55/0.944427.73/0.9151
SRCNN3D [20]50.97/0.999543.11/0.996938.24/0.990734.96/0.978932.67/0.963830.39/0.9407
isoSRN [Ours]55.98/0.999848.54/0.999144.33/0.997541.49/0.995039.31/0.991637.01/0.9858
isoSRN+ [Ours]56.18/0.999848.75/0.999144.56/0.997641.71/0.995339.57/0.992037.31/0.9867
Table 5. Performance comparison on Set7 dataset in terms of different scaling factors (×2∼×7). The best values are marked in red, and the second-best are marked in blue (PSNR (dB)/SSIM).
Table 5. Performance comparison on Set7 dataset in terms of different scaling factors (×2∼×7). The best values are marked in red, and the second-best are marked in blue (PSNR (dB)/SSIM).
MethodsTypeSR × 2SR × 3SR × 4SR × 5SR × 6SR × 7
CubeAvg [4]T143.12/0.992738.23/0.979735.75/0.965634.07/0.949932.84/0.933432.07/0.9200
NLM [56]44.62/0.994639.03/0.982536.65/0.971634.63/0.955933.34/0.941432.29/0.9266
SRCNN3D [20]49.58/0.998042.96/0.991339.75/0.982837.55/0.972035.93/0.959934.79/0.9481
ReCNN [41]52.46/0.998846.18/0.995343.22/0.991341.18/0.986639.59/0.980938.46/0.9752
VDSR3D [23]52.42/0.998846.38/0.995543.50/0.991741.52/0.987539.95/0.982238.88/0.9773
isoSRN [Ours]53.23/0.999046.76/0.995843.90/0.992442.04/0.988740.56/0.984339.51/0.9802
isoSRN+ [Ours]53.35/0.999146.87/0.996044.02/0.992642.16/0.989040.68/0.984739.63/0.9806
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Tian, J.; Xiao, C.; Zhu, H. Isotropic Brain MRI Reconstruction from Orthogonal Scans Using 3D Convolutional Neural Network. Sensors 2024, 24, 6639. https://doi.org/10.3390/s24206639

AMA Style

Tian J, Xiao C, Zhu H. Isotropic Brain MRI Reconstruction from Orthogonal Scans Using 3D Convolutional Neural Network. Sensors. 2024; 24(20):6639. https://doi.org/10.3390/s24206639

Chicago/Turabian Style

Tian, Jinsha, Canjun Xiao, and Hongjin Zhu. 2024. "Isotropic Brain MRI Reconstruction from Orthogonal Scans Using 3D Convolutional Neural Network" Sensors 24, no. 20: 6639. https://doi.org/10.3390/s24206639

APA Style

Tian, J., Xiao, C., & Zhu, H. (2024). Isotropic Brain MRI Reconstruction from Orthogonal Scans Using 3D Convolutional Neural Network. Sensors, 24(20), 6639. https://doi.org/10.3390/s24206639

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop