Next Article in Journal
Investigating the Impacts of Autonomous Vehicles on the Efficiency of Road Network and Traffic Demand: A Case Study of Qingdao, China
Previous Article in Journal
Grape Maturity Estimation Using Time-of-Flight and LiDAR Depth Cameras
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Complex Residual Attention U-Net for Fast Ultrasound Imaging from a Single Plane-Wave Equivalent to Diverging Wave Imaging

by
Ahmed Bentaleb
*,
Christophe Sintes
,
Pierre-Henri Conze
,
François Rousseau
,
Aziliz Guezou-Philippe
and
Chafiaa Hamitouche
Département Image et Traitement de l’Information, Institue Mines-Télécom (IMT) Atlantique, 29200 Brest, France
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(16), 5111; https://doi.org/10.3390/s24165111
Submission received: 6 June 2024 / Revised: 12 July 2024 / Accepted: 5 August 2024 / Published: 7 August 2024
(This article belongs to the Section Sensing and Imaging)

Abstract

:
Plane wave imaging persists as a focal point of research due to its high frame rate and low complexity. However, in spite of these advantages, its performance can be compromised by several factors such as noise, speckle, and artifacts that affect the image quality and resolution. In this paper, we propose an attention-based complex convolutional residual U-Net to reconstruct improved in-phase/quadrature complex data from a single insonification acquisition that matches diverging wave imaging. Our approach introduces an attention mechanism to the complex domain in conjunction with complex convolution to incorporate phase information and improve the image quality matching images obtained using coherent compounding imaging. To validate the effectiveness of this method, we trained our network on a simulated phased array dataset and evaluated it using in vitro and in vivo data. The experimental results show that our approach improved the ultrasound image quality by focusing the network’s attention on critical aspects of the complex data to identify and separate different regions of interest from background noise.

1. Introduction

Plane wave (PW) ultrasound (US) imaging [1] supports a high frame rate by transmitting a single unfocused beam that insonifies the field of view. This technique provides a high temporal resolution but limited lateral resolution and contrast. Multi-transmission of steered beams or diverging waves (DWs) [2] has successfully enhanced PW imaging; however, this technique is constrained by the number of transmitted waves, which directly affect the frame rate. This restricts its application in fast-moving structures such as the heart, and as such, PW imaging is the preferred method.
The transmitted waves (PWs or DWs) propagate through the target region and encounter impedance mismatches, causing an echo to reflect back to the transducer. In US image reconstruction, the reflected echo is received as a multichannel raw radio frequency (RF) signal. RF data are demodulated to produce in-phase/quadrature (IQ) data, followed by beamforming of the imaged target. Beamforming refers to the application of a time-of-flight correction process with spatial filtering to introduce selectivity into the signal, which eliminates undesirable interference [3].
Delay and sum (DAS) [4] is the most prevalent beamforming technique in medical US due to its low complexity and high frame rate. PW image beamforming with DAS results in a high frame rate but low image quality, whereas DW images are reconstructed from coherently compounding consecutive beamformed echos from each steered wave. Coherent compounding produces higher-quality images than PW imaging, but at the expense of a lower frame rate and a high computation cost related to the number of transmitted waves.
Deep learning has recently emerged as a critical component in the field of medical data analysis. Significant progress has been made in tasks such as image classification [5], segmentation [6,7], and liver and breast lesion classification [8,9], prompting US medical researchers to apply deep learning methods to their own work. Deep neural networks (DNNs) have been used for RF data interpolation [10] and reconstruction of B-mode images [11]. However, a more promising approach is the use of convolutional neural networks (CNNs), which been successfully applied to image processing-related tasks. For instance, CNNs have been used for compounding imaging [12], denoising, and speckle reduction [13], in addition to fully convolutional neural networks (FCNNs), which learned a minimum variance beamformer transformation [14] and direct image segmentation from RF data [15].
Most of the advancements achieved in the context of deep learning are based on real-valued data, whereas researchers have only recently begun to exploit the usability of complex-valued data. Complex-valued neural networks (CVNNs) were used in comparison with real-valued neural networks (RVNNs) [16] to identify the usability of such data and the learning of a complex representation of time series [17].
In US imaging, complex data comprise the raw demodulated IQ data from the RF signal and the complex DAS IQ data. As both contain phase and magnitude information, the latter are almost real (in other words, the module produces the US image); indeed, such data remain complex due to misknowledge with regard to the speed at which sound travels in human tissue.
A CNN was applied to complex data for speckle reduction [13] and improved multiline acquisition and transmission [18], where the complex data were trained as separate real and imaginary parts in a two-branch network structure. Consequently, such work failed to consider the nature of complex data. As demonstrated in [16], a complex-valued model provides a more constrained system than a model based on real-valued parameters. A recent study demonstrated the superior representational capacity of CVNNs in acoustic applications, such as speech spectrum prediction and music transcription [19]. A different approach was considered by [20,21]; in their work, a complex-valued CNN (CVCNN) was defined for fast US image reconstruction from DAS IQ data. The CVCNN consisted of a convolution between the complex data and complex weights represented as real and imaginary parts.
Current approaches to representing and processing IQ data in US imaging have used CNNs with complex convolutions without incorporating attention mechanisms. While an attention mechanism was applied to image despeckling in [22], it was limited to B-mode images and did not incorporate IQ data or complex data representations. To take advantage of the previously cited techniques for US image reconstruction, we propose a novel approach that combines the attention mechanism [23] with complex convolution based on a U-Net network [6] for US image reconstruction from a single PW acquisition. The reduction in transmitted waves will improve the frame rate, reduce the hardware complexity, and eventually improve the quality and resolution.
The aim of this work is to demonstrate the potential of CVNNs for IQ-based data in US image reconstruction tasks compared to image-based models using a simulated training dataset. While we will perform inference using in vivo samples, real-world data are different from simulations; therefore, the model requires further training and tuning using clinical data to be tested on real-world scenarios. In this work, we will address the following:
(1)
A complex residual attention U-Net network (C-Res-Att-UNet) for US image reconstruction from a single PW matching DW imaging quality. This innovative framework leverages the phase information present in complex IQ data to enhance the representation, ultimately reconstructing higher-quality US images.
(2)
A custom concatenation layer that takes into account complex data representation and a 2D max-pooling layer dedicated to down-sampling complex-valued data based on the indices of the maximum amplitude of the complex tensor.
(3)
A complex up-sampling technique that learns an up-sampling transformation based on sub-pixel convolutions [24] rather than interpolating the real and imaginary parts separately.
(4)
An equivalent network that is trained using a real-valued CNN named Res-Att-UNet, which uses B-mode image data.
(5)
Evaluation of C-Res-Att-UNet and Res-Att-UNet using a test set from the simulated dataset, in addition to samples from the PICMUS dataset [25] containing in vitro phantom, simulation, and in vivo carotid data.
The present study is structured as follows: Section 2 describes the custom-built layers, the proposed architecture, and the training strategy. Section 3 presents the experimental setup for the data acquisition, network training, and performance metrics. Section 4 presents the results of the work. The discussion is presented in Section 5. Finally, the conclusion is outlined in Section 6.

2. Materials and Methods

2.1. Complex Convolution

This work focuses on incorporating the phase information present in complex IQ data into the attention mechanism to facilitate more effective learning of ultrasound data. We start by defining a complex data frame, X = X r + j X i (where X r = Re ( X ) and X i = Im ( X ) are the real and imaginary components of X, respectively), and in the same manner, we describe a complex weight, W = W r + j W i . According to the definition provided in [19], the complex convolution Z of X with W is defined as follows:
Z = ( X r + j X i ) ( W r + j W i ) = X r W r X i W i + j ( X r W i + X i W r ) ,
where ( ) represents the convolution operator. Rearranging and rewriting (1) (with X as the input and Z as the output) in a matrix format, we obtain the following:
Re ( Z ) Im ( Z ) = W r W i W i W r X r X i .
In contrast to a purely real convolution in which the real and imaginary parts are considered to be part of a two-branch structure, in this paper, we construct a mathematical connection between the real and imaginary parts of the data frame and the convolution weights.
In order to update the weights after the forward pass, the condition for the back-propagation must be valid; i.e., the activation and loss functions must be differentiable. Sarroff et al. [17] stated that these functions must be holomorphic (complex differentiable functions that satisfy the Cauchy–Riemann conditions), whereas the authors of [16,19] showed that CVNNs can be optimized with real-valued activation and loss functions.
ReLU-based activation functions are commonly implemented in deep learning. We employ a complex ReLU ( C R e L U ) activation function [19]. This choice is aligned with the differentiability condition described in [16]. ( C R e L U ) is applied to the real and imaginary parts separately:
C R e L U ( Z ) = R e L U ( Re ( Z ) ) + j R e L U ( Im ( Z ) ) .
Following [16,19], the network is evaluated at the end of each epoch using a positive real-valued loss function consisting of a mean squared error between the desired output image Y and network-reconstructed image Y ^ :
L = 1 N N ( Y Y ^ ) 2 .

2.2. Complex Concatenation Layer

A regular concatenation layer works by stacking the inputs one by one. This method cannot be applied to complex data, as the stacking would mix the real and imaginary parts. To overcome this problem, we implemented a complex concatenation layer ( C Concat) that stacks the real components of the inputs followed by the imaginary components to achieve a correct stacking order, producing an output that is consistent with the definition of the complex convolution, (2). C Concat is defined as follows:
Re ( Z o u t p u t ) = Concat [ Re ( Z 1 ) , Re ( Z 2 ) ] Im ( Z o u t p u t ) = Concat [ Im ( Z 1 ) , Im ( Z 2 ) ] ,
where Z o u t p u t is the concatenated complex output from the complex inputs Z 1 , and Z 2 .

2.3. Complex Max-Pooling Layer

Usually, a pooling layer is used to reduce the spatial dimension of an input tensor across its features. The input is split into patches, and each patch is replaced by one value, which is defined by the pooling function. In max-pooling, the function is defined as the maximum value in the patch. Unfortunately, this could not be projected to the complex domain, where, in this work, the input tensor was defined as having real and imaginary components. The regular max-pooling layer’s output misplaces both real and imaginary components, resulting in the loss of the complex numbers (placing a real part in the same pool as a different imaginary part, or vice versa) and a consequent loss of phase information.
To overcome these issues, we suggest a complex max-pooling layer ( C Max-pooling), presented in Figure 1. C Max-pooling is based on the maximum magnitude of the complex number in a patch. The C Max-pooling function is defined as follows:
Z o u t p u t = Z a r g m a x   | Z | ,
where a r g m a x   | Z | is the index of the maximum magnitude used to extract the real and corresponding imaginary components. In this manner, the phase information is preserved by matching the correct components of the same complex number.

2.4. Complex Up-Sampling Layer

The up-sampling layer, a ubiquitous component in many neural network architectures, typically employs interpolation techniques to resize input data to the desired dimensions. This technique might be successful if we resize the real and imaginary parts, separately. However, it fails to improve the reconstruction of the complex ultrasound signals. Our objective is to refine the reconstruction process at each step, which necessitates a more sophisticated approach. The proposed complex up-sampling layer ( C Up-sampling) (Figure 2) is responsible for learning an up-sampling transformation based on periodic shuffling of the complex features from the complex convolution rather than relying on interpolation. This layer produces an up-sampled output that is not interpolated but rather learned from the rearranged complex features.

2.5. Proposed Network

The proposed network (Figure 3) contains two parts: an encoder and a decoder. The first part is composed of down-sampling blocks containing complex convolution layers, with batch normalization followed by C R e L U activation (blue arrow). The down-sampling is performed using the custom C Max-pooling layer, where the tensor input is reduced using the index of the maximum magnitude across the max-pooling window to extract the corresponding real and imaginary parts. Prior to the implementation of the C Max-pooling layer (red arrow), the output is used as a skip connection (purple arrow). At the bottom of the encoder, a residual complex convolutional block is used in place of the traditional fully convolutional block.
The decoder part of the network is the area in which the HQ frame x ^ is reconstructed based on information fed from the encoder. To reinforce the sensitivity and restoration accuracy, we employ an attention mechanism [23] in combination with the previously defined complex convolution (1), resulting in a complex attention gate (CAG) that combines the skip connection (purple arrow) from the encoder with information from the decoder C Up-sampling layers (cyan arrow). The skip connection provides more spatial information along with poor feature representation, while the attention gives more weight to the features of interest.
The features of CAG are defined as follows:
C A G f e a t u r e s = W S G c [ C R e L U ( S c W S + G c W G ) ] .
Meanwhile, the output of CAG is obtained as shown below:
C A G o u t p u t = W C A G c ( S × C A G f e a t u r e s ) ,
where ( × ) represents a complex multiplication of S, a complex skip connection with C A G f e a t u r e s . G is the complex gating signal. W ( . ) is the complex kernel at each step, while ( c ) represents the complex convolution defined in (2) between W ( . ) and different complex parameters along the CAG.
Instead of element-wise multiplication of the attention features with the skip connection signal, as defined in [23], a complex multiplication (8) is chosen, considering the nature of the presented complex data, followed by a complex convolution to enhance the output of the CAG and improve the overall learning process. The decoder is used symmetrically using up-sampling blocks with a C Up-sampling layer followed by a C Concat from the output of the CAG. It finishes with a complex convolution, leading to our HQ estimate x ˜ .

3. Experiment

3.1. Dataset Acquisition

A phased-array image configuration based on a VERMON P2-8 transducer (VERMON SA, Tours, France) (number of elements: 80; center frequency: 2.8 MHz; element width: 200 μm; kerf: 40 μm) was used to simulate a training dataset, the open-source ultrasound simulator (SIMUS) [26,27], from computer-generated phantom images. The phantom images were constructed by randomly distributing sets of geometric shapes with varying dimensions and three levels of brightness over an image (Figure 4a). One phantom image is passed to SIMUS to generate the scatter points that are used for the US RF data simulation. Low-quality (LQ) input data were simulated using a single insonification acquisition (Figure 4b), whereas the high-quality (HQ) target data were obtained by coherently compounding 20 acquisitions that corresponded to the acquisition from steered angles between ±45° (Figure 4c). The acquired RF data were demodulated (via downmixing and low-pass filtering) and beamformed using DAS [4], with a mean velocity of 1540 m/s over a grid image of 128 × 128 pixels that corresponded to a depth of 60 mm (from 10 mm to 70 mm) and a sectorial angle of 90°. A total of 6000 phantom images were simulated to generate training pairs (LQ, HQ). We used the complex IQ data for the C-Res-att-UNet model and performed bmode conversion to obtain images for the Res-att-UNet model.
In addition to the simulated dataset, we evaluated both models using the PICMUS dataset, containing in vitro, simulated, and in vivo data. The PICMUS dataset was not included in the training.

3.2. Network Training

The network was implemented in Python 3.9 using a Keras API [28] and a Keras Complex API [29]. A selection of 4350 samples were used as the training set, 1100 were used for validation, and the remaining 550 samples were used for testing. The network was trained using an Adam optimizer [30] with a batch size of 7 and an initial learning rate of 0.001. An early stopping mechanism was implemented with a learning-rate scheduler to prevent over-fitting. The learning rate was halved if there was no reduction in the validation loss for 10 epochs, and the training was terminated if there were 25 epochs with no reductions in the validation loss. The training was conducted using a NVIDIA GeForce RTX 3080 GPU (Nvidia, Santa Clara, CA, USA).

3.3. Evaluation Metrics

To objectively assess the reconstruction performance of the proposed approach, we employ the following evaluation metrics:
Contrast-to-noise ratio (CNR) measures the signal–intensity ratio between the region of interest and the background [31]:
C N R = | μ R μ B | σ R 2 + σ B 2 ,
where μ R , μ B ( σ R , σ B ) are the mean (standard deviation) of the region of interest and the background, respectively.
Structural similarity index (SSIM) measures the structural similarity between a given image and a reference image. The reference image is created via compounded imaging (20 DWs for the test set and 75 steered PWs for the PICMUS):
S S I M = ( 2 μ Y ^ μ Y + C 1 ) ( 2 σ Y ^ Y + C 2 ) ( μ Y ^ 2 + μ Y 2 + C 1 ) ( σ Y ^ 2 + σ Y 2 + C 2 ) ,
where μ Y ^ and μ Y ( σ Y ^ 2 and σ Y 2 ) are the means (variances) of Y ^ and Y, respectively, σ Y ^ Y is the covariance between Y ^ and Y, and C 1 and C 2 are two constants that stabilize the division with a weak denominator [21].
Peak signal-to-noise ratio (PSNR) is defined as the ratio of the image’s peak value to the distorting noise that degrades the quality of its representation [21]:
P S N R = 10 log 10 M A X Y 2 R M S E [ Y ^ , Y ] ,
where M A X Y is the maximum pixel value of the image and R M S E is the root mean square error.

4. Results

4.1. Learning Convergence

The training ended according to the implemented early stopping mechanism for both networks, C-Res-Att-UNet and Res-Att-UNet, ensuring that the models reached an optimal state. The weights obtained from the last training session are used for further analysis. C-Res-Att-UNet required 133 epochs, with an average training time per epoch of 150 s. Res-Att-UNet, on the other hand, required 80 epochs, with an average training time per epoch of 38 s. Neither network suffered from under- or over-fitting. C-Res-Att-UNet required a longer training time, as we implemented custom layers that handle complex data and therefore necessitate more multiplication–addition operations.

4.2. Image Quality

The overall image quality obtained from both models was similar to that of the images obtained via standard compounding of 20 DWs. To assess the quality of both models, we performed inference using a test set and the PICMUS dataset (using simulated, in vitro, and in vivo data).
(1) Test set: Figure 5 depicts samples from the test set displaying B-mode images of different techniques. C-Res-Att-UNet and Res-Att-UNet can produce images that are comparable to those created via DW imaging by effectively preserving important details and separating different regions from background noise. This improvement is attributed to the attention mechanism that is adapted to the complex domain in C-Res-Att-UNet and regular attention in Res-Att-UNet.
The previously defined evaluation metrics were computed over 40 samples from the test set for the three regions (background, anechoic, and hyperechoic), each represented by a 20 × 20 pixel square region (depicted in red, green, and blue in Figure 4). The results are presented in Table 1. We observe that C-Res-Att-UNet achieves the highest result in terms of CNR for the hyperechoic region, indicating superior contrast and visibility. For the anechoic region, Res-Att-UNet achieves the best result, while the reference of 20 DW achieves a lower result of same metric in the same region. Finally, comparing SSIM and PSNR using 20 DW images as the reference images, it is clear that Res-Att-UNet achieves the best PSNR value, while C-Res-Att-UNet presents a slightly better SSIM, indicating the closest similarity with the reference images. Overall, C-Res-Att-UNet achieves the most comparable performance to compounding imaging.
(2) PICMUS: Figure 6 illustrates the inference on the PICMUS dataset using the in vitro (CIRS phantom), simulated (point target and cyst target), and in vivo (carotid longitudinal section) data with a single PW. Both models successfully reconstructed images of similar visual quality in all four cases, offering a performance comparable to that of compounding imaging.
We computed the evaluation metrics for the image in Figure 6c for regions inside and outside the nine anechoic cysts. The results are presented in Table 2. Res-Att-UNet achieved superior results in terms of CNR, while C-Res-Att-UNet achieved the best results in terms of SSIM and PSNR, indicating the closest similarity with standard compounding of 75 PWs as a reference. Once again, C-Res-Att-UNet achieved the most comparable performance to compounding imaging.
Lastly, we computed the PSNR and SSIM metrics for the in vivo sample shown in Figure 6c. The Res-Att-UNet model achieved SSIM and PSNR values of 0.399 and 16.99, respectively. In comparison, the C-Res-Att-UNet model achieved higher values, with a SSIM of 0.595 and a PSNR of 18.54. Although both models showed improvements on the in vivo sample, they were not trained on real-world data. To ensure generalizability to clinical applications, further tuning using real-world data is necessary.

4.3. Lateral Resolution

The lateral profiles depicted in Figure 7 provide detailed insight into the performance of the trained models employing a single PW, the standard compounding of 75 PWs for in vitro data, and the point target represented in Figure 6. These profiles were taken at depths of 26 mm and 33 mm, respectively. In Figure 7, we observe that both models offers better reconstructions compared to single PW imaging (green circle). However, C-Res-Att-UNet (magenta square) demonstrates a high-quality reconstruction that is comparable to the target image (cyan asterisk) in both scenarios. The sharp reconstruction is related to the defined complex attention with the enhanced complex up-sampling layer.

4.4. Computational Performance

Table 3 presents the computational data of each model in terms of the number of training parameters, training time, and inference time on the GPU (inference was performed on the machine described in Section 3.2). Res-Att-UNet took around 50 min to train, whereas C-Res-Att-Unet took approximately 5.5 h to train. This indicates that a model that learns image-to-image mapping is significantly faster than a model attempting to learn complex-to-complex data mapping. This difference in training time has also been attributed to the built layers. For example, the concatenation layer for image data takes one concatenation step, while complex data take two concatenation steps for the real and imaginary parts and one last concatenation to produce a tensor that is compatible with a Keras Complex API. Another observation is the inference time on the GPU, which took 6 ms and 2 ms for C-Res-Att-UNet and Res-Att-UNet, respectively. We also compared the processing time of the PICMUS data with the provided scripts that ran on the CPU using MATLAB R2020a. The single PW took 1.02 s, while compounded imaging of 75 PWs took 42.84 s. Consequently, the marginal increase of 6 ms achieved by C-Res-Att-UNet remains unnoticeable with regard to compounded imaging, especially when considering the image quality achieved using a single PW.

5. Discussion

In this study, we presented a detailed analysis of a complex-valued UNet-based model for US image reconstruction from complex IQ signals in comparison to a standard DAS and UNet-based image model using a single PW. The results indicate that both Res-Att-UNet and C-Res-Att-UNet significantly improved the final US image, which was proved by the CNR, SSIM, and PSNR metrics compared to DAS. Specifically, C-Res-Att-UNet achieved the highest CNR in the hyperechoic region and the best SSIM, indicating the closest similarity to the standard compounding reference images in the simulated and PICMUS data.
In addition to the quantitative evaluation, a lateral resolution analysis was conducted at different depths (26 mm and 33 mm) for in vitro and simulated point target scenarios. The lateral profiles illustrate that both Res-Att-UNet and C-Res-Att-UNet offer superior image reconstruction quality compared to single PW imaging. Notably, C-Res-Att-UNet demonstrated a high-quality reconstruction that was comparable to the target image, attributed to its complex attention mechanism and enhanced complex up-sampling layer.
In summary, the analysis demonstrated that both Res-Att-UNet and C-Res-Att-UNet significantly enhanced US imaging quality from a single PW, achieving results close to those of standard compounding with multiple DWs. C-Res-Att-UNet consistently provided the highest SSIM and PSNR values, indicating its superior capability to preserve image structure and detail, making it a promising approach for fast and high-quality US imaging.
The downside of C-Res-Att-UNet was observed in the training time, as the network configuration doubled the number of addition–multiplication operations, leading to a slow learning time and relatively acceptable inference time concerning US imaging applications.
For future work, we would like to obtain real-world clinical data to improve and fine-tune the complex model to further investigate the performance of such an approach.

6. Conclusions

In this paper, we proposed a complex residual attention U-Net to achieve ultrasound image reconstruction using a single PW. The network uses complex convolution and attention mechanisms to incorporate phase information into training. We introduced additional layers to enhance complex data representation. An equivalent network of C-Res-Att-UNet named Res-Att-UNet for image data was trained for comparison. Our results showcase the performance of our approach to reconstructing images that are of a similar quality and resolution to compounded imaging in comparison to Res-Att-UNet. This improvement is attributed to the use of phase information and complex attention. However, our technique has several limitations: CVNNs require additional computation compared to RVNNs, which increases the training and inference time. Hence, further optimization is required. Nonetheless, this work will facilitate further investigations in complex data processing for a full CVCNN beamformer network, bypassing the DAS beamformer.

Author Contributions

Methodology: A.B. and C.S.; project administration: C.H.; software: A.B.; writing—original draft: A.B.; writing—review and editing: C.S., P.-H.C., F.R., A.G.-P. and C.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data can be made available from the corresponding author upon reasonable request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Sandrin, L.; Catheline, S.; Tanter, M.; Hennequin, X.; Fink, M. Time-resolved pulsed elastography with ultrafast ultrasonic imaging. Ultrason. Imaging 1999, 21, 259–272. [Google Scholar] [CrossRef] [PubMed]
  2. Hasegawa, H.; Kanai, H. High-frame-rate echocardiography using diverging transmit beams and parallel receive beamforming. J. Med. Ultrason. 2011, 38, 129–140. [Google Scholar] [CrossRef] [PubMed]
  3. Trees, H. Optimum Array Processing—Part IV of Detection, Estimation, and Modulation Theory; Wiley: Hoboken, NJ, USA, 2002. [Google Scholar]
  4. Perrot, V.; Polichetti, M.; Varray, F.; Garcia, D. So you think you can DAS? A viewpoint on delay-and-sum beamforming. Ultrasonics 2021, 111, 106309. [Google Scholar] [CrossRef] [PubMed]
  5. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. In Advances in Neural Information Processing Systems; Pereira, F., Burges, C., Bottou, L., Weinberger, K., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2012; Volume 25. [Google Scholar]
  6. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015: 18th International Conference, Munich, Germany, 5–9 October 2015; Springer: Berlin/Heidelberg, Germany, 2015. Proceedings, Part III. pp. 234–241. [Google Scholar]
  7. Li, X.; Chen, H.; Qi, X.; Dou, Q.; Fu, C.W.; Heng, P.A. H-DenseUNet: Hybrid Densely Connected UNet for Liver and Tumor Segmentation From CT Volumes. IEEE Trans. Med. Imaging 2018, 37, 2663–2674. [Google Scholar] [CrossRef] [PubMed]
  8. Wu, K.; Chen, X.; Ding, M. Deep learning based classification of focal liver lesions with contrast-enhanced ultrasound. Opt. Int. J. Light Electron Opt. 2014, 125, 4057–4063. [Google Scholar] [CrossRef]
  9. Cheng, J.Z.; Ni, D.; Chou, Y.H.; Qin, J.; Tiu, C.M.; Chang, Y.C.; Huang, C.S.; Shen, D.; Chen, C.M. Computer-aided diagnosis with deep learning architecture: Applications to breast lesions in US images and pulmonary nodules in CT scans. Sci. Rep. 2016, 6, 24454. [Google Scholar] [CrossRef] [PubMed]
  10. Yoon, Y.H.; Ye, J.C. Deep Learning for Accelerated Ultrasound Imaging. In Proceedings of the 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; pp. 6673–6676. [Google Scholar] [CrossRef]
  11. Luijten, B.; Cohen, R.; de Bruijn, F.J.; Schmeitz, H.A.; Mischi, M.; Eldar, Y.C.; van Sloun, R.J. Deep Learning for Fast Adaptive Beamforming. In Proceedings of the ICASSP 2019—2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019; pp. 1333–1337. [Google Scholar] [CrossRef]
  12. Gasse, M.; Millioz, F.; Roux, E.; Garcia, D.; Liebgott, H.; Friboulet, D. High-Quality Plane Wave Compounding Using Convolutional Neural Networks. IEEE Trans. Ultrason. Ferroelectr. Freq. Control 2017, 64, 1637–1639. [Google Scholar] [CrossRef] [PubMed]
  13. Hyun, D.; Brickson, L.L.; Looby, K.T.; Dahl, J.J. Beamforming and Speckle Reduction Using Neural Networks. IEEE Trans. Ultrason. Ferroelectr. Freq. Control 2019, 66, 898–910. [Google Scholar] [CrossRef] [PubMed]
  14. Simson, W.; Paschali, M.; Navab, N.; Zahnd, G. Deep Learning Beamforming for Sub-Sampled Ultrasound Data. In Proceedings of the 2018 IEEE International Ultrasonics Symposium (IUS), Kobe, Japan, 22–24 October 2018; pp. 1–4. [Google Scholar] [CrossRef]
  15. Nair, A.A.; Tran, T.D.; Reiter, A.; Lediju Bell, M.A. A Deep Learning Based Alternative to Beamforming Ultrasound Images. In Proceedings of the 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; pp. 3359–3363. [Google Scholar] [CrossRef]
  16. Hirose, A.; Yoshida, S. Generalization Characteristics of Complex-Valued Feedforward Neural Networks in Relation to Signal Coherence. IEEE Trans. Neural Netw. Learn. Syst. 2012, 23, 541–551. [Google Scholar] [CrossRef] [PubMed]
  17. Sarroff, A.M.; Shepardson, V.; Casey, M.A. Learning representations using complex-valued nets. arXiv 2015, arXiv:1511.06351. [Google Scholar]
  18. Senouf, O.; Vedula, S.; Zurakhov, G.; Bronstein, A.; Zibulevsky, M.; Michailovich, O.; Adam, D.; Blondheim, D. High frame-rate cardiac ultrasound imaging with deep learning. In Proceedings of the Medical Image Computing and Computer Assisted Intervention—MICCAI 2018: 21st International Conference, Granada, Spain, 16–20 September 2018; Springer: Berlin/Heidelberg, Germany, 2018. Proceedings, Part I. pp. 126–134. [Google Scholar]
  19. Trabelsi, C.; Bilaniuk, O.; Zhang, Y.; Serdyuk, D.; Subramanian, S.; Santos, J.F.; Mehri, S.; Rostamzadeh, N.; Bengio, Y.; Pal, C.J. Deep complex networks. arXiv 2017, arXiv:1705.09792. [Google Scholar]
  20. Lu, J.; Millioz, F.; Garcia, D.; Salles, S.; Ye, D.; Friboulet, D. Complex convolutional neural networks for fast diverging wave imaging. In Proceedings of the 2020 IEEE International Ultrasonics Symposium (IUS), Virtual Event, 6–11 September 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 1–3. [Google Scholar]
  21. Lu, J.; Millioz, F.; Garcia, D.; Salles, S.; Ye, D.; Friboulet, D. Complex convolutional neural networks for ultrafast ultrasound imaging reconstruction from in-phase/quadrature signal. IEEE Trans. Ultrason. Ferroelectr. Freq. Control 2021, 69, 592–603. [Google Scholar] [CrossRef] [PubMed]
  22. Lan, Y.; Zhang, X. Real-Time Ultrasound Image Despeckling Using Mixed-Attention Mechanism Based Residual UNet. IEEE Access 2020, 8, 195327–195340. [Google Scholar] [CrossRef]
  23. Oktay, O.; Schlemper, J.; Folgoc, L.L.; Lee, M.; Heinrich, M.; Misawa, K.; Mori, K.; McDonagh, S.; Hammerla, N.Y.; Kainz, B.; et al. Attention u-net: Learning where to look for the pancreas. arXiv 2018, arXiv:1804.03999. [Google Scholar]
  24. Shi, W.; Caballero, J.; Huszár, F.; Totz, J.; Aitken, A.P.; Bishop, R.; Rueckert, D.; Wang, Z. Real-Time Single Image and Video Super-Resolution Using an Efficient Sub-Pixel Convolutional Neural Network. arXiv 2016, arXiv:cs.CV/1609.05158. [Google Scholar]
  25. Liebgott, H.; Rodriguez-Molares, A.; Cervenansky, F.; Jensen, J.; Bernard, O. Plane-Wave Imaging Challenge in Medical Ultrasound. In Proceedings of the 2016 IEEE International Ultrasonics Symposium (IUS), Tours, France, 18–21 September 2016; pp. 1–4. [Google Scholar] [CrossRef]
  26. Garcia, D. SIMUS: An open-source simulator for medical ultrasound imaging. Part I: Theory & examples. Comput. Methods Programs Biomed. 2022, 218, 106726. [Google Scholar]
  27. Cigier, A.; Varray, F.; Garcia, D. SIMUS: An open-source simulator for medical ultrasound imaging. Part II: Comparison with four simulators. Comput. Methods Programs Biomed. 2022, 220, 106774. [Google Scholar] [CrossRef]
  28. Chollet, F. Keras. 2015. Available online: https://keras.io (accessed on 27 March 2015).
  29. Dramsch, J.S.; Trabelski, C.; Bilaniuk, O.; Serdyuk, D. Complex-Valued Neural Networks in Keras with Tensorflow. 2019. Available online: https://keras-complex.readthedocs.io/ (accessed on 7 September 2019). [CrossRef]
  30. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  31. Rindal, O.M.H.; Åsen, J.P.; Holm, S.; Austeng, A. Understanding contrast improvements from capon beamforming. In Proceedings of the 2014 IEEE International Ultrasonics Symposium, Chicago, IL, USA, 3–6 September 2014; pp. 1694–1697. [Google Scholar] [CrossRef]
Figure 1. Block diagram of the proposed complex max-pooling layer. Solid and dashed lines indicate real and imaginary parts, respectively.
Figure 1. Block diagram of the proposed complex max-pooling layer. Solid and dashed lines indicate real and imaginary parts, respectively.
Sensors 24 05111 g001
Figure 2. Block diagram of the proposed complex up-sampling layer with a factor of 2. The solid and dashed lines indicate real and imaginary parts, respectively.
Figure 2. Block diagram of the proposed complex up-sampling layer with a factor of 2. The solid and dashed lines indicate real and imaginary parts, respectively.
Sensors 24 05111 g002
Figure 3. Block diagram of the proposed complex residual attention U-Net. x ˜ is the low-quality (LQ) complex input and x ^ is the high-quality (HQ) reconstruction.
Figure 3. Block diagram of the proposed complex residual attention U-Net. x ˜ is the low-quality (LQ) complex input and x ^ is the high-quality (HQ) reconstruction.
Sensors 24 05111 g003
Figure 4. A sample from the phantom and training sets. The red area indicates a background region, the green area indicates an anechoic region, and the blue area indicates an hyperechoic region. (a) A computer-generated phantom used to simulate the dataset; (b) a B-mode image of low-quality IQ data acquired using single insonification, (c) a B-mode image of high-quality IQ data from standard compounding 20 steered acquisitions.
Figure 4. A sample from the phantom and training sets. The red area indicates a background region, the green area indicates an anechoic region, and the blue area indicates an hyperechoic region. (a) A computer-generated phantom used to simulate the dataset; (b) a B-mode image of low-quality IQ data acquired using single insonification, (c) a B-mode image of high-quality IQ data from standard compounding 20 steered acquisitions.
Sensors 24 05111 g004
Figure 5. Three B-mode samples (ac) from test datasets comparing different techniques. All results are from a single-plane wave insonification except for the standard compounding, which is obtained by coherently compounding 20 steered insonifications. Res-Att-UNet and C-Res-Att-UNet show visual improvements by separating different regions of noise.
Figure 5. Three B-mode samples (ac) from test datasets comparing different techniques. All results are from a single-plane wave insonification except for the standard compounding, which is obtained by coherently compounding 20 steered insonifications. Res-Att-UNet and C-Res-Att-UNet show visual improvements by separating different regions of noise.
Sensors 24 05111 g005
Figure 6. B-mode results of Res-Att-UNet and C-Res-Att-UNet with 1PW using the PICMUS dataset [25] compared to DAS (1PW) and standard compounding (75 PWS). (a) In vitro CIRS phantom, (b,c) simulated, and (d) in vivo carotid longitudinal section.
Figure 6. B-mode results of Res-Att-UNet and C-Res-Att-UNet with 1PW using the PICMUS dataset [25] compared to DAS (1PW) and standard compounding (75 PWS). (a) In vitro CIRS phantom, (b,c) simulated, and (d) in vivo carotid longitudinal section.
Sensors 24 05111 g006
Figure 7. Lateral profiles of (a) the in vitro Figure 6 at a 26 mm depth and (b) the simulated point target Figure 6 at a 33 mm depth. Depth is indicated with a dashed yellow line.
Figure 7. Lateral profiles of (a) the in vitro Figure 6 at a 26 mm depth and (b) the simulated point target Figure 6 at a 33 mm depth. Depth is indicated with a dashed yellow line.
Sensors 24 05111 g007
Table 1. CNR, SSIM, and PSNR of DAS (1PW), Res-Att-UNet (1 PW), C-Res-Att-UNet (1 PW), and standard compounding (20 DWs).
Table 1. CNR, SSIM, and PSNR of DAS (1PW), Res-Att-UNet (1 PW), C-Res-Att-UNet (1 PW), and standard compounding (20 DWs).
MethodHyperechoic Region CNRAnechoic Region CNRSSIMPSNR
DAS (1 PW)1.1831.0610.22714.35
Res-Att-UNet (1 PW)1.2781.0120.82523.86
C-Res-Att-UNet (1 PW)1.3140.9940.83621.74
Standard compounding (20 DWs)1.2200.9401
Bold highlights the best value for each metric across different methods.
Table 2. CNR, SSIM, and PSNR of DAS (1 PW), Res-Att-UNet (1 PW), C-Res-Att-UNet (1 PW), and standard compounding (75 PWs) of Figure 6c.
Table 2. CNR, SSIM, and PSNR of DAS (1 PW), Res-Att-UNet (1 PW), C-Res-Att-UNet (1 PW), and standard compounding (75 PWs) of Figure 6c.
MethodCNRSSIMPSNR
DAS (1 PW)1.1200.48911.316
Res-Att-UNet (1 PW)1.1570.67317.446
C-Res-Att-UNet (1 PW)0.9190.71817.611
Standard compounding (75 DWs)1.02411
Bold highlights the best value for each metric across different methods.
Table 3. Number of parameters, training time, and inference time of Res-Att-UNet and C-Res-Att-UNet.
Table 3. Number of parameters, training time, and inference time of Res-Att-UNet and C-Res-Att-UNet.
ModelNumber of ParametersTraining TimeInference Time (GPU)
Res-Att-UNet29 million50 min2 ms
C-Res-Att-UNet31 million5.5 h6 ms
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Bentaleb, A.; Sintes, C.; Conze, P.-H.; Rousseau, F.; Guezou-Philippe, A.; Hamitouche, C. Complex Residual Attention U-Net for Fast Ultrasound Imaging from a Single Plane-Wave Equivalent to Diverging Wave Imaging. Sensors 2024, 24, 5111. https://doi.org/10.3390/s24165111

AMA Style

Bentaleb A, Sintes C, Conze P-H, Rousseau F, Guezou-Philippe A, Hamitouche C. Complex Residual Attention U-Net for Fast Ultrasound Imaging from a Single Plane-Wave Equivalent to Diverging Wave Imaging. Sensors. 2024; 24(16):5111. https://doi.org/10.3390/s24165111

Chicago/Turabian Style

Bentaleb, Ahmed, Christophe Sintes, Pierre-Henri Conze, François Rousseau, Aziliz Guezou-Philippe, and Chafiaa Hamitouche. 2024. "Complex Residual Attention U-Net for Fast Ultrasound Imaging from a Single Plane-Wave Equivalent to Diverging Wave Imaging" Sensors 24, no. 16: 5111. https://doi.org/10.3390/s24165111

APA Style

Bentaleb, A., Sintes, C., Conze, P. -H., Rousseau, F., Guezou-Philippe, A., & Hamitouche, C. (2024). Complex Residual Attention U-Net for Fast Ultrasound Imaging from a Single Plane-Wave Equivalent to Diverging Wave Imaging. Sensors, 24(16), 5111. https://doi.org/10.3390/s24165111

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop