Next Article in Journal
Lymphocyte-to-Monocyte Ratio and Clinical Outcomes in Cholangiocarcinoma: A Systematic Review and Meta-Analysis
Next Article in Special Issue
Prognostic Comparison between Oncotype DX® and a 23-Gene Classifier, RecurIndex®, on the Taiwan Breast Cancer Population
Previous Article in Journal
Monitoring Leishmania infantum Infections in Female Lutzomyia longipalpis by Using DNA Extraction on Cation Exchange Paper and PCR Pool Testing
Previous Article in Special Issue
Residual Tumor Patterns of Breast Cancer on MRI after Neo-Adjuvant Chemotherapy: Impact on Clinicopathologic Factors and Prognosis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

BUViTNet: Breast Ultrasound Detection via Vision Transformers

1
Department of Medical IT Convergence Engineering, Kumoh National Institute of Technology, Gumi 39253, Korea
2
Department of IT Convergence Engineering, Kumoh National Institute of Technology, Gumi 39253, Korea
*
Author to whom correspondence should be addressed.
Diagnostics 2022, 12(11), 2654; https://doi.org/10.3390/diagnostics12112654
Submission received: 13 October 2022 / Revised: 26 October 2022 / Accepted: 26 October 2022 / Published: 1 November 2022
(This article belongs to the Special Issue Breast Cancer Metastasis, Diagnostic and Therapeutic Approaches 2022)

Abstract

:
Convolutional neural networks (CNNs) have enhanced ultrasound image-based early breast cancer detection. Vision transformers (ViTs) have recently surpassed CNNs as the most effective method for natural image analysis. ViTs have proven their capability of incorporating more global information than CNNs at lower layers, and their skip connections are more powerful than those of CNNs, which endows ViTs with superior performance. However, the effectiveness of ViTs in breast ultrasound imaging has not yet been investigated. Here, we present BUViTNet breast ultrasound detection via ViTs, where ViT-based multistage transfer learning is performed using ImageNet and cancer cell image datasets prior to transfer learning for classifying breast ultrasound images. We utilized two publicly available ultrasound breast image datasets, Mendeley and breast ultrasound images (BUSI), to train and evaluate our algorithm. The proposed method achieved the highest area under the receiver operating characteristics curve (AUC) of 1 ± 0, Matthew’s correlation coefficient (MCC) of 1 ± 0, and kappa score of 1 ± 0 on the Mendeley dataset. Furthermore, BUViTNet achieved the highest AUC of 0.968 ± 0.02, MCC of 0.961 ± 0.01, and kappa score of 0.959 ± 0.02 on the BUSI dataset. BUViTNet outperformed ViT trained from scratch, ViT-based conventional transfer learning, and CNN-based transfer learning in classifying breast ultrasound images (p < 0.01 in all cases). Our findings indicate that improved transformers are effective in analyzing breast images and can provide an improved diagnosis if used in clinical settings. Future work will consider the use of a wide range of datasets and parameters for optimized performance.

1. Introduction

Breast cancer is the most prevalent malignancy, accounting for 12.5% of new cases globally by 2020 [1,2]. The survival rate of breast cancer for at least five years after diagnosis varies among populations depending on the economic standing of nations; it is greater than 90% in high-income nations, 40–65% in middle-income nations, and less than 40% in low-income nations [2,3]. This difference is primarily attributable to the ability of high-income nations to detect breast cancer early [1]. Early detection expands treatment options and lowers the risk of breast-cancer-related deaths considerably. The primary imaging modalities for the early diagnosis of breast cancer include mammography and ultrasound [4].
Previously, ultrasonography was considered helpful only for cyst diagnosis. However, it improves local preoperative staging, guided interventional diagnosis, and differential diagnosis of benign and malignant tumors [5]. Mammography has a low sensitivity for thick breasts [6]. Women with dense parenchyma are far more likely to develop breast cancer [7]. Dense breast tissue can be examined using ultrasound [8]. Recent research has demonstrated that high-resolution ultrasonography increases the diagnosis of tiny tumors by three to four tumors per 1000 women without clinical or mammographic abnormalities [9]. Carcinomas observed on mammography and sonography have a similar stage distribution [10]. To overcome the limitations of mammography, ultrasound is frequently employed for curative diagnosis [11]. Ultrasound is noninvasive, widely available, easy to use, less expensive, and provides real-time imaging [12]. Moreover, ultrasound is safer and does not generate radiation [13]. More importantly, ultrasound helps detect tumors in women with dense breasts and detects and classifies breast lesions that cannot be interpreted adequately through mammography alone [14]. Doctors can use ultrasound to identify many clinically relevant regions such as benign cysts or normal tissues. Most women over the age of 30 years usually undergo ultrasonography along with mammography. For women under the age of 30, ultrasonography is sufficient to decide whether a biopsy is necessary in a particular area of the suspected breast.
However, early breast ultrasound diagnosis has limitations [15]. First, follow-up ultrasound, aspiration, or biopsy may be performed after a breast ultrasound image is interpreted [16]. Biopsy may be recommended to determine whether a suspicious abnormality is cancer. Most of the suspected problem locations detected by ultrasonography that require biopsy are noncancerous (false positives) [17]. Second, although ultrasound is a breast imaging method, annual mammography is still recommended [1]. Many tumors cannot be detected via ultrasonography. Numerous calcifications detected by mammography are invisible with ultrasonography [18]. In mammography, some early breast tumors simply manifest as calcifications [19,20]. Even for women with dense breasts, many institutions do not offer ultrasound screening, and some insurance plans might not pay for treatment. As a real-time test, ultrasound depends on the anomaly detected during the scan. Therefore, sophisticated tools and experienced professionals are required.
Deep learning (DL) has been used to overcome the limitations of ultrasound in the early detection of breast cancer [21]. Many studies have been conducted in the areas of synthetic imaging, object detection, segmentation, and imaging categorization of breast lesions using DL [22]. Many of these approaches have received the necessary legal certifications and are used in clinical settings [23]. Additionally, DL approaches have demonstrated ability to perform at levels comparable to those of human experts in a variety of breast cancer detection tasks and have the potential to help physicians with little training improve the diagnosis of breast cancer in clinical settings [24].
All prior investigations regarding DL for ultrasound-based early breast cancer diagnosis have used convolutional neural networks (CNNs) [25]. Vision transformers (ViTs), developed by Dosovitskiy et al. [26] in 2020, outperform state-of-the-art (SOTA) CNN algorithms in natural images categorization. ViTs demonstrate a superior performance by incorporating more global information than a CNN does in lower layers and having a stronger skip connection than that of ResNet. CNNs require several layers to calculate features computed by a smaller collection of lower layers of ViTs. All these attributes help ViTs outperform CNNs in classifying real-world images. However, not much research has been conducted on the application of ViTs in medical image analysis. This is mainly because of the heavy reliance of ViTs on extensive training data. ViTs do not outperform CNNs with small-scale datasets. The heavy reliance of ViTs on big data has limited its effective use in certain sectors. The same is true for medical images, as it can be difficult to locate substantial training datasets. To overcome this challenge, transfer learning has been extensively used in CNN-based medical image analysis [25,27,28]. The use of ViT models that have been pretrained on large natural image datasets and transfer learning via ViT models may help increase the performance of DL in medical image analysis. Hence, we present BUViTNet breast ultrasound detection via ViTs, a new transfer-learning strategy based on ViTs for the classification of ultrasound breast cancer images.
The proposed ViT-based transfer learning has multiple advantages over CNN-based transfer learning. The disadvantages of CNNs include their high processing cost, narrow focus on a portion of the image rather than the entire image, inability to encode relative spatial information, and inability to handle rotation and scale invariance without augmentation. CNNs are computationally expensive owing to the use of pixel arrays [29]. Because deeper layers are required to extract features from the entire image, CNNs require more training time and hardware. In addition, because of the information lost during processes such as pooling (average- or max-pooling), CNNs become slower and less accurate. A convolutional layer is the primary building block of a CNN. Its task is to identify significant details in the image pixels. Higher layers combine simple characteristics into more complex features, whereas deeper layers (closer to the input) learn to recognize simple features such as edges and color gradients. Very high-level features are finally combined in thick layers at the top of the network to create classification predictions. All higher-level neurons in a CNN receive low-level information. Then, these neurons perform additional convolutions to determine whether specific features are present. This is accomplished by repeating the knowledge across all the individual neurons after traversing the receptive field. The location and orientation of the object are not considered by a CNN when forming the predictions. They entirely lose all internal information about the position and orientation of the item and route all of the data to the same neurons that may not be able to process this type of input [30]. A CNN predicts output by observing an image and determining whether specific elements are present. If they are, the image is appropriately categorized. An artificial neuron produces only one scalar. Additionally, CNNs employ convolutional layers that produce a 2D matrix, with each number representing the result of the kernel’s convolution with a portion of the input volume. These layers duplicate the weights of each kernel across the entire input volume for each kernel. Therefore, we can consider the 2D matrix as the output of the replicated feature detector. The output of a convolutional layer is then produced by stacking all the 2D matrices of the kernel on top of one another. The next step is to attempt to attain perspective invariance in the neuronal activity. This is accomplished using max pooling, which selects the greatest value in each region of the 2D matrix after looking at each region in turn. Consequently, invariance is achieved. When the output is invariant, the input can be slightly altered without affecting the results. In other words, max pooling ensures that the network activities (neuron outputs) remain constant even if the object we wish to detect is slightly shifted in the input image, allowing the network to continue to recognize the object. However, because max pooling discards important data and does not encode the relative spatial relationships between features, the approach described above is ineffective [31]. Consequently, CNNs are not genuinely immune to significant input data alterations. The use of ViTs enables us to overcome these limitations of CNNs to detect breast cancer using ultrasound images. In this study, we also introduced a novel transfer-learning method to compensate for the ViT’s data hunger. The use of transfer learning enables ViT to be pretrained on a large number of natural and microscopic images and to be used to classify ultrasound images without overfitting.
Generally, the proposed method provides the following contributions.
  • Developed the first multistage transfer-learning method using vision transformers for breast cancer detection.
  • Utilized microscopic image datasets that have related image features to those of ultrasound images for intermediate-stage transfer learning to improve the performance of breast cancer early detection.
  • Carefully studied the characteristics of different vision transformers based on pretrained models for translation to ultrasound image-based breast cancer detection.
  • Investigated the effectiveness of the proposed BUViTNet method when applied to datasets from different sources as well as on mixed datasets from different origins.
  • Compared the performance of the BUViTNet method against vision transformers trained from scratch, conventional vision transformer-based transfer learning, and convolutional neural networks for breast cancer detection.

2. Materials and Methods

2.1. Dataset

We evaluated our method using two publicly available datasets: breast ultrasound image (BUSI) (Available online: https://scholar.cu.edu.eg/?q=afahmy/pages/dataset (access on 1 July 2022)) and Mendeley breast ultrasound (Available online: https://data.mendeley.com/datasets/wmy84gzngw/1 (access on 1 July 2022)) datasets. Moreover, we created a mixed BUSI and Mendeley dataset to evaluate the performance of the proposed method on datasets from different sources.
The BUSI dataset is categorized into three classes: benign (210 images), malignant (437 images), and normal (210 images). BUSI dataset images were taken from women between the ages of 25 and 75 years; hence, the dataset is preferred for studies involving early breast cancer detection in women below 40 years of age [32]. The dataset was collected in 2018 from 600 female patients. The dataset consists of 780 images, each with an average size of 500 × 500 pixels. The dataset images are PNG files. Representative images from the dataset are shown in Figure 1.
The Mendeley breast ultrasound dataset is divided into two categories: benign (100 images) and malignant (150 images) [33]. The dataset was collected in Brazil from women aged 30–70 years. The images in the dataset are in PNG format and of different sizes. This dataset is practical for our needs and was extensively used in numerous investigations. Representative images from this dataset are shown in Figure 2.
The mixed dataset is categorized into two classes: benign (310) and malignant (587). We mixed the datasets to evaluate the performance of the proposed method on datasets from different locations. We added benign images from the BUSI and Mendeley datasets to make a mixed benign image class, and the same was done for malignant images.

2.2. Preprocessing

The preprocessing of the data involved resizing the dataset images and balancing the number of images in each class. The images in both datasets utilized in this study had different pixel sizes from their corresponding sources. Therefore, we resized them to use them as input in our proposed method. Each image was resized to 224 × 224 pixels. Additionally, the datasets were class-imbalanced owing to the different number of images for each class. Class imbalance impinges upon the learning process by biasing the model toward the class having a larger dataset size. To prevent this, we implemented spatial transformation-based augmentations (horizontal flip, color jitter, sharpening, salt-and-pepper, gamma correction, random rotation, and height shift) to equalize the size of the images in each class. Subsequently, we created a BUSI dataset with three classes, each having 500 images; a Mendeley dataset with two classes, each having 500 images; and a mixed dataset with two classes, each having 1000 images. Following this, we categorized our dataset into training, validation, and test sets with a 7:2:1 ratio. Figure 3 depicts the dataset distribution after preprocessing the BUSI dataset. The training data were further augmented to enrich the training images. We performed random rotation, flip (both horizontal and vertical), shift (both width and height), shear, and other augmentations.

2.3. Proposed Method

The proposed method employs multistage transfer learning. A ViT model pretrained using the ImageNet dataset was first used for transfer learning with cancer cell line images to distinguish breast (MCF-7), lung (NCI-60), and cervical (HeLa) cancers. Figure 4 shows the acquisition and preparation of the cancer cell images for the intermediate stage transfer learning. The cancer cell images were captured under microscope, separated from the background, and then segmented to form patches of each cancer cell line. Details about acquisition and preprocessing of cancer cell lines can be found in [28]. Using the ImageNet pretrained model, transfer learning was employed to train a model for classifying the cancer cell line images as breast, lung, and cervical cancers.
Then, the model trained for cancer cell classification was used as a pretrained model for transfer learning on breast ultrasound images. Figure 5 depicts the multistage approach used in this study. Multistage transfer learning was applied for two main reasons. The first was to develop a model that was trained on large image datasets with features similar to those of ultrasound images, which can be acquired in large numbers and are ideal as intermediate-stage transfer learning domains. A previous study we conducted also proved that CNN models pretrained on cancer cell line images improved the performance of classifying breast ultrasound images. The second reason was that the use of cancer cell lines would optimize large ViT models, which are data-hungry, for small datasets. We utilized 28,560 cancer cell images [28] to train the ViT models in the intermediate transfer-learning stage before using them for classifying ultrasound images.
Multistage transfer learning was applied using three ViT models: a base model with 16 × 16 input patch size (vitb_16), a base model with 32 × 32 input patch size (vitb_32), and a large model with 32 × 32 input patch size (vitl_32). Table 1 lists their details.
The ViT-based transfer learning at both intermediate and task levels was implemented by replacing the last layer of each ViT model using a flattening layer followed by batch normalization, one dense layer followed by batch normalization, and a final dense layer as a classifier, as shown in Figure 6.
A transformer model is a neural network that learns context and meaning by tracking relationships in sequential data, such as the words in a sentence. Transformer models apply an evolving set of mathematical techniques, called attention or self-attention, to detect subtle ways by which even distant data elements in a series influence and depend on each other. ViTs use data sequences in the form of image patches. The ViTs used in the study operate as follows.
  • Conversion of images into patches: ViTs use image patches as tokens of words, as in the original paper Vaswani et al. [34]. For images, pixels can be considered; however, it increases the computational cost. Moreover, finding hardware to process high-resolution images such as medical images is challenging. Therefore, we propose the conversion of input images into patches, as in [26]. In this study, an image with H × W was converted into N patches of size P × P .
  • Flattening and patch embedding: The patches were then flattened and sent through a single feed-forward layer to obtain a linear patch projection. This feed-forward layer contained the embedding matrix E , as mentioned in Dosovitskiy et al. [26]. Matrix E is randomly generated.
  • Learnable embedding and positional embedding: Learnable embeddings are concatenated with patch projection, which is used later for classification. Because patches are not naturally formed in sequences, as in time sequence models, the transformers utilized positional embeddings to establish a certain order in the patches. The positional encoding matrix is randomly generated, as in patch embedding.
  • Multilayer perceptron (MLP) head: The outputs of the transformer encoder unit are passed to the MLP head for classification. Despite the multiple outputs from the transformer encoder, the MLP considered only one output related to the class embedding, whereas the other outputs are ignored. MLP outputs the probability distribution of the labels to which the corresponding images belong.

2.4. Implementation Details

We trained our models for 50 epochs using the Adam optimizer with a learning rate of 0.0001. A batch size of 16 and exponential decay were used for training. Our datasets were divided with a 7:2:1 ratio for training, validation, and testing. GELU was utilized as an activation function with an L2 regularizer for the ViT models. In CNNs, ReLu was utilized with an L2 regularizer. For experiments involving comparison with the proposed method, the same parameter settings were utilized to avoid bias in the results.

2.5. Experimental Settings

Four experiments were conducted to evaluate the proposed model. The first experiment was conducted to compare the performance of the proposed model in three ViT models (vitb_16, vitb_32, and vitl_32). This was done to select the best-performing model having the least computational cost. In the second experiment, a ViT model trained from scratch was implemented and compared with the proposed transfer-learning model. This experiment evaluated whether the ViT trained from scratch was more suitable for classifying breast ultrasound images than a ViT pretrained on different datasets prior to being applied to the task of classifying breast ultrasound images. The third experiment was performed to choose the transfer-learning approach (conventional ImageNet-only pretraining or the proposed transfer-learning method) appropriate for the ViT-based transfer- learning method in classifying breast ultrasound images. Finally, in the fourth experiment, the performance of ViT-based transfer learning was evaluated and compared with that of CNN-based transfer learning. This experiment provides an effective comparison between ViT-based and CNN-based transfer-learning methods for breast ultrasound image classification.

2.6. Evaluation Metrics

Metrics including accuracy, area under the receiver operating curve (AUC), F1-score, recall, and precision were used to assess the performance of our model. We used Matthew’s correlation coefficient (MCC) and kappa scores also to assess the performance of our model, owing to the uneven nature of our dataset and the superiority of these metrics in analyzing the performance. A 95% confidence interval was used to determine each outcome. In Table 2, TP, TN, FP, and FN stand for true positive, true negative, false positive, and false negative, respectively.

3. Results

The proposed method was evaluated using two datasets from different sources and their mixture, as listed in Table 3. The proposed method achieved an AUC of 1 ± 0, MCC score of 1 ± 0, and kappa score of 1 ± 0 on the Mendeley dataset for the vitb_16, vitb_32, and vitl_32 ViT models. Furthermore, the proposed method achieved the highest AUC of 0.968 ± 0.02, MCC score of 0.961 ± 0.01, and kappa score of 0.959 ± 0.02 on the BUSI dataset for the Vitb _16 ViT model. For the mixed dataset, the vitb_16 model achieved the highest AUC of 0.937 ± 0.03, MCC score of 0.924 ± 0.02, and kappa score of 0.919 ± 0.03.
Figure 7 shows the receiver operating characteristics (ROC) curves of the proposed BUViTNet method on the three datasets, using vitb_16 as a base model.
We compared the proposed method with ViT models trained from scratch using ultrasound images. This was performed to determine whether the ViT-based transfer-learning model performed better than a ViT model trained directly from scratch with ultrasound images. The highest AUC recorded with the models trained from scratch were 0.73 ± 0.2, 0.71 ± 0.07, and 0.7 ± 0.1 using vitb_16 on Mendeley, BUSI, and mixed datasets, respectively (Table 4). The proposed method significantly outperformed the ViT models trained from scratch with a p-value of less than 0.01 in all cases on all datasets.
Furthermore, the proposed transfer-learning method was compared with conventional ImageNet-pretrained ViT models. This experiment validates the superiority of the proposed method over traditional transfer learning. The highest AUCs achieved for conventional transfer learning using ViT models were 1 ± 0, 0.9548 ± 0.0183, and 0.9116 ± 0.0156 on Mendeley, BUSI, and mixed datasets, respectively (Table 5). The performances of the proposed method and traditional transfer learning on the Mendeley dataset were comparable; however, the proposed method outperformed the traditional transfer-learning method on BUSI and mixed datasets, with a p-value of less than 0.01.
Finally, we compared the proposed ViT-based transfer-learning method with the transfer-learning method using CNNs. To do so, we used three SOTA CNN architectures: ResNet50, EfficientNetB2, and InceptionNetV3. All implementation parameters were kept the same as those of the ViT-based transfer-learning method for a fair comparison. ResNet50-based transfer learning provided the highest AUC scores of 0.972 ± 0.01, 0.879 ± 0.2, and 0.836 ± 0.08 on Mendeley, BUSI, and mixed datasets, respectively (Table 6). The proposed ViT-based transfer-learning method performed better than the CNN-based transfer-learning method for breast ultrasound images, with a p-value of less than 0.01.

4. Discussion

In this study, we proposed a novel transfer-learning approach for ViT-based breast ultrasound image classification. Our approach uses a ViT model pretrained on the ImageNet dataset for transfer learning to classify cancer cell images. This model, trained on ImageNet and cancer cell images, was then used to classify breast ultrasound images. This novel transfer-learning approach enables a model to learn from a large number of natural and medical images before being used for classifying breast ultrasound images. The model leverages the features learned in the previous transfer-learning stages to use it for the target task. As a result, we were able to achieve the best performance in terms of all metrics used with our proposed model. We compared the proposed approach with ViT models trained from scratch, ViT-based conventional transfer learning, and CNN-based transfer learning. The proposed approach outperformed all of these models.
Regarding the performance of BUViTNet using various base models, vitb_16-based BUViTNet performed better than vitb_32 and vitl_32, thereby providing less computational complexity. The main reason for this might be the size of the patches used in these base models. The vitb_16 base model utilizes an input patch size of 16 × 16, whereas vitb_32 and vitl_32 base models utilize an input patch size of 32 × 32. The smaller the patch size, the higher the efficiency and effectiveness of the transformer encoder’s attention. This leads to a better performance in extracting local and global features, which consequently improves tumor classification. The other reason, specially, in the case of vitl_32 is that the network is larger and overfits the data compared to that of vitb_16 and vitb_32 models. Moreover, the vitb_16 model is less computationally complex compared to vitb_32 and vitl_32 models, making it preferable in our case.
Despite the superior performance of the proposed method in all the experiments, it has some disadvantages. As can be observed from the results in Table 4, ViT trained from scratch performed poorly. This is because ViTs are data intensive. ViTs require a large number of images to perform well when trained from scratch, owing to their large number of parameters. The ViT base model had 86 million parameters and the ViT large model had 300 million parameters. It is difficult to train from scratch using hundreds or thousands of ultrasound images, as was the case in our study. Therefore, the model, which was trained from scratch, performed poorly. Another observation made from the experiments was that ViTs are more computationally expensive compared to CNNs. Models using ViTs require more time to train owing to the large size of its architecture. Comparing the results in Table 3 and Table 6 in terms of the time taken for training the models, one can observe that CNNs were mostly faster than ViTs to train. Despite the larger number of parameters and higher computational cost compared with CNNs, our proposed ViTs performed better than CNNs in all cases. The transfer-learning method proposed in this study exhibited a superior performance.
Future work is directed towards optimizing the model using different parameters not included in this paper. Previous studies related to multistage learning methods for ultrasound images have shown the effect of different deep-learning parameters [25,28]. Thus, we will further our efforts in optimizing the proposed model using different deep- learning parameters. Furthermore, we have observed the slight performance difference when using different datasets from different sources, as can be observed in Table 3. The reason for this is obvious and it is due to the difference of imaging equipment, personnel, location, and other related factors. A good deep-learning model should entertain such variations and perform uniformly, irrespective of the source of the dataset. However, this requires the availability of diverse datasets from different locations and in large amounts, which are not currently available. Therefore, our next task will be considering the usage of different datasets across the globe from different locations and training our proposed model. The proposed method could also be translated to breast cancer early diagnosis via other modalities, such as mammogram and magnetic resonance imaging (MRI). Multistage transfer learning using natural images and microscopic images has shown improvement on CNNs’ performance for breast ultrasound image classification [27]. This could also be true for vision transformers’ performance on breast mammogram images. Therefore, the proposed method could be translated to breast cancer early detection using other modalities, such as mammograms.

Author Contributions

Conceptualization, G.A. and S.-w.C.; methodology, G.A. and S.-w.C.; software, G.A. and S.-w.C.; validation, G.A. and S.-w.C.; formal analysis, G.A. and S.-w.C.; investigation, G.A. and S.-w.C.; resources, S.-w.C.; data curation, G.A. and S.-w.C.; writing—original draft preparation, G.A. and S.-w.C.; writing—original draft preparation, G.A. and S.-w.C.; visualization, G.A. and S.-w.C.; supervision, S.-w.C.; project administration, S.-w.C.; funding acquisition, S.-w.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Research Foundation of Korea (NRF) Grant funded by the Korean Government (MSIT) [NRF-2019R1F1A1062397] and project for Industry-University-Research Institute platform cooperation R&D funded by the Korea Ministry of SMEs and Startups in 2022 (S3310765).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

In this study, we used publicly available breast ultrasound images, the Mendeley ultrasound dataset (https://data.mendeley.com/datasets/wmy84gzngw/1 (accessed on 1 July 2022)), and the Breast ultrasound dataset (BUSI) (https://scholar.cu.edu.eg/?q=afahmy/pages/dataset (accessed on 1 July 2022)). Cancer cell line images can be made available upon reasonable requests by contacting the corresponding authors.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the study design; collection, analyses, or interpretation of data; writing of the manuscript; or decision to publish the results.

References

  1. Siegel, R.L.; Miller, K.D.; Fuchs, H.E.; Jemal, A. Cancer Statistics, 2022. CA Cancer J. Clin. 2022, 72, 7–33. [Google Scholar] [CrossRef] [PubMed]
  2. Sung, H.; Ferlay, J.; Siegel, R.L.; Laversanne, M.; Soerjomataram, I.; Jemal, A.; Bray, F. Global Cancer Statistics 2020: GLOBOCAN Estimates of Incidence and Mortality Worldwide for 36 Cancers in 185 Countries. CA Cancer J. Clin. 2021, 71, 209–249. [Google Scholar] [CrossRef] [PubMed]
  3. Siegel, R.L.; Miller, K.D.; Jemal, A. Cancer Statistics, 2020. CA Cancer J. Clin. 2020, 70, 7–30. [Google Scholar] [CrossRef] [PubMed]
  4. Aggarwal, R.; Sounderajah, V.; Martin, G.; Ting, D.S.W.; Karthikesalingam, A.; King, D.; Ashrafian, H.; Darzi, A. Diagnostic Accuracy of Deep Learning in Medical Imaging: A Systematic Review and Meta-Analysis. NPJ Digit. Med. 2021, 4, 65. [Google Scholar] [CrossRef]
  5. Lima, Z.S.; Ebadi, M.R.; Amjad, G.; Younesi, L. Application of Imaging Technologies in Breast Cancer Detection: A Review Article. Open Access Maced. J. Med. Sci. 2019, 7, 838–848. [Google Scholar] [CrossRef] [Green Version]
  6. Hovda, T.; Tsuruda, K.; Hoff, S.R.; Sahlberg, K.K.; Hofvind, S. Radiological Review of Prior Screening Mammograms of Screen-Detected Breast Cancer. Eur. Radiol. 2021, 31, 2568–2579. [Google Scholar] [CrossRef]
  7. Rothschild, J.; Lourenco, A.P.; Mainiero, M.B. Screening Mammography Recall Rate: Does Practice Site Matter? Radiology 2013, 269, 348–353. [Google Scholar] [CrossRef]
  8. Geisel, J.; Raghu, M.; Hooley, R. The Role of Ultrasound in Breast Cancer Screening: The Case for and Against Ultrasound. Semin. Ultrasound CT MRI 2018, 39, 25–34. [Google Scholar] [CrossRef]
  9. Liu, H.; Zhan, H.; Sun, D.; Zhang, Y. Comparison of BSGI, MRI, Mammography, and Ultrasound for the Diagnosis of Breast Lesions and Their Correlations with Specific Molecular Subtypes in Chinese Women. BMC Med. Imaging 2020, 20, 98. [Google Scholar] [CrossRef]
  10. Mimura, T.; Okawa, S.; Kawaguchi, H.; Tanikawa, Y.; Hoshi, Y. Imaging the Human Thyroid Using Three-Dimensional Diffuse Optical Tomography: A Preliminary Study. Appl. Sci. 2021, 11, 1647. [Google Scholar] [CrossRef]
  11. Bene, I.B.; Ciurea, A.I.; Ciortea, C.A.; Dudea, S.M. Pros and Cons for Automated Breast Ultrasound (ABUS): A Narrative Review. J. Pers. Med. 2021, 11, 703. [Google Scholar] [CrossRef]
  12. Ayana, G.; Dese, K.; Raj, H.; Krishnamoorthy, J.; Kwa, T. De-Speckling Breast Cancer Ultrasound Images Using a Rotationally Invariant Block Matching Based Non-Local Means (RIBM-NLM) Method. Diagnostics 2022, 12, 862. [Google Scholar] [CrossRef] [PubMed]
  13. Ayana, G.; Ryu, J. Ultrasound-Responsive Nanocarriers for Breast Cancer Chemotherapy. Micromachines 2022, 13, 1508. [Google Scholar] [CrossRef] [PubMed]
  14. Yuan, W.H.; Hsu, H.C.; Chen, Y.Y.; Wu, C.H. Supplemental Breast Cancer-Screening Ultrasonography in Women with Dense Breasts: A Systematic Review and Meta-Analysis. Br. J. Cancer 2020, 123, 673–688. [Google Scholar] [CrossRef]
  15. Wang, L. Early Diagnosis of Breast Cancer. Sensors 2017, 17, 1572. [Google Scholar] [CrossRef] [Green Version]
  16. The American Cancer Society Medical and Editorial Content Team Breast Cancer Early Detection and Diagnosis. Available online: https://www.cancer.org%7C1.800.227.2345 (accessed on 8 August 2022).
  17. Yap, M.H.; Pons, G.; Marti, J.; Ganau, S.; Sentis, M.; Zwiggelaar, R.; Davison, A.K.; Marti, R. Automated Breast Ultrasound Lesions Detection Using Convolutional Neural Networks. IEEE J. Biomed. Heal. Inform. 2018, 22, 1218–1226. [Google Scholar] [CrossRef] [Green Version]
  18. Seely, J.M.; Alhassan, T. Screening for Breast Cancer in 2018—What Should We Be Doing Today? Curr. Oncol. 2018, 25, S115–S124. [Google Scholar] [CrossRef] [Green Version]
  19. Chougrad, H.; Zouaki, H.; Alheyane, O. Multi-Label Transfer Learning for the Early Diagnosis of Breast Cancer. Neurocomputing 2020, 392, 168–180. [Google Scholar] [CrossRef]
  20. Park, G.E.; Kang, B.J.; Kim, S.H.; Lee, J. Retrospective Review of Missed Cancer Detection and Its Mammography Findings with Artificial-Intelligence-Based, Computer-Aided Diagnosis. Diagnostics 2022, 12, 387. [Google Scholar] [CrossRef]
  21. Mridha, M.F.; Hamid, M.A.; Monowar, M.M.; Keya, A.J.; Ohi, A.Q.; Islam, M.R.; Kim, J.-M. A Comprehensive Survey on Deep-Learning-Based Breast Cancer Diagnosis. Cancers 2021, 13, 6116. [Google Scholar] [CrossRef]
  22. Oyelade, O.N.; Ezugwu, A.E.S. A State-of-the-Art Survey on Deep Learning Methods for Detection of Architectural Distortion from Digital Mammography. IEEE Access 2020, 8, 148644–148676. [Google Scholar] [CrossRef]
  23. Salim, M.; Wåhlin, E.; Dembrower, K.; Azavedo, E.; Foukakis, T.; Liu, Y.; Smith, K.; Eklund, M.; Strand, F. External Evaluation of 3 Commercial Artificial Intelligence Algorithms for Independent Assessment of Screening Mammograms. JAMA Oncol. 2020, 6, 1581–1588. [Google Scholar] [CrossRef] [PubMed]
  24. Murtaza, G.; Shuib, L.; Abdul Wahab, A.W.; Mujtaba, G.; Mujtaba, G.; Nweke, H.F.; Al-garadi, M.A.; Zulfiqar, F.; Raza, G.; Azmi, N.A. Deep Learning-Based Breast Cancer Classification through Medical Imaging Modalities: State of the Art and Research Challenges. Artif. Intell. Rev. 2020, 53, 1655–1720. [Google Scholar] [CrossRef]
  25. Ayana, G.; Dese, K.; Choe, S. Transfer Learning in Breast Cancer Diagnoses via Ultrasound Imaging. Cancers 2021, 13, 738. [Google Scholar] [CrossRef] [PubMed]
  26. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image Is Worth 16 × 16 Words: Transformers for Image Recognition at Scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  27. Ayana, G.; Park, J.; Choe, S.W. Patchless Multi-Stage Transfer Learning for Improved Mammographic Breast Mass Classification. Cancers 2022, 14, 1280. [Google Scholar] [CrossRef] [PubMed]
  28. Ayana, G.; Park, J.; Jeong, J.W.; Choe, S.W. A Novel Multistage Transfer Learning for Ultrasound Breast Cancer Image Classification. Diagnostics 2022, 12, 135. [Google Scholar] [CrossRef]
  29. Cuenat, S.; Couturier, R. Convolutional Neural Network (CNN) vs Vision Transformer (ViT) for Digital Holography. In Proceedings of the 2022 2nd International Conference on Computer, Control and Robotics (ICCCR), Shanghai, China, 18–20 March 2022; pp. 235–240. [Google Scholar] [CrossRef]
  30. Khan, A.; Sohail, A.; Zahoora, U.; Qureshi, A.S. A Survey of the Recent Architectures of Deep Convolutional Neural Networks. Artif. Intell. Rev. 2020, 53, 5455–5516. [Google Scholar] [CrossRef] [Green Version]
  31. Kiranyaz, S.; Avci, O.; Abdeljaber, O.; Ince, T.; Gabbouj, M.; Inman, D.J. 1D Convolutional Neural Networks and Applications: A Survey. Mech. Syst. Signal Process. 2021, 151, 107398. [Google Scholar] [CrossRef]
  32. Al-Dhabyani, W.; Gomaa, M.; Khaled, H.; Fahmy, A. Dataset of Breast Ultrasound Images. Data Br. 2020, 28, 104863. [Google Scholar] [CrossRef]
  33. Rodrigues, P.S. Breast Ultrasound Image. Mendeley Data 2018. [Google Scholar] [CrossRef]
  34. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention Is All You Need. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 6000–6010. [Google Scholar]
Figure 1. Representative images from BUSI dataset: (a) Benign; (b) Malignant; (c) Normal.
Figure 1. Representative images from BUSI dataset: (a) Benign; (b) Malignant; (c) Normal.
Diagnostics 12 02654 g001
Figure 2. Representative images from Mendeley breast ultrasound dataset: (a) benign; (b) malignant.
Figure 2. Representative images from Mendeley breast ultrasound dataset: (a) benign; (b) malignant.
Diagnostics 12 02654 g002
Figure 3. Dataset distribution after preprocessing the BUSI dataset.
Figure 3. Dataset distribution after preprocessing the BUSI dataset.
Diagnostics 12 02654 g003
Figure 4. Acquisition and preprocessing of cancer cell line images: (a) acquired image (b) binarized image (c) patch selection (d) final cancer cell images used for training. MCF-7, breast cancer cell; NCI-60, lung cancer cell; HeLa, cervical cancer cell.
Figure 4. Acquisition and preprocessing of cancer cell line images: (a) acquired image (b) binarized image (c) patch selection (d) final cancer cell images used for training. MCF-7, breast cancer cell; NCI-60, lung cancer cell; HeLa, cervical cancer cell.
Diagnostics 12 02654 g004
Figure 5. The proposed transfer-learning approach using the mixed ultrasound dataset.
Figure 5. The proposed transfer-learning approach using the mixed ultrasound dataset.
Diagnostics 12 02654 g005
Figure 6. The proposed vision transformer-based transfer-learning model. US, ultrasound; MLP, multilayer perceptron; F, flatten; BN, batch normalization; D, dense layer.
Figure 6. The proposed vision transformer-based transfer-learning model. US, ultrasound; MLP, multilayer perceptron; F, flatten; BN, batch normalization; D, dense layer.
Diagnostics 12 02654 g006
Figure 7. Receiver operating characteristics (ROC) curves of the proposed method on Mendeley, BUSI, and mixed datasets. BUSI, breast ultrasound image dataset.
Figure 7. Receiver operating characteristics (ROC) curves of the proposed method on Mendeley, BUSI, and mixed datasets. BUSI, breast ultrasound image dataset.
Diagnostics 12 02654 g007aDiagnostics 12 02654 g007b
Table 1. Vision transformer model details.
Table 1. Vision transformer model details.
ModelLayersHidden SizeMLP SizeHeadsPatch Size
vitb_161276830721216 × 16
vitb_321276830721232 × 32
vitl_3224102440961632 × 32
Table 2. Evaluation metrics.
Table 2. Evaluation metrics.
MetricsFormula
A c c u r a c y T P + T N T P + F P + F N + T N
P r e c i s i o n T P ( T P + F P )
R e c a l l T P ( T P + F N )
F 1   s c o r e T P T P + 1 2 ( F P + F N )
M C C   s c o r e T N × T P F N × F P ( T P + F P ) ( T P + F N ) ( T N + F P ) ( T N + F N )
K a p p a   s c o r e 2 × ( T P × T N F N × F P ) ( T P + F P ) × ( F P + T N ) + ( T P + F N ) × ( F N + T N )
Table 3. The performance of the proposed method: BUSI, breast ultrasound image dataset; LR, learning rate; s, seconds; AUC, area under receiver operating curve.
Table 3. The performance of the proposed method: BUSI, breast ultrasound image dataset; LR, learning rate; s, seconds; AUC, area under receiver operating curve.
ModelOptimizerLRAccuracy (95%)AUC (95%)F1 Score (95%)Time (s) (95%)Loss (95%)
Mendeley
vitb_16Adam0.00011 ± 01 ± 01 ± 0123.35 ± 2.77510.33 ± 0.0037
vitb_32Adam0.00011 ± 01 ± 01 ± 069.85 ± 1.42890.33 ± 0.0053
vitl_32Adam0.00011 ± 01 ± 01 ± 0157.16 ± 6.86670.33 ± 0.0031
BUSI
vitb_16Adam0.00010.952 ± 0.02960.968 ± 0.02320.966 ± 0.0323371.39 ± 3.17180.68 ± 0.0213
vitb_32Adam0.00010.944 ± 0.02420.958 ± 0.01830.942 ± 0.0204294.874 ± 2.28680.72 ± 0.0145
vitl_32Adam0.00010.928 ± 0.03090.947 ± 0.03150.922 ± 0.0333407.074 ± 5.1910.7± 0.0289
Mixed
vitb_16Adam0.00010.919 ± 0.01880.937 ± 0.02560.919 ± 0.0225407.024 ± 1.76520.43 ± 0.0051
vitb_32Adam0.00010.904 ± 0.00680.929 ± 0.00920.904 ± 0.0068265.434 ± 1.84550.45 ± 0.013
vitl_32Adam0.00010.891 ± 0.02910.914 ± 0.3430.894 ± 0.0335448.092 ± 5.02150.44 ± 0.008
Table 4. Performance results of ViT models trained from scratch. BUSI, breast ultrasound image dataset; LR, learning rate; s, seconds; AUC, area under receiver operating curve.
Table 4. Performance results of ViT models trained from scratch. BUSI, breast ultrasound image dataset; LR, learning rate; s, seconds; AUC, area under receiver operating curve.
ModelOptimizerLRAccuracy (95%)AUC (95%)F1 Score (95%)Time (s) (95%)Loss (95%)
Mendeley
vitb_16Adam0.00010.704 ± 0.30.730 ± 0.20.706 ± 0.3123.35 ± 2.77510.33 ± 0.0037
vitb_32Adam0.00010.692 ± 0.10.715 ± 0.20.693 ± 0.269.85 ± 1.42890.33 ± 0.0053
vitl_32Adam0.00010.673 ± 0.10.695 ± 0.30.671 ± 0.2157.16 ± 6.86670.33 ± 0.0031
BUSI
vitb_16Adam0.00010.694 ± 0.02960.710 ± 0.070.693 ± 0.0323371.39 ± 3.17180.68 ± 0.0213
vitb_32Adam0.00010.684 ± 0.02420.698 ± 0.01830.682 ± 0.0204294.874 ± 2.28680.72 ± 0.0145
vitl_32Adam0.00010.669 ± 0.03090.6851 ± 0.03150.667 ± 0.0333407.074 ± 5.1910.7 ± 0.0289
Mixed
vitb_16Adam0.00010.684 ± 0.01880.70 ± 0.1250.689 ± 0.0225407.024 ± 1.76520.43 ± 0.0051
vitb_32Adam0.00010.674 ± 0.00680.699 ± 0.00920.68 ± 0.0068267.434 ± 1.84550.45 ± 0.013
vitl_32Adam0.00010.66 ± 0.02910.689 ± 0.3430.67 ± 0.0335448.092 ± 5.02150.44 ± 0.008
Table 5. Performance of conventional transfer-learning-based ViT models. BUSI, breast ultrasound image dataset; LR, learning rate; s, seconds; AUC, area under receiver operating curve.
Table 5. Performance of conventional transfer-learning-based ViT models. BUSI, breast ultrasound image dataset; LR, learning rate; s, seconds; AUC, area under receiver operating curve.
ModelOptimizerLRAccuracy (95%)AUC (95%)F1 score (95%)Time (s) (95%)Loss (95%)
Mendeley
vitb_16Adam0.00011±01 ± 01 ± 0118.35 ± 2.77510.33 ± 0.0037
vitb_32Adam0.00011±01±01 ± 064.85 ± 1.42890.33 ± 0.0053
vitl_32Adam0.00011±01±01 ± 0152.16 ± 6.86670.33 ± 0.0031
BUSI
vitb_16Adam0.00010.942 ± 0.02960.9541 ± 0.02720.936 ± 0.0323366.39 ± 3.17180.68 ± 0.0213
vitb_32Adam0.00010.934 ± 0.02420.9548 ± 0.01830.932 ± 0.0204289.874 ± 2.28680.72 ± 0.0145
vitl_32Adam0.00010.918 ± 0.03090.9351 ± 0.03150.912 ± 0.0333402.074 ± 5.1910.7 ± 0.0289
Mixed
vitb_16Adam0.00010.914 ± 0.01880.9116 ± 0.01560.909 ± 0.0225402.024 ± 1.76520.43 ± 0.0051
vitb_32Adam0.00010.894 ± 0.00680.8799 ± 0.00920.884 ± 0.0068262.434 ± 1.84550.45 ± 0.013
vitl_32Adam0.00010.86 ± 0.02910.8499 ± 0.3430.844 ± 0.0335443.092 ± 5.02150.44 ± 0.008
Table 6. Performance of CNN-based transfer-learning models. BUSI, breast ultrasound image dataset; LR, learning rate; s, seconds; AUC, area under receiver operating curve.
Table 6. Performance of CNN-based transfer-learning models. BUSI, breast ultrasound image dataset; LR, learning rate; s, seconds; AUC, area under receiver operating curve.
ModelOptimizerLRAccuracy (95%)AUC (95%)F1 score (95%)Time (s) (95%)Loss (95%)
Mendeley
ResNet50Adam0.00010.965 ± 0.020.972 ± 0.010.964 ± 0.02113.35 ± 2.77510.33 ± 0.0037
EfficientNetB2Adam0.00010.961 ± 0.030.969 ± 00.956 ± 0.03115.85 ± 1.42890.33 ± 0.0053
InceptionNetV3Adam0.00010.947 ± 0.10.951 ± 0.010.941 ± 0.02123.16 ± 6.86670.33 ± 0.0031
BUSI
ResNet50Adam0.00010.862 ± 0.10.879 ± 0.20.864 ± 0.2261.39 ± 3.17180.68 ± 0.0213
EfficientNetB2Adam0.00010.851 ± 0.20.864 ± 0.10.856 ± 0.3270.874 ± 2.28680.72 ± 0.0145
InceptionNetV3Adam0.00010.854 ± 0.10.87 ± 0.050.859 ± 0.1272.074 ± 5.1910.7 ± 0.0289
Mixed
ResNet50Adam0.00010.838 ± 0.20.836 ± 0.080.841 ± 0.1302.024 ± 1.76520.43 ± 0.0051
EfficientNetB2Adam0.00010.824 ± 0.10.829 ± 0.10.824 ± 0.2362.434 ± 1.84550.45 ± 0.013
InceptionNetV3Adam0.00010.806 ± 0.070.8199 ± 0.10.804 ± 0.1343.092 ± 5.02150.44 ± 0.008
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Ayana, G.; Choe, S.-w. BUViTNet: Breast Ultrasound Detection via Vision Transformers. Diagnostics 2022, 12, 2654. https://doi.org/10.3390/diagnostics12112654

AMA Style

Ayana G, Choe S-w. BUViTNet: Breast Ultrasound Detection via Vision Transformers. Diagnostics. 2022; 12(11):2654. https://doi.org/10.3390/diagnostics12112654

Chicago/Turabian Style

Ayana, Gelan, and Se-woon Choe. 2022. "BUViTNet: Breast Ultrasound Detection via Vision Transformers" Diagnostics 12, no. 11: 2654. https://doi.org/10.3390/diagnostics12112654

APA Style

Ayana, G., & Choe, S. -w. (2022). BUViTNet: Breast Ultrasound Detection via Vision Transformers. Diagnostics, 12(11), 2654. https://doi.org/10.3390/diagnostics12112654

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop