Next Article in Journal
Vasa Previa and the Role of Fetal Fibronectin and Cervical Length Surveillance: A Review
Next Article in Special Issue
Evaluating the Margins of Breast Cancer Tumors by Using Digital Breast Tomosynthesis with Deep Learning: A Preliminary Assessment
Previous Article in Journal
Late Outcomes of Undiagnosed Unilateral Condylar Hyperplasia and Reoccurrence of Mandibular Asymmetry
Previous Article in Special Issue
Prognostic Molecular Biomarkers in Breast Cancer Lesions with Non-Mass Enhancement on MR
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Breast Delineation in Full-Field Digital Mammography Using the Segment Anything Model

by
Andrés Larroza
1,
Francisco Javier Pérez-Benito
1,
Raquel Tendero
1,
Juan Carlos Perez-Cortes
1,
Marta Román
2 and
Rafael Llobet
1,*
1
Instituto Tecnológico de la Informática, Universitat Politècnica de València, Camino de Vera s/n, 46022 València, Spain
2
Department of Epidemiology and Evaluation, IMIM (Hospital del Mar Research Institute), Passeig Marítim 25-29, 08003 Barcelona, Spain
*
Author to whom correspondence should be addressed.
Diagnostics 2024, 14(10), 1015; https://doi.org/10.3390/diagnostics14101015
Submission received: 17 April 2024 / Revised: 6 May 2024 / Accepted: 13 May 2024 / Published: 15 May 2024
(This article belongs to the Special Issue Advances in Breast Radiology)

Abstract

:
Breast cancer is a major health concern worldwide. Mammography, a cost-effective and accurate tool, is crucial in combating this issue. However, low contrast, noise, and artifacts can limit the diagnostic capabilities of radiologists. Computer-Aided Diagnosis (CAD) systems have been developed to overcome these challenges, with the accurate outlining of the breast being a critical step for further analysis. This study introduces the SAM-breast model, an adaptation of the Segment Anything Model (SAM) for segmenting the breast region in mammograms. This method enhances the delineation of the breast and the exclusion of the pectoral muscle in both medio lateral-oblique (MLO) and cranio-caudal (CC) views. We trained the models using a large, multi-center proprietary dataset of 2492 mammograms. The proposed SAM-breast model achieved the highest overall Dice Similarity Coefficient (DSC) of 99.22% ± 1.13 and Intersection over Union (IoU) 98.48% ± 2.10 over independent test images from five different datasets (two proprietary and three publicly available). The results are consistent across the different datasets, regardless of the vendor or image resolution. Compared with other baseline and deep learning-based methods, the proposed method exhibits enhanced performance. The SAM-breast model demonstrates the power of the SAM to adapt when it is tailored to specific tasks, in this case, the delineation of the breast in mammograms. Comprehensive evaluations across diverse datasets—both private and public—attest to the method’s robustness, flexibility, and generalization capabilities.

1. Introduction

Breast cancer represents a significant global health challenge. Mammography is the primary imaging modality used for breast cancer screening and is associated with a significant decrease in breast cancer mortality. Ongoing technological advancements have yielded improved imaging features, facilitating the early detection of pathological signs associated with the disease [1].
Percent density (PD) refers to the relative amount of fibroglandular tissue within the breast area compared to the overall breast area. Advanced computational methods, including deep learning algorithms, have been employed to improve the accuracy of breast density estimation from mammograms. These methods analyze digital mammograms to classify and segment dense and fatty tissues, providing a quantifiable measure of breast density. This quantification is crucial as breast density is associated with breast cancer risk [2], and its precise assessment can enhance cancer detection and screening effectiveness [3,4].
One fundamental task to estimate the PD is the correct delineation of the breast tissue, excluding other areas that are not of interest that appear in the mammogram, such as background, text, pectoral muscle, and/or abdomen. Once the breast region is isolated, quantitative analysis can be performed to assess various characteristics, such as breast density. Knowledge of the breast contour also allows for further analysis of breast abnormalities such as bilateral asymmetry. Standardizing the breast segmentation process helps ensure consistency and reliability in breast tissue analysis. It is essential in large-scale screening programs where multiple radiologists or automated systems may be involved in interpreting mammograms [5].
Previous breast delineation methods incorporate threshold-based segmentation, morphological manipulation, and a combination of Hough transforms and texture features [6,7]. Sansone et al. [8] analyzed two popular packages that have been proven robust against various situations and were suitable for pectoral muscle removal. Gudhe et al. [3] proposed a new deep learning architecture that automatically estimates the area-based breast percentage density from mammograms using a weight-adaptive multitask learning approach, which improves the baseline traditional methods. The pectoral muscle is mainly seen in medio lateral-oblique (MLO) views. For this reason, most of the available tools are aimed at the pectoral muscle exclusion in MLO views rather than in cranio-caudal (CC) views in which the muscle is sometimes present.
In our previous study [9], we introduced a fully automated framework for dense-tissue segmentation. It included breast detection, pectoral muscle exclusion, and dense-tissue segmentation. The dense-tissue segmentation step was implemented using a Convolutional Neural Network (CNN) architecture named CM-Ynet. However, breast detection and exclusion of the pectoral muscle were carried out using traditional image processing methods. While these methods generally performed well, they encountered challenges in detecting the pectoral muscle in CC images and excluding unwanted regions, such as the abdomen. In response to these challenges, this work introduces a deep learning-based method for breast region segmentation, with a specific focus on enhancing pectoral muscle removal in CC views.
The Segment Anything Model (SAM), released in 2023 by META AI, represents a significant advancement in image segmentation. As the largest model of its kind, the SAM has been trained on billions of labeled images, covering a vast array of object types and scenarios. This extensive training enables the SAM to perform segmentation tasks on nearly any object, regardless of its complexity, without additional specific training. The SAM operates by using prompts or seed points, which can be either positive (to specify areas for segmentation) or negative (to exclude certain areas). Additionally, the SAM can utilize a bounding box approach, segmenting all objects within the specified area [10]. While the SAM is a state-of-the-art research advancement in natural image segmentation, it does not perform satisfactorily when directly applied to medical image segmentation. Recent research suggests that this limitation primarily arises from the significant domain gap between natural images and medical images [11,12].
The main focus of current research is to adapt the SAM to various medical image segmentation tasks. Previous studies have explored different ways to fine-tune the SAM for specific datasets or modalities, such as retinal, abdominal, and cardiac images [13,14,15,16]. Wu et al. [17] proposed the Medical SAM Adapter, which incorporates medical knowledge into the SAM using a simple adapter technique, and tested it on 19 segmentation tasks. More recently, a thorough study on using the SAM for medical 2D images was introduced, distinguished by its extensive data collection, detailed analysis of fine-tuning options, and comprehensive performance evaluation [18]. However, none of these studies addressed the breast segmentation problem.
In this paper, we investigate and evaluate the SAM for breast delineation using a large-scale multi-center dataset that aims to remove all non-breast tissue in MLO and CC views, thereby providing a robust preprocessing method for further mammogram evaluations. We compare our proposed SAM-breast model with traditional and deep learning-based breast segmentation approaches, evaluating their performance on both proprietary and publicly available datasets.

2. Materials and Methods

2.1. Datasets

We used five datasets for this study, two proprietary and three publicly available.
  • GVA (proprietary): A multi-center dataset that covers 11 medical centers of the Generalitat Valenciana (GVA) as part of the Spanish breast cancer screening network. It includes 1785 women with ages from 45 to 70. The CC and MLO views were available for 10 out of 11 of the centers, while one center only collected the CC view. This dataset was used for training, validation, and testing. The dataset was randomly partitioned into 75% (2492 mammograms) for training and validation (10%) and 25% for testing (844 mammograms). All the mammograms are of the type “for presentation”. Further details about this dataset can be found in our previous work [9].
  • IMIM (proprietary): A dataset composed of 881 images obtained at the Hospital del Mar Research Institute (IMIM). It was included solely for testing purposes to better evaluate the generalization performance of the models. This dataset consists of images from three different acquisition devices. One of these devices (Hologic Lorad Selenia, Marlborough, MA, USA) is older and was used to obtain images back in 2012. As a result, the image quality is lower, making the segmentation task more challenging. Only CC views were provided for this dataset, and they are also of the type “for presentation”.
  • CSAW-S (public): The CSAW-S dataset, released by Matsoukas et al. [19], is a companion subset of CSAW [20]. This subset contains mammograms with expert radiologist labels for cancer and complementary labels of breast anatomy made by non-experts. The anonymized dataset contains mammograms from 150 cases of breast cancer, some of them including both MLO and CC views. We generated the breast masks for our experiments by combining the provided mammary gland and pectoral muscle labels, thus obtaining a total of 270 images with breast mask segmentations.
  • InBreast (public): A well-known publicly available dataset [21]. It has ground truth annotations for the pectoral muscle in MLO views. We used these annotations to generate the ground truth breast mask for a total of 200 images.
  • Mini-MIAS (public): The Mini-MIAS database [22], created by the Mammographic Image Analysis Society (MIAS), is a resource that has been extensively used in prior research. It contains 322 digitized mammographic films in MLO view. The breast masks that we used for evaluation were obtained from Verboom et al. [23].
A summary of the datasets is presented in Table 1, showing the number of images per view available in each split. Examples for each dataset are shown in Figure 1.
In the proprietary datasets, ground truth labels were generated by two expert radiologists using the Segment Anything Model (SAM) within CVAT (Computer Vision Annotation Tool), v2.5.1. Consequently, the experts were able to use prompting points to select only the breast area, thereby excluding all unwanted regions. When the SAM failed to provide accurate delineation, conventional CVAT tools were employed for refinement. An example of a labeled mammogram in CVAT is depicted in Figure 2. For the public datasets, the labels were directly extracted from the previously mentioned references.

2.2. Models

In this section, we describe the different segmentation methods that we used to compare with our proposed SAM-breast model.

2.2.1. Thresholding

Thresholding is a simple yet effective method for image segmentation. This technique assigns pixels to the foreground or background based on whether their intensity is above or below a certain threshold. Thresholding methods are widely employed in the delineation of breast boundaries [8].
The thresholding method that we implemented was presented in our previous work and includes two steps: breast detection and pectoral muscle exclusion [9].
In the breast detection step, an iterative algorithm based on connected components is used to obtain the gray level threshold that distinguishes the breast from the background. This is performed by considering the most frequent pixel value as the background and determining a range of possible breast thresholds from all unique values in the image. The image is then binarized using the first possible threshold before applying the scan-plus-array-based union-find (SAUF) algorithm. This process continues until only two homogeneous components are detected.
The pectoral muscle exclusion step assumes that the pectoral muscle appears as a triangle in one of the top corners of the image. An algorithm based on negative gradient changes is implemented, which requires the breast image to be left-oriented. A Gaussian filter with σ = 3 and a 50-pixel moving window is applied to smooth edges and remove isolated bright pixels. The muscle border, being well-defined, tends to be the last remaining part after the smoothing process. A polygon is iteratively built to enclose the exclusion area by selecting the pixel with the lowest gradient every 50 rows until the column of the selected pixel is close enough to the left image border. The first pixel from the top left corner is taken as the vertex that closes the polygon.

2.2.2. MedSegDiff

MedSegDiff-V2 is a novel transformer-based diffusion framework proposed for medical image segmentation. It integrates two cutting-edge techniques, the Diffusion Probabilistic Model (DPM) and vision transformer mechanisms. These techniques, which have been previously applied independently in the field, are combined in MedSegDiff-V2 to harness their collective strengths [24].
We trained MedSegDiff-V2 with the code provided by the original authors using the suggested hyperparameters (https://github.com/KidsWithTokens/MedSegDiff (accessed on 10 January 2024)). We processed the test datasets using five distinct embeddings. The final segmentation mask was then obtained by applying a threshold of 0.5.

2.2.3. SAM

The Segment Anything Model (SAM) is a foundational model for image segmentation. It is designed to handle the promptable segmentation task while addressing real-world constraints. The SAM’s architecture consists of three primary modules: an image encoder, a prompt encoder, and a mask decoder. A single image embedding is produced by the image encoder, while different prompt encoding modules are specifically designed for efficient encoding of various prompt modes. Combining the image embedding with prompt encodings, a lightweight decoder then generates the segmentation masks [10].
In our evaluations, we utilized the ViT-H version of the original SAM. This model was employed with three prompting points, which were randomly placed across the ground segmentation masks in the test set.

2.2.4. SAM-Adapter

Adapters, first used in Natural Language Processing (NLP), allow for efficient fine-tuning of large pre-trained models for specific tasks. Adaptation requires less than 5% of total parameters to be learned, enables quick updates, and is effective not only in NLP but also in computer vision. The Medical SAM Adapter, proposed by Wu et al. [17], is a specialized model designed to adapt the SAM architecture for medical image segmentation. It retains the pre-trained SAM parameters and introduces adapter modules at specific points in the architecture. The adapter module is a bottleneck model that uses a down-projection to compress the embedding to a smaller dimension and an up-projection to expand it back to its original size. The SAM encoder uses two adapters for each ViT block, positioned to optimize the multi-head attention and MLP layer. The SAM decoder utilizes three adapters for each ViT block, with modifications to incorporate prompt information, adjust the MLP-enhanced embedding, and handle the residue connection of the image embedding-to-prompt cross-attention. The final output is produced after adaptation, residue connection, and layer normalization.
We trained the Medical SAM Adapter with the code provided by the original authors (https://github.com/KidsWithTokens/Medical-SAM-Adapter (accessed on 23 November 2023)), using the training strategy described below. At inference, we used three prompting points randomly placed across the ground segmentation masks.

2.2.5. SAM-Breast

We propose the SAM-breast model, which is also based on the original SAM architecture but with modifications to enhance breast delineation. In our model, we chose to freeze the weights of the image encoder. This decision was influenced by the considerable scale of the SAM, particularly since the encoder represents a major portion of the model’s weights. Modifying both the encoder and decoder not only would demand extensive hardware resources but has also been shown to compromise segmentation performance.
We eliminated the prompt encoder, thus removing the need for prompts during both the training and inference phases. The SAM decoder uses both prompt tokens and image embeddings, along with trainable output tokens for generating masks and predicting mask confidence. These output tokens, including foreground and background mask tokens, are combined with prompt tokens to form auxiliary embeddings. The decoder employs a two-way attention module that performs self-attention and cross-attention between tokens and the image embedding. The image embedding is then upscaled and combined with the foreground mask token to produce the mask. Following the method proposed by Hu et al., 2023 [15], we removed the prompt token from the auxiliary embeddings, making it non-promptable. As our objective is solely to predict the breast mask, we did not duplicate the embeddings as conducted in the original paper. Additionally, rather than training the entire mask decoder, we integrated two adapters for each ViT block, following the description by Wu et al., 2023 [17]. Consequently, we fine-tuned only the adapters within the mask decoder while still leveraging the original pre-trained weights. The architecture of SAM-breast is illustrated in Figure 3.

2.2.6. Training Strategy

The following training strategy was implemented for the SAM-adapter and SAM-breast models.
The algorithms were implemented in Pytorch and trained for a maximum of 50 epochs. The epoch with the lowest validation loss was saved and used for test predictions. All the models were trained on an NVIDIA Tesla V100 (Santa Clara, CA, USA) using a batch size of 4 and the Adam optimizer with default parameters. We chose an initial learning rate of 1 × 10 5 , which was divided by 2 whenever the validation loss did not decrease by more than 1 % in one epoch. Two types of loss functions are normally used to train segmentation models: Binary Cross-Entropy (BCE) and the Dice Coefficient [25]. In our preliminary testings, we determined that the Dice loss had better convergence and achieved better results for this specific task.
The input images were resized to 1024 × 1024 pixels. To enhance the model’s generalization capabilities across a wide spectrum of mammograms, we randomly performed on-the-fly data augmentation during training. This process was specifically designed to introduce variations in image intensities and contrast, thereby exposing the model to a broader range of image variations during training. The augmentation process incorporated several transformations. These included random flipping and cropping of images, histogram shifting, and contrast adjustment within a specified range. Additionally, Gaussian noise was added to the images, and the intensity of the images was randomly scaled and shifted. These transformations were carefully chosen to simulate all possible cases that the model may encounter, thereby enhancing its robustness and predictive accuracy.

2.3. Evaluation

To measure the performance of the models, we selected a set of segmentation metrics commonly used for the evaluation of medical image segmentation [26] and used the implementations of the Monai library [27]. The selected metrics are as follows:
  • Dice Similarity Coefficient (DSC): The DSC is a widely used metric for assessing the similarity between the predicted segmentation and the ground truth. It calculates the ratio of twice the intersection of the two segmentations to the sum of the number of pixels in each segmentation. We express this metric as a percentage, ranging from 0 to 100%, with a higher value signifying a superior model performance. The DSC can be expressed as:
    D S C ( X , Y ) = 2 | X Y | | X | + | Y |
    where X and Y denote the binary segmentation and the ground truth, respectively.
  • Intersection over Union (IoU): The IoU, also known as the Jaccard index, measures the overlap between two segmentations. While it bears resemblance to the DSC, its calculation method differs. The IoU can provide a more stringent measure of segmentation performance than the DSC because it penalizes false positives and false negatives equally. It also ranges from 0 to 100%, with higher values indicating better performance. The formula for IoU is given by:
    I o U ( X , Y ) = | X Y | | X Y |
  • Hausdorff Distance (HD): The HD is a metric based on spatial distance. It calculates the maximum distance from a point in one set to the closest point in the other set, allowing for the scoring of localization similarity by focusing on contour delineation. Compared to the overlapping metrics (DSC and IoU), the HD is more sensitive to the boundary. The HD is measured in pixels, with lower values indicating better performance. The formula is given by:
    H D ( X , Y ) = max max x X min y Y d ( x , y ) , max y Y min x X d ( x , y )
    where d ( x , y ) represents the distance between points x and y.
  • Average Surface Distance (ASD): This is another metric used to evaluate the quality of image segmentation. It calculates the average distance from each point on the boundary of the predicted segmentation to the closest point on the boundary of the ground truth segmentation. The ASD is less sensitive to outliers than HD, as it considers the average performance over all points, not just the worst one. It is also measured in pixels, with lower values indicating better performance. The formula for the ASD is given by:
    A S D ( X , Y ) = 1 | X | x X min y Y d ( x , y )
    where | X | is the measure of the surface X and d ( x , y ) is the Euclidean distance between points x and y.

3. Results

All the models described were tested on the different datasets, and a summary of the results is presented in Table 2. Based on these results, our proposed SAM-breast model performs the best for most metrics across all datasets. The only exception is the MedSegDiff model, which shows the best distance metrics (HD and ASD). However, when considering the overall results and visual assessments, SAM-breast shows superior performance. Segmentation examples from each model and dataset are depicted in Figure 4.
In Table 3, we present the results for the images in CC views. This table demonstrates that all models perform exceptionally well in CC views. However, the metrics do not account for the small percentage of pectoral muscle present in the CC views. This is further illustrated in Figure 5, where it is evident that the proposed SAM-breast model successfully excludes the pectoral muscle, even in these images.
The MLO views are significantly more complex due to the presence of the pectoral muscle. Consequently, Table 4 displays lower metrics primarily for the thresholding method and the original SAM.
As outlined earlier, the proprietary datasets utilized in our study were sourced from various centers and acquisition devices. Table 5 presents the results corresponding to each distinct device within the proprietary datasets. This analysis confirms that the metrics obtained are consistent across different acquisition devices.

4. Discussion

We have presented the SAM-breast model for delineating the breast area in digital mammograms. Our results demonstrate the superior performance of the proposed model compared to traditional methods and other deep learning models.
The traditional thresholding technique used in this study has proven to be successful in a large number of scenarios. However, it does not perform as expected when the image contains the abdomen or when the pectoral muscle is included in the CC views [9]. The SAM-breast model effectively addresses this issue by excluding the pectoral muscle in both the MLO and CC views.
The effectiveness of the MedSegDiff model has been thoroughly evaluated across a broad range of medical image segmentation tasks. Specifically, it has been tested on 20 tasks involving different image modalities. This extensive testing underscores the model’s versatility and adaptability to various types of medical imaging data. It has demonstrated a significant improvement in segmentation accuracy compared to previous image segmentation models [24]. However, in our tests, MedSegDiff did not perform as well as SAM-breast. One drawback of the MedSegDiff model is that it takes about two minutes on a single GPU to produce one prediction, while SAM-breast only takes about one second.
The original SAM can generate segmentation masks with impressive speed and quality in most natural images [10]. However, its performance was not optimal on the different mammogram test sets. These results reinforce the necessity of adapting the SAM for specific tasks, especially for medical images.
To compare our results with previously published methodologies, we calculated the accuracy of the SAM-breast model for the Mini-MIAS and InBreast datasets. This was performed in light of the fact that accuracy is frequently employed as an evaluation metric in the majority of studies conducted on these datasets. We present a comparison with this metric in Table 6. The most recent study that aligns with our work is by Zhou et al. [28]. They implemented the Deeplab v3+ model, incorporating preprocessing steps such as noise suppression and contrast enhancement, which resulted in a DSC of 98.48% and IoU of 97.39% for the Mini-MIAS dataset. These values are slightly higher than those achieved by our model on the same dataset (DSC 98.07% and IoU 96.29%). However, it is important to note that their model was trained and tested on different subsets of the Mini-MIAS dataset. In contrast, our proposed model was trained on a proprietary dataset, implying that the public datasets were not seen during the training phase. This is indicative of the robust generalization performance of our model. For example, Zhou et al. also evaluated their model independently on the entire InBreast dataset (comprising 410 images), achieving a DSC of 98.48% and an IoU of 97.14%, whereas our model, tested on 200 MLO images of the InBreast dataset, achieved a DSC of 99.27% and an IoU of 98.55%. Therefore, our model consistently delivers high performance, even when analyzing images with different resolutions and originating from various vendors. We would like to highlight that our approach did not involve any image preprocessing. Instead, our focus was on extensive data augmentation during the training phase. This was achieved by simulating various contrast and intensity variations, thereby enabling us to develop a model that exhibits robust performance across a wide range of scenarios.

5. Conclusions

In this study, we presented the SAM-breast model, a specialized adaptation of the SAM, designed for the precise segmentation of the breast region in mammographic imaging. Our findings indicate that the SAM-breast model proficiently delineates the breast area, effectively excluding the pectoral muscle across both MLO and, notably, CC views where its presence is less frequent. The robustness of our model was validated across diverse datasets sourced from various acquisition devices, showcasing consistent performance throughout. The success of the SAM-breast model underscores the versatility and adaptability of the SAM framework when customized for specific medical imaging tasks, such as the segmentation of breast tissue in mammograms. This advancement holds significant promise for enhancing the accuracy and efficiency of breast cancer screening protocols.

Author Contributions

Conceptualization, A.L., F.J.P.-B. and R.L.; methodology, A.L.; software, A.L.; validation, A.L., F.J.P.-B., M.R. and R.L.; formal analysis, J.C.P.-C. and R.L.; investigation, A.L. and F.J.P.-B.; resources, M.R.; data curation, M.R. and R.T.; writing—original draft preparation, A.L.; writing—review and editing, F.J.P.-B., J.C.P.-C., M.R., R.L. and R.T.; visualization, A.L.; supervision, J.C.P.-C. and R.L.; project administration, J.C.P.-C. and R.L.; funding acquisition, M.R. and J.C.P.-C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Generalitat Valenciana through IVACE and the European Union through FEDER funds, PANAKEIA project IMDEEA/2023/92.

Institutional Review Board Statement

Approval for the usage of IMIM data was granted by the Ethics Committee at IMIM (Approval code: 2017/772/I, 15 November 2017).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the GVA dataset. Patient consent was waived for the IMIM dataset since anonymized retrospective data were used.

Data Availability Statement

The public datasets used in this study can be found at CSAW-S https://zenodo.org/records/4030660#.X2HD15MzZhE (accessed on 7 August 2023), InBreast https://www.kaggle.com/datasets/tommyngx/inbreast2012 (accessed on 30 October 2023), Mini-MIAS http://peipa.essex.ac.uk/info/mias.html (accessed on 11 January 2024), https://zenodo.org/records/10149914 (accessed on 11 January 2024).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Gegios, A.R.; Peterson, M.S.; Fowler, A.M. Breast Cancer Screening and Diagnosis: Recent Advances in Imaging and Current Limitations. PET Clin. 2023, 18, 459–471. [Google Scholar] [CrossRef]
  2. Advani, S.M.; Zhu, W.; Demb, J.; Sprague, B.L.; Onega, T.; Henderson, L.M.; Buist, D.S.M.; Zhang, D.; Schousboe, J.T.; Walter, L.C.; et al. Association of Breast Density with Breast Cancer Risk among Women Aged 65 Years or Older by Age Group and Body Mass Index. JAMA Netw. Open 2021, 4, e2122810. [Google Scholar] [CrossRef] [PubMed]
  3. Gudhe, N.R.; Behravan, H.; Sudah, M.; Okuma, H.; Vanninen, R.; Kosma, V.M.; Mannermaa, A. Area-based breast percentage density estimation in mammograms using weight-adaptive multitask learning. Sci. Rep. 2022, 12, 12060. [Google Scholar] [CrossRef]
  4. Lopez-Almazan, H.; Pérez-Benito, F.J.; Larroza, A.; Perez-Cortes, J.C.; Pollan, M.; Perez-Gomez, B.; Trejo, D.S.; Casals, M.; Llobet, R. A deep learning framework to classify breast density with noisy labels regularization. Comput. Methods Programs Biomed. 2022, 221, 106885. [Google Scholar] [CrossRef]
  5. Michael, E.; Ma, H.; Li, H.; Kulwa, F.; Li, J. Breast cancer segmentation methods: Current status and future potentials. Biomed Res. Int. 2021, 2021, 9962109. [Google Scholar] [CrossRef]
  6. Hazarika, M.; Mahanta, L.B. A New Breast Border Extraction and Contrast Enhancement Technique with Digital Mammogram Images for Improved Detection of Breast Cancer. Asian Pac. J. Cancer Prev. 2018, 19, 2141–2148. [Google Scholar] [PubMed]
  7. Bora, V.B.; Kothari, A.G.; Keskar, A.G. Robust automatic pectoral muscle segmentation from mammograms using texture gradient and Euclidean Distance Regression. J. Digit. Imaging 2016, 29, 115–125. [Google Scholar] [CrossRef]
  8. Sansone, M.; Marrone, S.; Di Salvio, G.; Belfiore, M.P.; Gatta, G.; Fusco, R.; Vanore, L.; Zuiani, C.; Grassi, F.; Vietri, M.T.; et al. Comparison between two packages for pectoral muscle removal on mammographic images. Radiol. Medica 2022, 127, 848–856. [Google Scholar] [CrossRef] [PubMed]
  9. Larroza, A.; Pérez-Benito, F.J.; Perez-Cortes, J.C.; Román, M.; Pollán, M.; Pérez-Gómez, B.; Salas-Trejo, D.; Casals, M.; Llobet, R. Breast Dense Tissue Segmentation with Noisy Labels: A Hybrid Threshold-Based and Mask-Based Approach. Diagnostics 2022, 12, 1822. [Google Scholar] [CrossRef]
  10. Kirillov, A.; Mintun, E.; Ravi, N.; Mao, H.; Rolland, C.; Gustafson, L.; Xiao, T.; Whitehead, S.; Berg, A.C.; Lo, W.Y.; et al. Segment Anything. arXiv 2023, arXiv:2304.02643. [Google Scholar]
  11. Mazurowski, M.A.; Dong, H.; Gu, H.; Yang, J.; Konz, N.; Zhang, Y. Segment anything model for medical image analysis: An experimental study. Med. Image Anal. 2023, 89, 102918. [Google Scholar] [CrossRef] [PubMed]
  12. He, S.; Bao, R.; Li, J.; Stout, J.; Bjornerud, A.; Grant, P.E.; Ou, Y. Computer-Vision Benchmark Segment-Anything Model (SAM) in Medical Images: Accuracy in 12 Datasets. arXiv 2023, arXiv:2304.09324. [Google Scholar]
  13. Deng, G.; Zou, K.; Ren, K.; Wang, M.; Yuan, X.; Ying, S.; Fu, H. SAM-U: Multi-box prompts triggered uncertainty estimation for reliable SAM in medical image. arXiv 2023, arXiv:2307.04973. [Google Scholar]
  14. Zhang, K.; Liu, D. Customized segment anything model for medical image segmentation. arXiv 2023, arXiv:2304.13785. [Google Scholar]
  15. Hu, X.; Xu, X.; Shi, Y. How to Efficiently Adapt Large Segmentation Model (SAM) to Medical Images. arXiv 2023, arXiv:2306.13731. [Google Scholar]
  16. Ma, J.; Wang, B. Segment anything in medical images. arXiv 2023, arXiv:2304.12306. [Google Scholar] [CrossRef] [PubMed]
  17. Wu, J.; Zhang, Y.; Fu, R.; Fang, H.; Liu, Y.; Wang, Z.; Xu, Y.; Jin, Y. Medical SAM Adapter: Adapting Segment Anything Model for Medical Image Segmentation. arXiv 2023, arXiv:2304.12620. [Google Scholar]
  18. Cheng, J.; Ye, J.; Deng, Z.; Chen, J.; Li, T.; Wang, H.; Su, Y.; Huang, Z.; Chen, J.; Jiang, L.; et al. SAM-Med2D. arXiv 2023, arXiv:2308.16184. [Google Scholar]
  19. Matsoukas, C.; Hernandez, A.B.I.; Liu, Y.; Dembrower, K.; Miranda, G.; Konuk, E.; Haslum, J.F.; Zouzos, A.; Lindholm, P.; Strand, F.; et al. Adding Seemingly Uninformative Labels Helps in Low Data Regimes. arXiv 2020, arXiv:2008.00807. [Google Scholar]
  20. Dembrower, K.; Lindholm, P.; Strand, F. A multi-million mammography image dataset and population-based screening cohort for the training and evaluation of deep neural networks—The cohort of Screen-Aged Women (CSAW). J. Digit. Imaging 2020, 33, 408–413. [Google Scholar] [CrossRef]
  21. Moreira, I.C.; Amaral, I.; Domingues, I.; Cardoso, A.; Cardoso, M.J.; Cardoso, J.S. INbreast: Toward a Full-field Digital Mammographic Database. Acad. Radiol. 2012, 19, 236–248. [Google Scholar] [CrossRef] [PubMed]
  22. Suckling, J.; Parker, J.; Dance, D.; Astley, S.; Hutt, I. Mammographic Image Analysis Society (MIAS) Database v1.21. Apollo—University of Cambridge Repository. 2015. Available online: http://peipa.essex.ac.uk/info/mias.html (accessed on 11 January 2024).
  23. Verboom, S.D.; Caballo, M.; Peters, J.; Gommers, J.; van den Oever, D.; Broeders, M.; Teuwen, J.; Sechopoulos, I. Segmentation Masks Mini-MIAS [Data Set]. 2023. Available online: https://zenodo.org/records/10149914 (accessed on 11 January 2024).
  24. Wu, J.; Ji, W.; Fu, H.; Xu, M.; Jin, Y.; Xu, Y. MedSegDiff-V2: Diffusion based Medical Image Segmentation with Transformer. Proc. Aaai Conf. Artif. Intell. 2024, 38, 6030–6038. [Google Scholar] [CrossRef]
  25. Jadon, S. A survey of loss functions for semantic segmentation. In Proceedings of the 2020 IEEE Conference on Computational Intelligence in Bioinformatics and Computational Biology (CIBCB), Via del Mar, Chile, 27–29 October 2020; pp. 1–7. [Google Scholar]
  26. Müller, D.; Soto-Rey, I.; Kramer, F. Towards a guideline for evaluation metrics in medical image segmentation. BMC Res. Notes 2022, 15, 1–11. [Google Scholar] [CrossRef] [PubMed]
  27. Maintainers, P.M. MONAI: A PyTorch-based, open-source framework for deep learning in healthcare imaging. arXiv 2020, arXiv:2003.05597. [Google Scholar]
  28. Zhou, K.; Li, W.; Zhao, D. Deep learning-based breast region extraction of mammographic images combining pre-processing methods and semantic segmentation supported by Deeplab v3+. Technol. Health Care 2022, 30, S173–S190. [Google Scholar] [CrossRef] [PubMed]
  29. Lbachir, I.A.; Es-Salhi, R.; Daoudi, I.; Tallal, S. A New Mammogram Preprocessing Method for Computer-Aided Diagnosis Systems. In Proceedings of the 2017 IEEE/ACS 14th International Conference on Computer Systems and Applications (AICCSA), Hammamet, Tunisia, 30 October–3 November 2017; pp. 166–171. [Google Scholar] [CrossRef]
  30. Taghanaki, S.A.; Liu, Y.; Miles, B.; Hamarneh, G. Geometry-Based Pectoral Muscle Segmentation from MLO Mammogram Views. IEEE Trans. Biomed. Eng. 2017, 64, 2662–2671. [Google Scholar] [CrossRef]
  31. Rahman, M.A.; Jha, R.K.; Gupta, A.K. Gabor phase response based scheme for accurate pectoral muscle boundary detection. IET Image Process. 2019, 13, 771–778. [Google Scholar] [CrossRef]
Figure 1. Sample images for each dataset. The white arrows indicate the pectoral muscle. Note the presence of the pectoral muscle in the cranio-caudal (CC) view of the IMIM dataset (second image).
Figure 1. Sample images for each dataset. The white arrows indicate the pectoral muscle. Note the presence of the pectoral muscle in the cranio-caudal (CC) view of the IMIM dataset (second image).
Diagnostics 14 01015 g001
Figure 2. Main screen of CVAT, v2.5.1. Interaction with SAM is demonstrated through labeling with three positive (green) points and one negative (red) point.
Figure 2. Main screen of CVAT, v2.5.1. Interaction with SAM is demonstrated through labeling with three positive (green) points and one negative (red) point.
Diagnostics 14 01015 g002
Figure 3. The architecture of SAM-breast. The encoder and layers of the mask decoder are frozen. Only the adapters incorporated in each ViT block are learned during training.
Figure 3. The architecture of SAM-breast. The encoder and layers of the mask decoder are frozen. Only the adapters incorporated in each ViT block are learned during training.
Diagnostics 14 01015 g003
Figure 4. Representative examples for each dataset and model tested. The prompting points used for the SAM and SAM-adapter models are also indicated as pink stars. Note the correct exclusion of the pectoral muscle in the CC view image (second row).
Figure 4. Representative examples for each dataset and model tested. The prompting points used for the SAM and SAM-adapter models are also indicated as pink stars. Note the correct exclusion of the pectoral muscle in the CC view image (second row).
Diagnostics 14 01015 g004
Figure 5. Cranio-caudal (CC) examples where the pectoral muscle is present. The prompting points used for the SAM and SAM-adapter models are also indicated as pink stars.
Figure 5. Cranio-caudal (CC) examples where the pectoral muscle is present. The prompting points used for the SAM and SAM-adapter models are also indicated as pink stars.
Diagnostics 14 01015 g005
Table 1. Summary of the datasets used for training and testing.
Table 1. Summary of the datasets used for training and testing.
DatasetSplit#MLO#CCTotal
GVATrain105211902242
GVAValidation111139250
GVATest395449844
IMIMTest0881881
CSAW-STest126144270
InBreastTest2000200
Mini-MIASTest3220322
MLO: medio lateral-oblique. CC: cranio-caudal.
Table 2. A summary of the results for each dataset.
Table 2. A summary of the results for each dataset.
DatasetModelDSCIoUHDASD
Thresholding95.81 ± 4.8092.31 ± 7.58101.87 ± 91.4410.21 ± 16.66
MedSegDiff98.40 ± 1.1896.88 ± 2.175.96 ± 7.580.96 ± 0.89
GVASAM95.32 ± 5.7391.60 ± 9.80111.70 ± 130.1219.78 ± 40.17
SAM-adapter99.21 ± 0.8798.45 ± 1.6520.88 ± 27.951.85 ± 2.21
SAM-breast99.37 ± 0.8698.76 ± 1.6118.72 ± 30.191.64 ± 2.95
Thresholding97.14 ± 8.0095.32 ± 11.2458.20 ± 100.5213.88 ± 45.56
MedSegDiff98.91 ± 1.4597.87 ± 2.595.77 ± 9.140.77 ± 1.12
IMIMSAM99.22 ± 1.0498.47 ± 1.9542.10 ± 69.114.14 ± 15.83
SAM-adapter99.49 ± 0.5998.98 ± 1.1418.20 ± 21.841.45 ± 1.90
SAM-breast99.54 ± 0.5299.10 ± 115.09 ± 19.311.27 ± 1.74
Thresholding92.48 ± 8.3686.94 ± 12.30152.50 ± 104.6323.04 ± 21.50
MedSegDiff97.19 ± 3.9894.80 ± 6.7415.26 ± 23.052.83 ± 7.06
CSAW-SSAM94.26 ± 6.7089.85 ± 11.10129.10 ± 126.8720.39 ± 29.58
SAM-adapter98.72 ± 1.4197.52 ± 2.6058.15 ± 75.963.21 ± 3.81
SAM-breast99.10 ± 0.9698.22 ± 1.8351.89 ± 75.162.54 ± 3.22
Thresholding94.70 ± 2.7490.06 ± 4.77146.33 ± 72.5312.66 ± 7.02
MedSegDiff98.37 ± 1.2396.82 ± 2.307.56 ± 8.271.10 ± 0.88
InBreastSAM92.04 ± 5.1685.65 ± 8.48170.89 ± 71.2128.77 ± 16.92
SAM-adapter99.05 ± 1.1898.13 ± 2.1530.98 ± 36.242.59 ± 3.29
SAM-breast99.27 ± 0.8598.55 ± 1.6225.37 ± 32.762.08 ± 2.42
Thresholding81.30 ± 7.3669.11 ± 10.05187.58 ± 57.4652.02 ± 13.64
MedSegDiff87.81 ± 6.5978.85 ± 9.9846.18 ± 16.2410.79 ± 5.38
Mini-MIASSAM89.04 ± 6.2480.78 ± 9.61199.71 ± 79.2040.45 ± 26.70
SAM-adapter95.35 ± 4.3791.43 ± 7.4381.66 ± 60.2613.68 ± 12.47
SAM-breast98.07 ± 2.1196.29 ± 3.8246.06 ± 41.706.55 ± 8.13
DSC: Dice Similarity Coefficient. IoU: Intersection over Union. HD: Hausdorff Distance. ASD: Average Surface Distance. Values are shown as mean ± standard deviation. The best result for each evaluation metric and dataset is highlighted in bold.
Table 3. Results for images in CC view.
Table 3. Results for images in CC view.
DatasetModelDSCIoUHDASD
Thresholding98.17 ± 3.6596.59 ± 4.8842.57 ± 56.906.14 ± 19.33
MedSegDiff98.73 ± 0.6897.49 ± 1.314.70 ± 6.960.77 ± 0.50
GVASAM99.25 ± 1.3998.54 ± 2.4245.80 ± 104.896.73 ± 36.58
SAM-adapter99.49 ± 0.4898.99 ± 0.9415.02 ± 24.521.21 ± 1.31
SAM-breast99.60 ± 0.4299.21 ± 0.8212.96 ± 23.970.96 ± 1.08
Thresholding97.14 ± 8.0095.32 ± 11.2458.20 ± 100.5213.88 ± 45.56
MedSegDiff98.91 ± 1.4597.87 ± 2.595.77 ± 9.140.77 ± 1.12
IMIMSAM99.22 ± 1.0498.47 ± 1.9542.10 ± 69.114.14 ± 15.83
SAM-adapter99.49 ± 0.5998.98 ± 1.1418.20 ± 21.841.45 ± 1.90
SAM-breast99.54 ± 0.5299.10 ± 115.09 ± 19.311.27 ± 1.74
Thresholding89.72 ± 9.1482.37 ± 12.57193.76 ± 99.2433.17 ± 21.63
MedSegDiff97.18 ± 4.0094.76 ± 6.6616.13 ± 21.392.59 ± 5.91
CSAW-SSAM91.59 ± 7.0585.21 ± 11.37167.77 ± 108.2426.39 ± 22.34
SAM-adapter98.56 ± 1.3997.20 ± 2.5479.14 ± 86.933.53 ± 3.71
SAM-breast98.99 ± 0.8098.02 ± 1.5371.05 ± 84.742.86 ± 3.23
DSC: Dice Similarity Coefficient. IoU: Intersection over Union. HD: Hausdorff Distance. ASD: Average Surface Distance. Values are shown as mean ± standard deviation. The best result for each evaluation metric and dataset is highlighted in bold.
Table 4. Results for images in MLO view.
Table 4. Results for images in MLO view.
DatasetModelDSCIoUHDASD
Thresholding93.13 ± 4.5487.44 ± 7.15169.28 ± 75.1314.84 ± 11.33
MedSegDiff98.03 ± 1.4996.18 ± 2.697.38 ± 8.011.17 ± 1.15
GVASAM90.86 ± 5.5383.70 ± 9.03186.62 ± 114.6434.62 ± 38.94
SAM-adapter98.90 ± 1.0897.84 ± 2.0327.56 ± 30.072.57 ± 2.74
SAM-breast99.11 ± 1.1198.25 ± 2.0725.89 ± 34.872.40 ± 4.02
Thresholding95.62 ± 6.0192.16 ± 9.66105.36 ± 89.9511.47 ± 14.36
MedSegDiff97.22 ± 3.9794.84 ± 6.8614.25 ± 24.873.11 ± 8.19
CSAW-SSAM97.32 ± 4.6895.15 ± 8.0384.91 ± 132.4613.53 ± 34.98
SAM-adapter98.91 ± 1.4197.88 ± 2.6234.15 ± 51.812.83 ± 3.90
SAM-breast99.21 ± 1.1198.46 ± 2.0930 ± 55.112.17 ± 3.17
Thresholding94.70 ± 2.7490.06 ± 4.77146.33 ± 72.5312.66 ± 7.02
MedSegDiff98.37 ± 1.2396.82 ± 2.307.56 ± 8.271.10 ± 0.88
InBreastSAM92.04 ± 5.1685.65 ± 8.48170.89 ± 71.2128.77 ± 16.92
SAM-adapter99.05 ± 1.1898.13 ± 2.1530.98 ± 36.242.59 ± 3.29
SAM-breast99.27 ± 0.8598.55 ± 1.6225.37 ± 32.762.08 ± 2.42
Thresholding81.30 ± 7.3669.11 ± 10.05187.58 ± 57.4652.02 ± 13.64
MedSegDiff87.81 ± 6.5978.85 ± 9.9846.18 ± 16.2410.79 ± 5.38
Mini-MIASSAM89.04 ± 6.2480.78 ± 9.61199.71 ± 79.2040.45 ± 26.70
SAM-adapter95.34 ± 4.3891.41 ± 7.4481.86 ± 60.2413.71 ± 12.47
SAM-breast98.07 ± 2.1196.29 ± 3.8246.06 ± 41.106.55 ± 8.13
DSC: Dice Similarity Coefficient. IoU: Intersection over Union. HD: Hausdorff Distance. ASD: Average Surface Distance. Values are shown as mean ± standard deviation. The best result for each evaluation metric and dataset is highlighted in bold.
Table 5. Results for different devices using the proposed SAM-breast model.
Table 5. Results for different devices using the proposed SAM-breast model.
DatasetDeviceDSCIoUHDASD
01-FUJIFILM99.12 ± 1.6498.30 ± 2.9819.98 ± 26.201.94 ± 3.07
02-FUJIFILM99.59 ± 0.4499.19 ± 0.8610.41 ± 14.680.96 ± 1.07
04-IMS Giotto99.37 ± 1.0898.78 ± 2.0312.85 ± 10.221.25 ± 1.32
05-FUJIFILM99.55 ± 0.3499.10 ± 0.6713.98 ± 26.481.05 ± 1.06
07-HOLOGIC99.27 ± 1.1298.58 ± 2.1427.21 ± 62.353.95 ± 10.97
GVA10-SIEMENS99.06 ± 1.2698.16 ± 2.3327.94 ± 33.872.48 ± 4.33
11-FUJIFILM99.44 ± 0.5498.90 ± 1.0514.40 ± 24.221.37 ± 1.94
13-FUJIFILM99.43 ± 0.8498.87 ± 1.5919.72 ± 40.301.54 ± 3.47
18-IMS Giotto99.14 ± 0.6798.30 ± 1.3019.43 ± 22.832.16 ± 2.02
20-Unknown99.13 ± 0.7398.29 ± 1.4128.27 ± 41.292.04 ± 2.34
99-GE99.46 ± 0.3898.93 ± 0.7617.72 ± 32.021.52 ± 1.10
FUJIFILM99.34 ± 0.4798.69 ± 0.9216.25 ± 18.371.86 ± 1.55
IMIMGE99.28 ± 0.7198.59 ± 1.3823.98 ± 30.831.97 ± 2.91
HOLOGIC99.31 ± 0.5598.64 ± 1.0514.70 ± 17.751.71 ± 1.61
DSC: Dice Similarity Coefficient. IoU: Intersection over Union. HD: Hausdorff Distance. ASD: Average Surface Distance. Values are shown as mean ± standard deviation. FUJIFILM (Tokyo, Japan), IMS Giotto (Sasso Marconi, Italy), HOLOGIC (Marlborough, MA, USA), SIEMENS (Munich, Germany), GE (Boston, MA, USA).
Table 6. A comparison of the accuracy of various published methods.
Table 6. A comparison of the accuracy of various published methods.
DatasetMethodYearImagesAccuracy
Lbachir et al. [29]20174090%
Taghanaki et al. [30]201719796%
InBreastRahman et al. [31]201920094.50%
Zhou et al. [28]202241099.12%
SAM-breast (proposed)202420099.56%
Lbchir et al. [29]201732298.75%
Taghanaki et al. [30]201732295%
Mini-MIASRahman et al. [31]201920097.50%
Zhou et al. [28]202210298.98%
SAM-breast (proposed)202432298.76%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Larroza, A.; Pérez-Benito, F.J.; Tendero, R.; Perez-Cortes, J.C.; Román, M.; Llobet, R. Breast Delineation in Full-Field Digital Mammography Using the Segment Anything Model. Diagnostics 2024, 14, 1015. https://doi.org/10.3390/diagnostics14101015

AMA Style

Larroza A, Pérez-Benito FJ, Tendero R, Perez-Cortes JC, Román M, Llobet R. Breast Delineation in Full-Field Digital Mammography Using the Segment Anything Model. Diagnostics. 2024; 14(10):1015. https://doi.org/10.3390/diagnostics14101015

Chicago/Turabian Style

Larroza, Andrés, Francisco Javier Pérez-Benito, Raquel Tendero, Juan Carlos Perez-Cortes, Marta Román, and Rafael Llobet. 2024. "Breast Delineation in Full-Field Digital Mammography Using the Segment Anything Model" Diagnostics 14, no. 10: 1015. https://doi.org/10.3390/diagnostics14101015

APA Style

Larroza, A., Pérez-Benito, F. J., Tendero, R., Perez-Cortes, J. C., Román, M., & Llobet, R. (2024). Breast Delineation in Full-Field Digital Mammography Using the Segment Anything Model. Diagnostics, 14(10), 1015. https://doi.org/10.3390/diagnostics14101015

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop