Next Article in Journal
CPM-GFDM: A Novel Combination of Continuous Phase Modulation and Generalized Frequency Division Multiplexing for Wireless Communication
Next Article in Special Issue
Rainfall Similarity Search Based on Deep Learning by Using Precipitation Images
Previous Article in Journal
Study on the Median Opening Length of a Freeway Work Zone Based on a Naturalistic Driving Experiment
Previous Article in Special Issue
An Adaptive Dynamic Channel Allocation Algorithm Based on a Temporal–Spatial Correlation Analysis for LEO Satellite Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Attention-Based Active Learning Framework for Segmentation of Breast Cancer in Mammograms

1
School of Artificial Intelligence, Zhejiang College of Security Technology, Wenzhou 325000, China
2
College of Computer and Information, Hohai University, Nanjing 210093, China
3
Key Laboratory of Symbolic Computation and Knowledge Engineering of Ministry of Education, Jilin University, Changchun 130015, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(2), 852; https://doi.org/10.3390/app13020852
Submission received: 5 December 2022 / Revised: 30 December 2022 / Accepted: 6 January 2023 / Published: 7 January 2023
(This article belongs to the Special Issue Deep Learning and Edge Computing for Internet of Things)

Abstract

:
Breast cancer is one of most serious malignant tumors that affect women’s health. To carry out the early screening of breast cancer, mammography provides breast cancer images for doctors’ efficient diagnosis. However, breast cancer lumps can vary in size and shape, bringing difficulties for the accurate recognition of both humans and machines. Moreover, the annotation of such images requires expert medical knowledge, which increases the cost of collecting datasets to boost the performance of deep learning methods. To alleviate these problems, we propose an attention-based active learning framework for breast cancer segmentation in mammograms; the framework consists of a basic breast cancer segmentation model, an attention-based sampling scheme and an active learning strategy for labelling. The basic segmentation model performs multi-scale feature fusion and enhancement on the basis of UNet, thus improving the distinguishing representation capability of the extracted features for further segmentation. Afterwards, the proposed attention-based sampling scheme assigns different weights for unlabeled breast cancer images by evaluating their uncertainty with the basic segmentation model. Finally, the active learning strategy selects unlabeled images with the highest weights for manual labeling, thus boosting the performance of the basic segmentation model via retraining with new labeled samples. Testing on four datasets, experimental results show that the proposed framework could greatly improve segmentation accuracy by about 15% compared with an existing method, while largely decreasing the cost of data annotation.

1. Introduction

Breast cancer is one of the most serious malignant tumors that threatens the health of women. It is reported that about 12.5 percent of women are affected by breast cancer worldwide [1]. In China, the incidence of breast cancer is increasing by 0.5% per year [2], which has become one of the most dangerous and deadly diseases for women’s health. Early-stage breast cancer screening can aid in detecting disease early, thus largely increasing the chances of recovery [3]. These screenings are conducted by experienced doctors, who check the existence of malignant lesions in mammograms for further diagnosis and evaluation. However, such operations are not only annoying and time-consuming, but also carry the risk of wrong or missed detections with manual examinations [4], which is reported to be as high as 30% in breast cancer screening [5]. Since the advent of artificial intelligence (AI), researchers have adopted computer-aided detection (CAD) technologies to considerably reduce the amount of work involved in breast cancer screening. In the early stages of this application, CAD systems with traditional image processing algorithms were applied to aid in breast cancer diagnosis (masses, microcalcification etc.) [6]. With the rapid development of deep learning, the accuracy of medical image segmentation, regarded as a subfield of image segmentation, has been significantly improved. For example, Wang et al. [7] proposed a breast tumor semantic segmentation method based on a convolutional neural network (CNN), which provided accurate results with the extracted biomarkers in radiology imaging. Later, Hann et al. [8] utilized a multiple UNet network to generate multiple segmentations for fusion, which were further thresholded to generate the final segmentation result.
However, these impressive achievements of deep learning in breast cancer screening were built on a large amount of labeled data, and accuracy would greatly drop without it [9]. Due to privacy issues, it is difficult to obtain sufficient breast cancer images for labeling. Moreover, labeling generally requires medical experts with professional knowledge and experience, making it time-consuming and costly to acquire enough labeled data. Achieving desirable segmentation results with fewer samples has thus recently become a research focus.
Benefiting from its ability to achieve high performance with few labeled samples, active learning has become a feasible approach for less-thoroughly labeled breast cancer data [10]. Several successful applications have been implemented to show the remarkable power of active learning in medical image analysis. For example, Vishwesh et al. [11] combined an active learning framework with deep learning for medical image segmentation, where new active learning strategies guided the segmentation model to learn the diversity of uncertain and unlabeled data, thus greatly achieving convergence in accuracy with less labelling. Later, Li et al. [12] proposed a novel active learning framework for histopathology image analysis, where two groups of unlabeled data were selected in each training iteration, one annotated by experts and the other selected from high-confidence unlabeled samples to assign pseudo-labels. Both manual labeling and pseudo-label generation Were able to largely alleviate the problem of scarce labeled samples.
Based on the advantages of using deep learning and active learning for automatic screening tasks, we propose an attention-based active breast cancer segmentation model which is capable of achieving desirable segmentation results without a high quantity of labeled images. The proposed model consists of a basic segmentation model, an attention-based sampling scheme and an active learning based labeling strategy. Specifically, a multi-scale fusion and enhancement module based on UNet is first adopted for segmentation. Afterwards, a novel attention mechanism is used to evaluate the similarity between the unlabeled and segmented samples, thus offering weights as criteria to measure the uncertainty or informativeness of unlabeled samples with respect to the trained and basic segmentation model. Finally, an active learning strategy is used to sort unlabeled samples with weights, thus determining selections that need to be further manually labelled. We observed that these selected samples often contained appearance or shape features which were unacknowledged by the basic segmentation model subjected to the current training dataset. With iterations of retraining with the most informative unlabeled samples, the proposed model stably approached the upper bound of segmentation performance for high accuracy.
The contributions of this paper are summarized as follows:
  • We propose an attention-based active breast cancer segmentation framework which effectively improved the accuracy of segmentation with few training samples, thus alleviating the high cost of labeling breast cancer images.
  • A novel attention-based sampling scheme is proposed which measures the most informative unlabeled samples via calculating similarity weights.
  • We adopt an active learning strategy for the global optimization of accuracy performance, which iteratively selects appropriate unlabeled samples to first manually label and then retrain the model to boost performance.
The rest of this paper is organized as follows. Section 2 reviews related work on the segmentation of breast cancer images. Section 3 presents an overview of the proposed method. Details of the basic segmentation model, attention-based sampling scheme and active-learning-based labeling strategy are also discussed in Section 3. Section 4 presents and discusses the experimental results. Finally, Section 5 concludes the paper.

2. Related Work

In this section, we give a brief literature review, including prior works on traditional segmentation methods for breast cancer images, deep-learning-based segmentation methods for breast cancer images, as well as active learning methods.

2.1. Traditional Segmentation Methods for Breast Cancer Images

Traditional segmentation methods usually apply image processing technologies for segmentation. However, they generally suffer from the drawbacks of low accuracy and are sensitive to the quality of sampled images.
For example, Cheng et al. [13] proposed a near-automatic ultrasound image segmentation algorithm which builds a solid foundation of computer-aided diagnosis for breast cancer. Later, Eziddin et al. [14] proposed the segmentation of mammograms using an iterative fusion process of information obtained from multiple knowledge sources, including context information, image processing algorithms, prior knowledge and so on. Gnonnou et al. [15] proposed a structural method to separate breast margins at pixel-level, thus accurately extracting tumor regions. Later, Kaushal et al. [16] proposed an automated segmentation technique followed by self-driven post-processing operations to detect cancerous cells effectively. Recently, Jing et al. [17] proposed a simple but effective segmentation method with the concept of global thresholding, which successfully segmented tumor regions in breast histopathology images. During the process, partial contrast stretching and median filtering are specially designed to improve image quality for segmentation.

2.2. Deep-Learning-Based Segmentation Methods for Breast Cancer Images

Inspired by the remarkable performance of deep learning methods in image classification and segmentation tasks [18,19], researchers have proposed several works on the segmentation of breast cancer images with various kinds of networks.
For example, Su et al. [20] proposed a fast scanning deep convolutional neural network (FCNN) to achieve pixel-wise region segmentation, successfully eliminating the redundant computation of the original CNN without sacrificing performance. Later, Simin et al. [21] proposed the combination of deep learning with traditional features for medical image classification, where a CNN model is first used to extract image features, and then support vector machines are used for feature learning and classification. Then, Roy et al. [22] used a dropout strategy to generate different Monte Carlo segmentations, where they computed the dissimilarity of these segmentations to measure the structural uncertainty of the image. In such way, they could confidently choose the best matching segmentations from candidates for output.
Despite their strengths, the shortage of sufficient training data affects the performance of deep learning methods, and researchers have focused on active learning to improve the effectiveness of deep learning. For example, Shen et al. [23] proposed a novel deep active learning model for the image segmentation of breast cancer on immunohistochemistry images. They not only achieved significant performance improvements in the segmentation of breast cancer images, but the system also showed promise for implementation as a real-world application.
Recently, difference comparisons between multiple candidate segmentation maps has become an effective method for sampling in active learning. For example, Wang et al. [24] believe that easy samples tend to obtain similar segmentations in K models, where they use K different models to segment images and measure the similarity of outputs, thus building connections between different models for further comparisons. Recently, Zhang et al. [25] generated two segmentation maps before and after processing of their proposed attention module, thus calculating the similarity coefficient of maps to guide sampling of their active learning framework.

2.3. Active Learning Methods

To reduce the cost of labeling, active learning selects the most valuable and informative samples from unlabeled samples for the labeling task, which relieves the dependence of deep learning models on large training datasets. Due to its ideal function in achieving desirable performance with few labelled samples, researchers in the medical image analysis community have proposed several methods for CAD.
For example, Ayerdi et al. [26] proposed an interactive image segmentation system using active learning, which allows rapid segmentation without the requirement of manual intervention. Later, Sharma et al. [27] adopted active learning to perform biomedical image segmentation with limited labeled data, where they combined UNet and an active learning query strategy to select additional samples for annotation, thus capturing the most uncertain and representative samples. Then, Li et al. [28] proposed a deep active learning framework which combines an attention-gated fully convolutional network (ag-FCN) and a distributional difference-based active learning algorithm (dd-AL) to iteratively annotate samples. Later, Lai et al. [29] proposed a semi-supervised active learning framework with region-based selection criteria which iteratively selects regions for annotation queries to rapidly expand the diversity and number of marker sets.
Most recently, Gaillochet et al. [30] proposed a test-time augmentation method for active learning in medical image segmentation, which exploits the power of uncertain information provided by data transformation. Bai et al. [31] proposed a difference-based active learning (DEAL) method for bleed segmentation, which successfully bridged the gap between class activation maps (CAMs) and ground truth with few annotations.

3. Method

To deal with the high labeling cost of breast cancer images, we propose an attention-based deep active learning framework for segmentation in mammograms. First, the overall structure is given, offering a global view of how the proposed framework works. Then, we present the basic breast cancer segmentation model with multi-scale feature fusion and enhancement. Afterwards, we describe an attention-based sampling scheme to assign weights for unlabeled samples under uncertainty. Finally, we describe an active-learning-based labelling strategy to choose the unlabeled samples for manually labeling and retraining, thus reducing the cost of manually labeling a large quantity of unlabeled samples.

3.1. Overall Structure

The existing segmentation models based on deep learning generally require a large number of labeled images for training at substantial cost. Thus,  it is crucial to achieve as high an accuracy as possible for segmentation with few labeled samples. To achieve this goal, we propose the overall framework as shown in Figure 1, which consists of three steps, i.e., the basic segmentation model, attention-based sampling and active-learning-based labelling.
During Step A, a labeled set of breast cancer images are first used to train the basic segmentation model, which obtains distinguished feature maps via multi-scale feature fusion and enhancement. In Step B, uncertainty sampling is first adopted to classify good and bad segmentation results for breast cancer in mammograms. Then, a novel attention mechanism is built to calculate weights for samples of the unlabeled breast cancer set. During Step C, unlabeled samples with higher weights, implying that they are more informative for learning, are selected to be manually labeled by professional medical experts. All these samples can then be further used to retrain the breast cancer segmentation model, thus boosting performance in an iterative manner.
On the basis of structure design, we propose a loss function to involve the traditional IoU loss L o s s I and binary cross-entropy loss L o s s B , i.e., L o s s = L o s s I + L o s s B . Specifically, IoU loss L o s s I can be calculated via
L o s s I = 1 i I p i p ^ i i I p i + p ^ i p i p ^ i ,
where I refers to the input breast cancer image, i represents each pixel in the image, and p ^ i and p i represent the predicted and true labeled values for each pixel, respectively. Moreover, the binary cross-entropy loss L o s s B can be calculated via
L o s s B = i I p i log ( p ^ i ) + ( 1 p i ) log ( 1 p ^ i ) ,
Note that L o s s I and L o s s B are calculated in each iteration stage to achieve convergence of the training process.

3.2. Basic Segmentation Model for Breast Cancer

Although skip connections in UNet could avoid the loss of detailed information caused by continuous down-sampling, they cannot capture multi-scale information with strong restrictions on locality. To obtain multi-scale feature maps, we cascade the feature maps at multiple layers with different receptive fields, where skip connections are used across different layers.
On the basis of the UNet segmentation model, the proposed basic model further involves the strength of the promotion feature module (PFM) [32] for the features of each output layer, which fuses multi-scale feature maps to enhance their representation capability. Note that the PFM works as a feature fusion and enhanced block in our former work, which not only fuses the features from multiple scales, but also selectively forgets useless information and enhances informative information, thus constructing more effective feature representation.
Specifically, an input image I in the labeled breast cancer set is sent to the proposed basic segmentation model for feature extraction:
F i = S e g t i ( I ) ,
where F i refers to feature maps corresponding to the ith output layer of the UNet model, and the function S e g t ( ) refers to the segmentation model during the tth iteration of retraining.
Then, feature maps of multiple layers are sent to different PFMs for fusion and enhancement:
F E i = G e i ( G f i ( F i ) ) , w h e r e i = 1 , 2 , . . . , n
where F E i represents the output feature map after processing of the ith PFM, functions G e and G f represent enhance and forget operations in PFM, and n is the total number of layers in the segmentation model.
In the later fusion step, the original feature map F i corresponding to the i-th layer is fused with its enhanced version F E i via skip connections:
F i 1 = F E i X U P ( F i ) , w h e r e i = 1 , 2 , . . . , n
where F i 1 refers to the output feature map of the i 1 th layer, and the function X U P ( ) is an up-sampling operation.
Finally, the generated feature map F P 1 combines the high-layer semantic features with shallow features, thus enhancing representation capability for segmentation via multi-scale feature extraction fusion. We set n = 5 for all experiments in this paper.

3.3. Attention-Based Sampling Scheme

To obtain more segmentation related knowledge with as few labeled samples as possible, it is essential to obtain more distinguished feature representation from the training stage. We thus judge the informativeness for one specific sample based on its segmentation result with the following equation:
I n f o i = 1 , i f U i α 0 , o t h e r w i s e
where I n f o i implies whether the ith sample is useful for learning knowledge or not, α is a pre-set parameter based on segmentation performance of experiments, and U i calculates similarity coefficient with the following equation:
U i = S i , q S i , g S i , q S i , g ,
where S q and S g respectively represent the predicted and ground-truth feature maps of the ith breast cancer image, which is used to represent the uncertainty and guides the selection of unlabeled breast cancer images. We consider the segmentation result of the i-th breast cancer image as good only if I n f o i = 1 ; otherwise, it is considered bad.
With such criteria for judging informativeness, we choose samples from an unlabeled breast cancer set, which are either dissimilar to good segmentation samples or similar to bad segmentation samples, thus greatly improving the learning capability of the segmentation model for the features of difficult samples. Essentially, a soft attention model is generally formed as a dimension of interpretability into internal representations by selectively focusing on specific information. The core procedure of soft attention model [33] can calculate weights based on similarity between an input signal and pre-trained weights. Therefore, we propose a novel attention mechanism which assigns weights based on similarity calculations between unlabeled and labeled samples. In other words, the proposed attention mechanism assigns smaller weights if the unlabeled samples are more similar to the good segmentation samples. On the contrary, it would give higher weights to unlabeled samples with greater similarity to bad samples.
Defining the input unlabeled breast cancer image as query Q and the set of labeled segmentation samples as W, a multi-layer perceptron(MLP) is utilized to calculate the similarity or correlation between Q and one of the pre-trained samples W i as s i m ( Q , W i ) = M L P ( Q , W i ) . Afterwards, we adopt the Softmax function to perform normalization on the calculated similarity and emphasize the informative parts based on their inherent ability:
α i = s o f t m a x ( s i m ( Q , W i ) ) = e s i m ( Q , W i ) j = 1 L e s i m ( Q , W i ) ;
where L refers to the number of samples in the labeled segmentation results.

3.4. Active Labeling Strategy

Essentially, we believe that unlabeled samples with larger attention weight could contribute to the classification capability of the segmentation model, thus boosting the segmentation performance of the model by using these samples for retraining. Therefore, we propose a labeling strategy for breast cancer samples based on an active learning method, as described in Algorithm 1.
Specifically, we sort the unlabeled samples by weights and choose the unlabeled samples with higher weights for manual labeling, represented as Step 8 in the algorithm. The specific processes of sample calculation, selection and manual labeling and retraining are shown in Figure 2. We use both similarity and dissimilarity weights to select unlabeled images. Then, these samples are first roughly labeled by automatic labeling software and then manually adjusted by experts. Afterwards, they are added as labeled samples into the set of labeled breast cancer images. Meanwhile, these samples are deleted from the unlabeled image set, which can be represented as
U t + 1 = U t I t L t + 1 = L t + I t
where t refers to the iteration time of training, and L t U t = to ensure the consistent processing of different iterations.
Finally, the newly constructed labeled set L t + 1 is used to retrain the basic segmentation model. With all these steps, an active learning iteration process is completed, where the performance of the breast cancer segmentation model could be improved step by step.
Algorithm 1 Labeling strategy of breast cancer samples based on active learning.
Require: 
Unlabeled sample set U, labeled sample set L
Ensure: 
Labeled dataset L t
  1:
While:
  2:
if L then
  3:
   Train the breast cancer segmentation model
  4:
   Use the updated segmentation model to infer the labeled breast cancer images in the labeled sample set L, and output the feature maps F P
  5:
   Calculate the similarity of the i-th image
  6:
   Calculate the attention weights
  7:
end if
  8:
Sort the breast cancer images in the unlabeled sample set according to attention weights
  9:
Manually label the selected images which are assigned with higher weights
 10:
Update U t + 1 and L t + 1
 11:
Retrain the segmentation model for breast cancer images
 12:
EndWhile

4. Experimental Results

This section first introduces the datasets and measurements. Then, it describes the ablation experiments constructed to verify the effectiveness of the proposed modules. Afterwards, we demonstrate the performance of the segmentation models on four datasets to verify the effectiveness of the proposed framework. We also qualitatively compare the proposed active learning framework with existing methods. Finally, we offer the implementation details for readers’ convenience.

4.1. Datasets and Measurements

We collected breast cancer images from a cooperating hospital, which are not released due to privacy reasons. The dataset includes 1462 labeled breast cancer images, with resolution 6781 × 3676 pixels. Since images were acquired by different scanners, we divided all our samples into four parts based on the type of scanner, i.e., Breast-A, Breast-B, Breast-C and Breast-D.
Various measurements were used to verify the effectiveness of the breast cancer image segmentation results, i.e., the mean Dice similarity coefficient (mDice), the mean intersection ratio (mIoU), and the mean absolute error (MAE). Where T P , F P and F N indicate true positive, false positive and false negative samples, mDice can be calculated as:
m D i c e = 2 T P F P + 2 T P + F N .
Note that a higher mDice implies a greater similarity between two samples.
IoU is defined as the area of the intersection divided by the area of the union of the predicted bounding box, which can be evaluated by
I o U = area B p B g t area B p B g t ,
where B p and B g t are the predicted and ground-truth segmentation results, respectively.
By calculating the Euclidean distance between the predicted and the ground-truth results, MAE can be defined as:
M A E = 1 n i = 1 n | Y i Y i ^ | ,
where n refers to the total number of samples, and Y i and Y i ^ are the predicted and ground-truth labels, respectively. Note that a lower MAE implies a better segmentation result.

4.2. Ablation Experiments

To evaluate the effectiveness of the attention-based sampling mechanism and the active labeling strategy, we designed several ablation experiments as shown in Table 1. Note that PFMs represent multiple promotion feature modules used in the basic segmentation model, Att refers to the proposed attention-based sampling scheme, and Act is the proposed active labeling strategy.
After adding PFMs, Att and Act, the segmentation performance on breast cancer images gradually improved on our four breast cancer image datasets, proving the effectiveness of these three modules. Specifically, we found that UNet+PFMs could achieve more precise boundaries of polyp regions and performed accurate segmentation when compared with the basic network (i.e., UNet). However, the shallow usage of boundary information without multi-scale refinement for boundary regions leads to uncompact performance towards larger and more regularized-shape polyp regions. In contrast, due to the usage of PFMs to extract a more distinguishing feature map by fusing multi-scale information, boundaries achieved by the proposed method were much more obvious with clear contour lines, thus providing better segmentation performance.
The ablation experiment on Att proved that the attention-based sampling scheme improved segmentation performance on all datasets. The attention-based design helped in effectively selecting more valuable samples for the further manual labeling process. It is beneficial to focus on the most informative unlabeled samples, which brings the feature information required by the current model for performance improvement with the fewest updating iterations.
Act, representing the active labeling strategy, enlarges size of the labeled breast cancer set, thereby improving the effect of tumor segmentation. Due to the guidance of optimized selection on unlabeled samples, we observed that informative samples for the current trained model were added to the labeled samples set for further retraining.

4.3. Comparative Experiments

In this subsection, we describe our comparative experiments and present heatmap visualizations, segmentation results and the effectiveness of active learning.
Figure 3 shows the generated heatmap of breast cancer segmentation achieved by the proposed framework, the comparative method and the ground truth on four datasets. We used UNet in comparisons of heatmap visualization and segmentation results. The last layer of the network can generate a heatmap for each input breast cancer image, which can be used to generate segmentation results. Comparison of heatmaps shows that the proposed framework could more accurately identify the breast cancer region, and thus obtained a better performance in the segmentation task. Even in the case of a blurred image boundary, the heatmap implies that further segmentation results would maintain high accuracy by focusing on the dominant parts of cancer regions.
Figure 4 shows the qualitative comparison results of the breast cancer segmentation. Compared with the comparative segmentation model, i.e., UNet, the proposed framework achieved better segmentation results that were similar to the ground-truth results. Moreover, the proposed active learning strategy could refine the distinguished feature information of breast cancer using unlabeled samples, achieving more accurate pixel-level classification results. In addition, the proposed framework generates and refines the boundary region through an effective iterative update strategy, thus achieving global optimization progressively.
To verify that the active learning strategy can effectively reduce the cost of labeling, we conducted comparative experiments by selecting random sampling and CoreSet [34] as the comparison methods. Figure 5 shows plots of mean Dice for each iteration during the active learning. It is worth noting that the proposed active learning strategy not only had a higher mean Dice value, but also converged in fewer iterations. This proves that the proposed strategy selected more informative samples in each iteration, thus reducing the cost of labeling samples. Without measuring the uncertainty of sample labeling, other comparison methods might suffer from unstable convergence with increasing iterations because they adopt samples without helpful information, or even containing noisy information. Although all methods converged eventually, the compared methods tended to have lower mean Dice values due to the influence of noisy samples.

4.4. Implementation Details

Due to the scarcity of training samples, we used various data-enhancement methods to expand the training samples so that they met the requirements of model training. The size of input breast images was first adjusted to 352 × 352 pixels in the training and inference process. Then, we used image flipping to expand the number of training samples, in both horizontal and vertical directions. Finally, we not only randomly adjusted the contrast, brightness and sharpness of breast cancer images, but also randomly dilated and eroded image labels. All these operations were designed for data enhancement.
All experiments were carried out under the Linux Ubuntu operating system with a single Titan V GPU. We use Adaptive Moment Estimation (Adam) as the optimizer, while the initial learning rate was set to 0.0001 and the learning rate was adjusted using learning rate decay. In the active learning strategy experiment, we used 100 unlabeled breast cancer images for initialization, and set the number of iterations to 10 for training of the basic segmentation model. Afterwards, we added the selected unlabeled samples to the labeled dataset, and trained 25 epochs for retraining in each iteration.

5. Conclusions

Due to the high cost of labeling training samples, herein we propose an attention-based active learning framework for the segmentation of breast cancer in mammograms. Specifically, we propose an attention sampling scheme to assign weights for unlabeled samples by evaluating their uncertainty. We also propose an active labeling strategy to select valuable unlabeled samples for manual labeling, thus enlarging the scale of the training set and improving the performance of the segmentation model. Testing on four datasets, experimental results showed that the proposed framework could greatly improve segmentation accuracy. The active learning scheme and attention strategy we adopted can be easily applied to other models and effectively reduce the data size required for model training.
In the future, we will try to introduce semi-supervised learning and unsupervised learning in active learning to further improve the generalization ability of the segmentation model on different datasets. Moreover, we will design specific algorithms to solve problems in breast cancer segmentation such as microcalcification and architectural distortion, thus improving the segmentation accuracy.

Author Contributions

Conceptualization, X.F. and Y.W. (Yirui Wu); methodology, X.F., H.H. and H.C.; software, X.F. and H.C.; validation, B.L., Y.W. (Yirui Wu) and Q.H.; formal analysis, X.F.; investigation, Y.W. (Yansong Wang); resources, X.F.; data curation, X.F. and Y.W. (Yirui Wu); writing—original draft preparation, X.F., H.C. and Y.W. (Yirui Wu); writing—review and editing, X.F., H.C. and Y.W. (Yirui Wu);visualization, B.L., Y.W. (Yansong Wang) and Q.H.; supervision, Y.W. (Yirui Wu) and Q.H.; project administration, Y.W. (Yirui Wu) and Q.H.; funding acquisition, X.F. and Y.W. (Yirui Wu). All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by a grant from General Scientific Research Project of Zhejiang Education Department (Y202147224), National Key R&D Program of China under Grant No. 2021YFB3900601, the Fundamental Research Funds for the Central Universities under Grant B220202074, the Fundamental Research Funds for the Central Universities, JLU, Joint Foundation of the Ministry of Education (No. 8091B022123).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Data available on request due to restrictions eg privacy or ethical.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Elmoufidi, A. Deep Multiple Instance Learning for Automatic Breast Cancer Assessment Using Digital Mammography. IEEE Trans. Instrum. Meas. 2022, 71, 1–13. [Google Scholar] [CrossRef]
  2. Loizidou, K.; Skouroumouni, G.; Nikolaou, C.; Pitris, C. Automatic Breast Mass Segmentation and Classification Using Subtraction of Temporally Sequential Digital Mammograms. IEEE J. Transl. Eng. Health Med. 2022, 10, 1–11. [Google Scholar] [CrossRef] [PubMed]
  3. Seely, J.; Alhassan, T. Screening for breast cancer in 2018—What should we be doing today? Curr. Oncol. 2018, 25, 115–124. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Huang, Q.; Miao, Z.; Zhou, S.; Chang, C.; Li, X. Dense Prediction and Local Fusion of Superpixels: A Framework for Breast Anatomy Segmentation in Ultrasound Image With Scarce Data. IEEE Trans. Instrum. Meas. 2021, 70, 1–8. [Google Scholar] [CrossRef]
  5. Kim, H.E.; Kim, H.H.; Han, B.K.; Kim, K.H.; Han, K.; Nam, H.; Lee, E.H.; Kim, E.K. Changes in cancer detection and false-positive recall in mammography using artificial intelligence: A retrospective, multireader study. Lancet Digit. Health 2020, 2, e138–e148. [Google Scholar] [CrossRef] [Green Version]
  6. Chen, C.; Wang, Y.; Niu, J.; Liu, X.; Li, Q.; Gong, X. Domain Knowledge Powered Deep Learning for Breast Cancer Diagnosis Based on Contrast-Enhanced Ultrasound Videos. IEEE Trans. Med. Imaging 2021, 40, 2439–2451. [Google Scholar] [CrossRef] [PubMed]
  7. Wang, Y.; Jin, Z.; Tokuda, Y.; Naoi, Y.; Tomiyama, N.; Suzuki, K. Development of Deep-learning Segmentation for Breast Cancer in MR Images based on Neural Network Convolution. In Proceedings of the 2019 8th International Conference on Computing and Pattern Recognition, Beijing, China, 23–25 October 2019; pp. 187–191. [Google Scholar]
  8. Hann, E.; Biasiolli, L.; Zhang, Q.; Popescu, I.A.; Werys, K.; Lukaschuk, E.; Carapella, V.; Paiva, J.M.; Aung, N.; Rayner, J.J.; et al. Quality control-driven image segmentation towards reliable automatic image analysis in large-scale cardiovascular magnetic resonance aortic cine imaging. In International Conference on Medical Image Computing and Computer-Assisted Intervention; Springer: Cham, Switzerland, 2019; pp. 750–758. [Google Scholar]
  9. Chen, J.; Jiao, J.; He, S.; Han, G.; Qin, J. Few-Shot Breast Cancer Metastases Classification via Unsupervised Cell Ranking. IEEE ACM Trans. Comput. Biol. Bioinform. 2021, 18, 1914–1923. [Google Scholar] [CrossRef] [PubMed]
  10. Belharbi, S.; Ayed, I.B.; McCaffrey, L.; Granger, E. Deep Active Learning for Joint Classification & Segmentation with Weak Annotator. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 3–8 January 2021; pp. 3337–3346. [Google Scholar]
  11. Nath, V.; Yang, D.; Landman, B.A.; Xu, D.; Roth, H.R. Diminishing Uncertainty within the Training Pool: Active Learning for Medical Image Segmentation. IEEE Trans. Med. Imaging 2021, 40, 2534–2547. [Google Scholar] [CrossRef] [PubMed]
  12. Li, W.; Li, J.; Wang, Z.; Polson, J.; Sisk, A.E.; Sajed, D.P.; Speier, W.; Arnold, C.W. PathAL: An Active Learning Framework for Histopathology Image Analysis. IEEE Trans. Med. Imaging 2022, 41, 1176–1187. [Google Scholar] [CrossRef] [PubMed]
  13. Cheng, J.Z.; Chen, K.W.; Chou, Y.H.; Chen, C.M. Cell-based image partition and edge grouping: A nearly automatic ultrasound image segmentation algorithm for breast cancer computer aided diagnosis. In Proceedings of the Medical Imaging 2008: Computer-Aided Diagnosis, San Diego, CA, USA, 19–21 February 2008; Volume 6915, pp. 743–754. [Google Scholar]
  14. Eziddin, W.; Montagner, J.; Solaiman, B. An iterative possibilistic image segmentation system: Application to breast cancer detection. In Proceedings of the 2010 13th International Conference on Information Fusion, Edinburgh, UK, 26–29 July 2010; pp. 1–8. [Google Scholar]
  15. Gnonnou, C.; Smaoui, N. Segmentation and 3D reconstruction of MRI images for breast cancer detection. In Proceedings of the International Image Processing, Applications and Systems Conference, Sfax, Tunisia, 5–7 November 2014; pp. 1–6. [Google Scholar]
  16. Kaushal, C.; Singla, A. Automated segmentation technique with self-driven post-processing for histopathological breast cancer images. CAAI Trans. Intell. Technol. 2020, 5, 294–300. [Google Scholar] [CrossRef]
  17. Jing, T.Y.; Mustafa, N.; Yazid, H.; Rahman, K.S.A. Segmentation of Tumour Regions for Tubule Formation Assessment on Breast Cancer Histopathology Images. In Proceedings of the 11th International Conference on Robotics, Vision, Signal Processing and Power Applications; Springer: Berlin/Heidelberg, Germany, 2022; pp. 170–176. [Google Scholar]
  18. Shi, G.; Wu, Y.; Liu, J.; Wan, S.; Wang, W.; Lu, T. Incremental Few-Shot Semantic Segmentation via Embedding Adaptive-Update and Hyper-class Representation. In Proceedings of the ACM International Conference on Multimedia, Tokyo, Japan, 13–16 December 2022; pp. 5547–5556. [Google Scholar]
  19. Wu, Y.; Guo, H.; Chakraborty, C.; Khosravi, M.; Berretti, S.; Wan, S. Edge Computing Driven Low-Light Image Dynamic Enhancement for Object Detection. IEEE Trans. Netw. Sci. Eng. 2022. [Google Scholar] [CrossRef]
  20. Su, H.; Liu, F.; Xie, Y.; Xing, F.; Meyyappan, S.; Yang, L. Region segmentation in histopathological breast cancer images using deep convolutional neural network. In Proceedings of the 2015 IEEE 12th International Symposium on Biomedical Imaging (ISBI), Brooklyn, NY, USA, 16–19 April 2015; pp. 55–58. [Google Scholar]
  21. He, S.; Ruan, J.; Long, Y.; Wang, J.; Wu, C.; Ye, G.; Zhou, J.; Yue, J.; Zhang, Y. Combining deep learning with traditional features for classification and segmentation of pathological images of breast cancer. In Proceedings of the 2018 11th International Symposium on Computational Intelligence and Design (ISCID), Hangzhou, China, 8–9 December 2018; Volume 1, pp. 3–6. [Google Scholar]
  22. Roy, A.G.; Conjeti, S.; Navab, N.; Wachinger, C. Inherent brain segmentation quality control from fully convnet monte carlo sampling. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Granada, Spain, 16–20 September 2018; Springer: Berlin/Heidelberg, Germany, 2018; pp. 664–672. [Google Scholar]
  23. Shen, H.; Tian, K.; Dong, P.; Zhang, J.; Yan, K.; Che, S.; Yao, J.; Luo, P.; Han, X. Deep active learning for breast cancer segmentation on immunohistochemistry images. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Lima, Peru, 4–8 October 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 509–518. [Google Scholar]
  24. Wang, J.; Chen, Z.; Wang, L.; Zhou, Q. An Active Learning with Two-step Query for Medical Image Segmentation. In Proceedings of the 2019 International Conference on Medical Imaging Physics and Engineering (ICMIPE), Shenzhen, China, 22–24 November 2019; pp. 1–5. [Google Scholar]
  25. Zhang, Z.; Li, J.; Tian, C.; Zhong, Z.; Jiao, Z.; Gao, X. Quality-driven deep active learning method for 3D brain MRI segmentation. Neurocomputing 2021, 446, 106–117. [Google Scholar] [CrossRef]
  26. Ayerdi, B.; Graña, M. Random forest active learning for retinal image segmentation. In Proceedings of the 9th International Conference on Computer Recognition Systems CORES, Wroclaw, Poland, 25–27 May 2015; Springer: Berlin/Heidelberg, Germany, 2016; pp. 213–221. [Google Scholar]
  27. Sharma, D.; Shanis, Z.; Reddy, C.K.; Gerber, S.; Enquobahrie, A. Active learning technique for multimodal brain tumor segmentation using limited labeled images. In Domain Adaptation and Representation Transfer and Medical Image Learning with Less Labels and Imperfect Data; Springer: Berlin/Heidelberg, Germany, 2019; pp. 148–156. [Google Scholar]
  28. Li, H.; Yin, Z. Attention, suggestion and annotation: A deep active learning framework for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Lima, Peru, 4–8 October 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 3–13. [Google Scholar]
  29. Lai, Z.; Wang, C.; Oliveira, L.C.; Dugger, B.N.; Cheung, S.C.; Chuah, C.N. Joint Semi-supervised and Active Learning for Segmentation of Gigapixel Pathology Images with Cost-Effective Labeling. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 591–600. [Google Scholar]
  30. Gaillochet, M.; Desrosiers, C.; Lombaert, H. TAAL: Test-Time Augmentation for Active Learning in Medical Image Segmentation. In Proceedings of the MICCAI Workshop on Data Augmentation, Labelling, and Imperfections, Singapore, 22 September 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 43–53. [Google Scholar]
  31. Bai, F.; Xing, X.; Shen, Y.; Ma, H.; Meng, M.Q.H. Discrepancy-Based Active Learning for Weakly Supervised Bleeding Segmentation in Wireless Capsule Endoscopy Images. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Singapore, 18–22 September 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 24–34. [Google Scholar]
  32. Wu, B.; Wu, Y.; Wan, S. An Image Enhancement Method for Few-shot Classification. In Proceedings of the IEEE International Conference on Embedded and Ubiquitous Computing, Shenyang, China, 20–22 October 2021; pp. 1–7. [Google Scholar]
  33. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is All you Need. In Proceedings of the NIPS, Long Beach, CA, USA, 4–9 December 2017; pp. 5998–6008. [Google Scholar]
  34. Sener, O.; Savarese, S. Active Learning for Convolutional Neural Networks: A Core-Set Approach. In Proceedings of the International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
Figure 1. Framework of the proposed framework, which consists of (A) Basic Segmentation Model, (B) Attention-based Sampling, and (C) Active-Learning-based Labeling.
Figure 1. Framework of the proposed framework, which consists of (A) Basic Segmentation Model, (B) Attention-based Sampling, and (C) Active-Learning-based Labeling.
Applsci 13 00852 g001
Figure 2. Key steps, i.e., Calculation, Selection, and Manual Labeling, in the proposed active-learning-based labeling strategy.
Figure 2. Key steps, i.e., Calculation, Selection, and Manual Labeling, in the proposed active-learning-based labeling strategy.
Applsci 13 00852 g002
Figure 3. Heatmaps of breast cancer segmentation achieved by the proposed framework, the comparative method and the ground truth results. (AD) refer to Breast-A, Breast-B, Breast-C and Breast-D datasets, respectively.
Figure 3. Heatmaps of breast cancer segmentation achieved by the proposed framework, the comparative method and the ground truth results. (AD) refer to Breast-A, Breast-B, Breast-C and Breast-D datasets, respectively.
Applsci 13 00852 g003
Figure 4. Qualitative comparisons between the segmentation results achieved by the proposed framework, the comparative method, and the ground truth. (AD) refer to Breast-A, Breast-B, Breast-C and Breast-D datasets, respectively.
Figure 4. Qualitative comparisons between the segmentation results achieved by the proposed framework, the comparative method, and the ground truth. (AD) refer to Breast-A, Breast-B, Breast-C and Breast-D datasets, respectively.
Applsci 13 00852 g004
Figure 5. Plots of mean Dice coefficients, comparing different sampling strategies. It’s noted that RS and CS refer to random selecting sampling and CoreSet sampling [34], respectively.
Figure 5. Plots of mean Dice coefficients, comparing different sampling strategies. It’s noted that RS and CS refer to random selecting sampling and CoreSet sampling [34], respectively.
Applsci 13 00852 g005
Table 1. Ablation experiments with different network structure designs on Breast-A, Breast-B, Breast-C and Breast-D datasets, where PFMs, Att and Act represent multiple promotion feature modules used in the basic segmentation model, the proposed attention-based sampling scheme and the proposed active labeling strategy, respectively.
Table 1. Ablation experiments with different network structure designs on Breast-A, Breast-B, Breast-C and Breast-D datasets, where PFMs, Att and Act represent multiple promotion feature modules used in the basic segmentation model, the proposed attention-based sampling scheme and the proposed active labeling strategy, respectively.
DatasetMethodmDiceIoUMAE
Breast-AUNet0.3560.2660.018
UNet+PFMs0.4120.3320.013
UNet+PFMs+Att0.4310.3610.013
UNet+PFMs+Att+Act0.4620.3940.009
Breast-BUNet0.4030.290.013
UNet+PFMs0.4920.3710.009
UNet+PFMs+Att0.5150.3820.112
UNet+PFMs+Att+Act0.5430.4010.006
Breast-CUNet0.5530.4290.033
UNet+PFMs0.6770.5120.026
UNet+PFMs+Att0.6930.5180.029
UNet+PFMs+Att+Act0.7250.5330.023
Breast-DUNet0.3690.2610.040
UNet+PFMs0.4510.3820.311
UNet+PFMs+Att0.4780.3960.335
UNet+PFMs+Att+Act0.5120.4220.027
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Fu, X.; Cao, H.; Hu, H.; Lian, B.; Wang, Y.; Huang, Q.; Wu, Y. Attention-Based Active Learning Framework for Segmentation of Breast Cancer in Mammograms. Appl. Sci. 2023, 13, 852. https://doi.org/10.3390/app13020852

AMA Style

Fu X, Cao H, Hu H, Lian B, Wang Y, Huang Q, Wu Y. Attention-Based Active Learning Framework for Segmentation of Breast Cancer in Mammograms. Applied Sciences. 2023; 13(2):852. https://doi.org/10.3390/app13020852

Chicago/Turabian Style

Fu, Xianjun, Hao Cao, Hexuan Hu, Bobo Lian, Yansong Wang, Qian Huang, and Yirui Wu. 2023. "Attention-Based Active Learning Framework for Segmentation of Breast Cancer in Mammograms" Applied Sciences 13, no. 2: 852. https://doi.org/10.3390/app13020852

APA Style

Fu, X., Cao, H., Hu, H., Lian, B., Wang, Y., Huang, Q., & Wu, Y. (2023). Attention-Based Active Learning Framework for Segmentation of Breast Cancer in Mammograms. Applied Sciences, 13(2), 852. https://doi.org/10.3390/app13020852

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop