Next Article in Journal
upU-Net Approaches for Background Emission Removal in Fluorescence Microscopy
Next Article in Special Issue
Multiple Papillomas of the Breast: A Review of Current Evidence and Challenges
Previous Article in Journal
Comparison of Ultrasound Image Classifier Deep Learning Algorithms for Shrapnel Detection
Previous Article in Special Issue
BI-RADS BERT and Using Section Segmentation to Understand Radiology Reports
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Image Augmentation Techniques for Mammogram Analysis

1
Computer Science and Engineering Department, School of Technology, Pandit Deendayal Energy University, Gandhinagar 382007, India
2
Department of Computing and Informatics, Bournemouth University, Poole BH12 5BB, UK
*
Authors to whom correspondence should be addressed.
J. Imaging 2022, 8(5), 141; https://doi.org/10.3390/jimaging8050141
Submission received: 15 March 2022 / Revised: 19 April 2022 / Accepted: 22 April 2022 / Published: 20 May 2022

Abstract

:
Research in the medical imaging field using deep learning approaches has become progressively contingent. Scientific findings reveal that supervised deep learning methods’ performance heavily depends on training set size, which expert radiologists must manually annotate. The latter is quite a tiring and time-consuming task. Therefore, most of the freely accessible biomedical image datasets are small-sized. Furthermore, it is challenging to have big-sized medical image datasets due to privacy and legal issues. Consequently, not a small number of supervised deep learning models are prone to overfitting and cannot produce generalized output. One of the most popular methods to mitigate the issue above goes under the name of data augmentation. This technique helps increase training set size by utilizing various transformations and has been publicized to improve the model performance when tested on new data. This article surveyed different data augmentation techniques employed on mammogram images. The article aims to provide insights into basic and deep learning-based augmentation techniques.

1. Introduction

Amongst various artificial intelligence fields, Deep Learning (DL) is widely adopted for the processing and analysis of radiological images. DL has been successfully applied to multiple Computer Vision tasks such as Object Segmentation, Detection, and Classification, mainly thanks to accuracy rates achieved by convolutional neural networks (CNNs). CNNs have the capabilities to automatically learn features through several network layers from a large set of labelled datasets [1]. Concerning the biomedical image analysis topic, CNNs have been successfully utilised for various tasks such as lesion or tumour classification, suspicious region detection, and abnormality detection [2,3,4]. DL-based solutions serve as a second opinion tool for expert radiologists and assist them in decision-making, and proper treatment planning [5]. However, there needs to be a large amount of ground truth to build a DL model capable of inferring knowledge from data and avoiding the model being very accurate only on the training dataset images. The latter goes under the name of overfitting [6,7] and represents a critical issue to overcome to have a model capable of delivering appropriate knowledge inference capabilities on a given application domain. Furthermore, having high-quality and manually annotated data is a time-consuming and expert dependent task. Unfortunately, that is quite common in the context of mammogram analysis [8,9,10]. One of the most challenging tasks for DL models is the generalisation, with generalisation being the capability of models to recognise those categories they were trained for on new data [11,12]. The model with poor generalisation generally does not perform well due to high overfitting on the training set. Overfitting can be observed somehow in the plot showing validation accuracy at every epoch of the training phase [1]. Figure 1 shows the pictorial representation of models with and without overfitting. The training and validation loss curve is progressively and simultaneously reducing, which is a perfect circumstance, as shown in Figure 1 (left). The right side of the figure shows overfitting, in which the validation loss begins to grow after a certain number of epochs. In contrast, the training loss keeps decreasing. That is due to the model’s inability to work effectively with unknown or new data. One of the reasons for this phenomenon might be a lack of enough training samples. The validation error of suitable DL models should continue to decrease along with the training error. Data augmentation methods can help achieve this task. Augmented data can characterise the inclusive set of input data points and minimise the distance between validation and training data. Data augmentation techniques apply alterations to training datasets to produce more samples. Moreover, this technique helps the model avoid learning features too specific to the original data, resulting in a more generalised model with improved performance on the test dataset. Class distribution imbalance in datasets is another common challenge. For instance, binary classification problems occur when one class (the minority class) holds considerably fewer samples than the other class (the majority class). Due to this, the model may get biased towards the majority class, possibly resulting in misclassification. Augmenting the minority class images may be used to mitigate the imbalance problem. Data augmentation is not the only approach to reduce the effect of overfitting and class imbalance. Other options for avoiding overfitting in DL models are also explored in the literature (see Figure 2).
Batch Normalization: Batch Normalization can overcome the side-effect of overfitting by diminishing the internal covariate shift and instability in the distributions of Deeper networks’ layer activations. For each mini-batch, batch normalisation standardises the inputs to a layer. That has the effect of bringing the learning process into balance.
Dropout: Dropout applies during the training phase to get randomly selected neurons ignored. That avoids the so-called layer’s “over-reliance” on a few inputs. Still, apart from that, it also prevents neurons from co-adapting to training data.
Transfer Learning (TL): Transfer learning improves models’ performances on new and unknown data. The main point with TL is employing pre-trained models to be fine-tuned on a specific application domain using a small-sized dataset.
Pre-training: Model pre-training is similar to TL; the only difference is that model architectures can be defined, and weights are transferred.
Early-stopping: It allows providing an arbitrarily large number of training epochs to suddenly stop training if the model does not perform well on the validation set.

1.1. Research Contribution

Image augmentation techniques have been applied to mammogram datasets to increase the training set size, allowing data-hungry learners to benefit from more representative data. A review is conducted to summarise image augmentation techniques used in medical imaging applications such as deep learning-based for breast cancer diagnosis. The two following main categories of image augmentation techniques are surveyed here: (1) Basic image augmentation techniques and (2) Advanced augmentation techniques. The search terms used in the study are combinations of keywords such as “data augmentation”, “image augmentation”, “deep learning”, “breast cancer”, and “mammograms”. Articles that do not utilise or discuss data or image augmentation were discarded in this study. The research mainly focuses on image augmentation for mammogram images. Therefore, articles whose subject is on other imaging modalities such as CT scans, Breast MRI, Breast ultrasounds, Histopathology, etc., are excluded. Articles on image augmentation used in the literature for breast image analysis applications are also summarised according to the dataset, model, technique, tasks performed, etc. In the scientific literature, comprehensive and insightful surveys on image augmentation methods are present; some are specific to medical images. For example, the authors in [1] suggested several data augmentation solutions as ways to tackle models overfitting due to low-sized datasets. Another article [5] presents a thorough evaluation of the data augmentation methods employed in the broad topic of medical image analysis. In further detail, the authors focused on CT and MRI. However, another article reports recent advancements in data-augmentation techniques for brain MRI [13] by examining the papers submitted to the Multimodal Brain Tumor Segmentation Challenge (BraTS 2018 edition [14]). This paper aims to quickly access the research field and form an appropriate groundwork for the domain. This work examines several articles from various conferences, books and indexed journals out of scientific databases such as Scopus, IEEE, Web of Science and PubMed in compliance with PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) [15] recommendations. Figure 3 depicts the selection procedure. From a data augmentation standpoint, the goal of this work is to provide insights into the broader area of mammography image analysis. Mammograms are the primary matter of discussion in the article. The goal is to give readers an understanding of basic and deep learning-based augmentation approaches. Therefore, other than other review articles on the topic, the main goal here is to survey different approaches for data augmentation and check through the pros and cons of mammogram analysis related tasks. The impact of data augmentation techniques is analysed by checking through the dataset size increase and the pre and post-augmentation accuracy rates of models over a specific task.

1.2. Paper Topology

The paper is structured as follows: Section 1 provides background and context for image augmentation within the broad topic of a deep learning-based CAD system for medical imaging. Section 2 delves into various image augmentation techniques used in practice. Advanced image augmentation methods are showcased in Section 3. Insights into test-time augmentation are provided in Section 4. Discussions and conclusions, Section 5 and Section 6, respectively end the paper.

2. Basic Image Augmentation Techniques

Data augmentation techniques have been used to increase the size of the training set to provide more illustrative training samples to large-capacity learners [16]. Data augmentation encompasses a comprehensive range of techniques by inserting random variations into the existing training samples while preserving class labels. The purpose of data augmentation is to improve the model knowledge inference capability. One of the most meaningful principles adopted in data augmentation relates to the physical phenomenon of a state perturbation. The latter takes the form of slightly changed versions of images. Consequently, it increases the dataset size, allowing the network to infer knowledge from a broader pool of images. Therefore, when using deep learning in computer vision tasks, three types of data augmentation are the most likely; (1) Dataset generation and expansion. (2) On-the-fly data augmentation. (3) Amalgamation of Dataset generation and on-the-fly data augmentation. As already mentioned and widely covered in the scientific literature, supervised DL models [17] need a large amount of training data to unleash their knowledge inference capabilities fully. In the worst-case scenario, only one image is available, and data augmentation comes into play to produce a complete image collection. The task carries out random transformations (rotation, flipping, etc.) and other effects on the original image. Then, the newly generated images feed the DL model during the training phase. Methods like generation and expansion can forge N number of images. However, these approaches are not exempt from flaws: employing images generated by these methods does not necessarily improve models’ generalisation abilities. On-the-fly data augmentation (sometimes also called in-place) is the second type of data augmentation [18]. On-the-fly data augmentation helps DL model training see new variations of images at each epoch. It takes image batches as input and then applies a series of random transformations and other effects on each image in the batch. It finally returns a randomly altered image batch.

2.1. Geometric Transformations

In geometric transformation, an original image undergoes various modifications such as translation, rotation, scaling, flipping, or resizing to increase the training dataset size [5]. These conventional data augmentation techniques produce somewhat correlated images [19] and hence offer significantly less improvement to the model training and generalisation over test data. However, these transformations lead to a significant increase in the training dataset; therefore, they are widely used in the domain [13]. This section presents the most commonly used geometric transformations for computer-aided breast cancer diagnosis. It also briefly surveys methods building on the data augmentation methods mentioned above.

2.1.1. Flipping

Flipping generates a mirror image of an image with both horizontal or vertical axes. The horizontal axis is more preferred over vertical flipping because the top and bottom parts of an image may not be interchangeable always [13]. However, flipping cannot always be a label-preserving transformation (e.g., MNIST dataset) [1]. For example, in datasets such as DDSM [20] and CBIS-DDSM [21], most of the breast profiles are on the left side of the mammograms. Making uniform direction of the breast in mammograms makes padding easier to perform during preprocessing steps.

2.1.2. Rotation

Images are rotated leftward or rightward across an axis within the range [1°, 359°]. The rotation angle determines the safety of this augmentation technique. The possibility of keeping the label post-transformation is known as a Data Augmentation method’s safety. An image label might no longer be preserved with an increase in rotation degree. For example, rotation transformation is possibly safe on medical image datasets (X-ray, mammograms, Breast MRI, etc.) as well as on images of other datasets like ImageNet [22], but not on images of 9 and 6 for digit identification tasks.

2.1.3. Translation

Translation applies to prevent positional bias [1]. This transformation translates the whole image by a given translation vector along a specific direction. It helps the network learn geographically invariant properties rather than focusing on features present in a single spatial location [13]. In the case of breast mammograms, translation of images can generate suitable augmented images. After the translation, padding or pixel replication usually comes into play to fill out the leftover space. The process keeps the image dimensions [1].

2.1.4. Scaling

Scaled versions of images are added to the training set; deep neural networks can learn features regardless of their original scale. Furthermore, scaling can be applied using scaling factors in different directions. For example, breast lesions may vary in size; this transformation can bring realistic augmented images into the training dataset. Figure 4 shows examples of geometric transformations applied to MIAS [23] images.
Limited dataset size is one of the most common barriers in the medical research domain. Therefore, the scientific literature provides a wide range of techniques to handle this issue. For example, Costa et al. [24] employed data augmentation to create new images based on their original clinical mammography dataset, and they compared the results across different CNN architectures. The authors used geometric transformations such as rotation by varying degrees, flipping and adding Poisson noise. The model performs better when more regions of interest are added to the training step using data augmentation techniques. A new CNN model for identifying architectural distortion is proposed by Oyelade et al. in [8], which uses rotation, flipping, shearing and scaling for data augmentation. In their article, Cha et al. [25] stated, “Deep learning algorithms can improve performance by expanding their training set with synthetic examples”.
Horizontal and vertical flipping methods were used by Omonigho et al. [26] to augment the training set. By augmenting the training set with scaling, horizontal flip, and rotation, the authors could achieve 95.70% overall accuracy on the modified AlexNet model. In another study [27], Rahman et al. showed how specific pre-processing, transfer learning, and data augmentation approaches may help overcome the dataset size bottleneck in medical imaging applications. Geometric transformations such as reflection, translation, random scaling and random rotations were applied to the DDSM mammogram dataset.
Shi et al. [28] implemented a customised CNN to classify BI-RADS [29] density of mammogram images. The MIAS dataset was augmented using various transformations such as zooming, flipping, rotation and shifting. The authors carried out five-fold cross-validation of the model, which yielded an average test accuracy of 83.6%. Still, it is paramount to keep a certain level of variety between the images. Therefore, Khan et al. [30] developed a mammogram classification system and adopted random horizontal and vertical shifts, random shear and zoom as data augmentation techniques.
Zhang et al. [31] performed data augmentation through reflection and rotation. Initially, each original image underwent horizontal flipping, and then original and reflected images were rotated by 90°, 180°, and 270° degrees, respectively. As a result, the dataset increased eight times in size. The authors evaluated seven different architectures and concluded that models built and optimised using data augmentation and transfer learning had a lot of potential for automatic breast cancer detection. Bruno et al. [12] extracted patches from mammogram datasets such as MiniMIAS [32] and their own freely accessible dataset called SuREMaPP. Image transformations such as translations, horizontal reflections, and crops were employed in the study to generate augmented patches. Figure 5 shows an example of patches and augmented patches out of geometric transformations.
Assari et al. [33] suggested a BI-RADS-based CAD system for breast mass discrimination. The study makes use of GoogleNet for transfer learning. The authors used the DDSM breast mammography dataset. The authors used augmentation techniques such as horizontal flipping, clipped ROIs, and random zero-mean Gaussian noise to address the issue of overfitting. Another work by authors of [34] presented a system for mass detection and classification. To address the issue of class imbalance and small-sized datasets, the authors used two different augmentation approaches. In the first method, the authors have augmented the whole dataset and then divided it into training and testing sets.
Muduli et al. [35] ran various geometric transformations and Gaussian noise to generate a large number of samples: 2240 for Inbreast, 3200 for MIAS, and 28,800 for DDSM. The resulting images helped the authors work out the training of a deep CNN for breast cancer classification. A new domain generalization method is proposed in [36] to aid mammography lesion detection techniques. Domain-invariant features were embedded in a range of datasets using a multi-style and multi-view contrastive learning technique. The results showed that the domain generalisation technique is successful and can significantly improve both seen and unseen lesion detection tasks.

2.2. Pixel Level Augmentation

Pixel-level augmentation is quite helpful for research in medical imaging fields, as medical images are obtained with several technologies and imaging modalities; hence, they can be essentially assorted in pixel intensities [13]. In pixel-level augmentation, intensities of pixels are perturbed with random noise and a given probability, also called random intensity variation. In addition, a pixel-level augmentation modifies the brightness of an image. Among others, gamma correction (and all its variants), image blurring, and image sharpening represent forms of pixel-level augmentation [37,38,39].

2.3. Pseudo-Colour Augmentation

Pseudo-colour augmentation applies to colour channels spaces. Isolating a single colour channel, such as R, G, or B, is the first step for colour augmentation consisting of deriving a colour histogram that describes the image allows further advanced colour augmentations. Mammograms are turned into pseudo-colour pictures to assess the effectiveness of Mask R-CNN. The latter is carried out using multi-scale morphological sifting, which boosts mass-like patterns. Mask R-CNN is then used with transfer learning to detect and segment masses on pseudo-colour images at the same time [40].

2.4. Random Erasing

Random erasing is another data augmentation technique [41] complementary to the previously described ones. The main goal of this technique is to make a model robust against occlusions in images. One of the most meaningful features is the learning phase being parameter-free. In Figure 6, some examples of random erasing image augmentation are given.

2.5. Kernel Filters

Kernel filters rely on spatial filtering techniques to sharpen or smooth pixel values. It uses M × M size filtering masks. Along with transformations such as padding, flipping, and cropping, Kang et al. [42] used a kernel filter swapping pixel values with a n × n sliding window. The experiments were carried out on four different datasets such as SVHN [43], MNIST [44], CIFAR-10 [45], and STL-10 [46]. A different data augmentation method involves combining images by averaging their pixel values. The main point is to generate different samples to have the model be able to generalise information from data [1]. Various kernel filters such as gaussian blur, mean filter, median filter, Laplacian filter, etc., can be employed for data augmentation purposes.
Figure 7 shows some images generated with various kernel filters. Adedigba et al. in [47] employed the augmented dataset to train five state-of-the-art models. The authors used Gaussian blurring and additions of white noise and geometric transformations to increase the training set. The experiments showed DensNet remarkably achieving the highest training and validation accuracy (99.01% and 99.99%, respectively). Artificially generated mammograms and data augmentation techniques are applied by Yemini et al. in [48] to increase and balance the available database at training time. Along with flipping transformation, the authors of this work used Gaussian noise and changed image brightness to generate new images from the original samples.

3. Advanced Augmentation Techniques

In the last decade, several methods have been proposed to generate new samples from a reference dataset. The main goal has been to overcome all issues related to basic data augmentation techniques. This section groups advanced techniques for data augmentation into two main categories: GAN-based augmentation, and NST augmentation.

3.1. GAN-Based Augmentation

GANs (Generative Adversarial Networks) belong to the family of unsupervised deep learning algorithms capable of extracting hidden underlying properties from data and employing them in decision-making. The fundamental goal of a GAN is to develop new image samples (by a generator) that the discriminator will not be able to tell apart from the original ones (Both network branches compete against each other and gradually learn to produce better results) [49]. GANs are reliant on two main components, namely, generator and discriminator. Scientific literature shows they are also used to learn noise augmentations. In adversarial training, one model classifies examples while another adds noise to deceive the classifier. The adversarial model is then given a loss function by the classification model, allowing it to improve itself to create better noise. Including images from adversarial training might help models acquire more robust features that are less sensitive to noise distortions. Although it has been proven that using adversarial search to inject noise improves performance in adversarial cases, it is unclear if this is effective for decreasing overfitting. That is still currently an open challenge and has researchers investigating the link between adversarial attack resistance and actual performance on test datasets [1]. In addition, several DL-based augmentation systems employ adversarial training (including GAN-based and other adversarial learning networks) [50,51]. GANs are a widely used data augmentation approach to detect patterns and variances in image samples from the training dataset [52,53]. They have also been used for breast mass detection [53], mass classification [10] as well as mass segmentation [54]. The realistic level of artificially generated images for medical scenarios is still a debate matter [5]. On the other side, Shen et al. [55] provide a unique strategy based on GANs for generating varied mass images to perform contextual infilling by incepting synthetic masses into mammograms. Furthermore, their system automatically annotates the generated mass from patches. As shown in Figure 8, a mammogram image is transformed into a new one containing the incepted mass. Shen et al. carried out similar experiments on a private dataset too (Figure 9). In [56], the authors generated synthetic data via Cycle GAN. They carried out mass classification tests over 412 images: 212 with cancerous mass and 202 with no cancerous mass. The findings revealed that synthetically generated images, along with domain transformation from unrelated masses, could be used to increase the training sample size and improve the mass classification accuracy rates.
A deep neural system to support tumour recognition in mammograms was proposed in [57], with GANs serving as a data augmentation tool. The research was carried out utilising a large-sized database containing around 10,000 mammographic images from the DDSM dataset. A class-conditional GAN (ciGAN) was trained to perform contextual in-filling, which is subsequently used to synthesise lesions onto healthy screening mammograms in [9]. The authors also showed ciGAN-synthesized samples for cancerous to non-cancerous and non-cancerous to cancerous transformations. Experimental research involving specialist doctors in the assessment of GANs on the generation of medical images was presented in [58]. Some promising results showed that the developments of GAN-based image synthesis could successfully apply to high-resolution medical imaging. Figure 10 provides examples of original and synthetic mammograms from [58]. Users may modify or enrich existing datasets by effortlessly putting a real breast mass or micro-calcification cluster retrieved from a source digital mammography into a different region of another mammogram. Findings of a reader experiment that compared the realism of inserted lesions to clinical lesions were shown in [59]. Using the receiver operating characteristic (ROC) technique, radiologist ratings showed that injected lesions cannot be consistently discriminated from clinical lesions. Swiderski et al. [57] remarked a ResNet-50 classifier trained on GAN-augmented data produced better AUROC than when trained solely with traditionally augmented data. The adoption of data augmentation as an overfitting mitigation technique was investigated in [25]. “In silico” procedural analytic breast and breast mass modelling algorithms were used to create synthetic mammograms. They were then projected into mammographic pictures using simulated X-ray projections.

3.2. Neural Style Transfer (NST)

Deep Learning proved effective even in mixing styles out of different images. Neural Style Transfer is meaningfully representative of the quality levels achieved on this [60]. The overall goal is to alter visual representations formed in CNNs [61]. Neural Style Transfer is well known for its uses in creative application domains, but it can also be used to augment data. The technique manipulates the sequential representations across a CNN to transfer the style of one image to another while keeping the original content [1]. Gatys et al. [62] first proposed NST, which typically takes two input images: a content image C to be transferred and a style reference image S. It executes feature learning from the representations of Fl(C) and Fl(S) in layer l of a neural style transfer network [63]. However, suppose the image styles from different datasets are way too far. In that case, it may cause a wide domain gap undermining deep learning models’ capacities to target a specific scenario of interest. Wang et al. [63] proposed a multi-resolution and multi-reference NST network to address style diversity in mammograms. With very high resolution, the network can normalise styles from several vendors (e.g., GE Healthcare (GE) and United Imaging Healthcare (UIH) to the same style baseline.
A novel approach for detecting abnormal and normal regions from mammograms has been presented by Ramadan et al. [64]. They combined a cheat sheet containing standard features retrieved from ROIs with data augmentation boosting CNN performances in breast cancer detection. As a result, the accuracy rate improved by at least 12.2% and precision by at least 2.2%.
Based on the identification of masses in the projections, the authors of [65] assessed the usage of data augmentation and the selection of non-overlapping areas of interest (ROI). Zhang et al. [31] combined data augmentation and transfer learning techniques with CNN models to improve the performance of the classifiers for mammogram images.

3.3. Other Techniques

Evolutionary algorithms deal with problem solutions’ optimisations by building on Darwin’s natural selection theory. There are some crossings elements between evolutionary algorithms and the broad pool of machine learning methods for classification, regression, and clustering tasks. In this section, a specific evolutionary algorithm is described for biomedical image augmentation: crossover [66]. The crossover technique is proposed in [66] for medical image classification problems using CNNs. The method creates sample pairs through two-point crossover on already existing training datasets. From N training samples, authors could generate N new samples. The process was examined on the Mini-MIAS dataset with the VGG-16 and VGG-19 pre-trained models. The “natural deformation data augmentation approach” is proposed by Cao et al. [67] as a new data augmentation method based on local elastic deformation. The essential notion is that only the BMass is elastically deformed in a picture containing BMass to replicate the natural changing of BMass, while the local background region in contact with BMass changes accordingly. The research by Chen et al. [68] paved the way for the employability of virtual adversarial training (VAT) to improve the performance of semi-supervised classification of malignant and benign masses from mammograms. This especially applies to medical scenarios with unlabeled medical images. By employing augmentation methods such as Cutout [69] and RandConv [70], Garrucho et al. [71] ran comparisons between eight different object detection models to detect breast masses from mammogram repositories such as OPTIMAM [72], INbreast [73], BCDR [74]. Another study by Tran et al. [75] presented a Transparency strategy-based technique for generating abnormal instances by changing the pixel values. Experiments confirm the proposed approach improving the BI-RADS classification task on mammography assessments. The augmentation method was compared with The CutMix [76] augmentation approach and outperforms the same.
Table 1 summarises basic and advanced augmentation techniques with their strength and limitations. Finally, Table 2 presents a summary of methods that adopted image augmentation strategies to improve the model performance and counter overfitting. Furthermore, we also summarise some articles by highlighting pre and post-augmentation performances as well as pre and post augmentation dataset sizes (see Table 3).

4. Test-Time Augmentation (TTA)

Over the last few years, a new image augmentation technique has increasingly caught researchers’ interest. It goes under the name of TTA, which stands for Test Time Augmentation. Wang et al. [99] provided the scientific community with a mathematic formulation of TTA. They present TTA as an inference problem with hidden parameters and prior distributions. Therefore, images are considered the results of an elaboration process with hidden parameters. The final goal is to evaluate structure-wise uncertainty associated with image transformations and noise. Other than the previously mentioned techniques, TTA creates various augmented images of the test set, feeds these augmented images to the trained model, and finally returns an ensemble of those predictions to get a more assertive response [100]. Figure 11 shows the process of both train and test time augmentation, while, in Figure 12, test-time data augmentation framework is depicted.
TTA has conveyed new possibilities to the medical imaging field by measuring the strength and network consistency as practical issues [101]. TTA can be used for those methods which modify an incoming example with affine, pixel-level, or elastic transformations in the case of lesion classification from mammograms. The research community has focused on training data augmentations, while data transformation before inference has yet to be fully explored. TTA combines numerous inference findings utilising various data augmentations to categorise one image (see Figure 12). Kim et al. [100] presented a TTA method that is instance-aware and based on a loss predictor. They improved image classification performance with the dynamic use of TTA transformations. The authors of [102] employed TTA for U-Net [103] to tackle medical image segmentation. Another study [104] employed TTA with the model making predictions on five, 224 × 224 image patches, as well as horizontally reflected patches (for a total of ten patches), and then averaging the outputs on over the ten patches with the softmax layer. An inference approach called Mixup Inference (MI), reliant on simple geometric intuitions was proposed by Pang et al. [105]. The method mixes inputs with additional random samples. Vedalankar et al. [89] addressed the analysis of architectural distortion in mammograms with an integrated solution based on AlexNet and SVM. However, the solution heavily relies on TTA as the data augmentation technique for mammogram images.

5. Discussion

This section discusses data augmentation and its employment in mammogram analysis and related tasks. The paper spans the main data augmentation approaches as listed in Table 1. Most of them build on geometric transformations, noise injection, kernel filters, mixing images, random erasing, generative adversarial training, and neural style transfer. In Figure 4, Figure 5 and Figure 7 examples of basic geometric transformations and image filtering show how simple operations and modifications allow increasing datasets’ volumes. Conversely, things gradually become more complex when surveying advanced image augmentation methods. In Figure 8, Figure 9 and Figure 10, pictures respectively taken from Shen et al.’s [55], and Korkinot et al.’s [58] demonstrate the level of sophistication achieved by more recently introduced deep learning techniques such as GANs in generating realistic mammograms. It is not straightforward to discriminate synthetic images from real ones, especially to non-expert eyes. Apart from the considerations mentioned above, it is necessary to span the performances of those models that heavily rely on data augmentation, as shown in Table 2. Around thirty methods tackling tasks such as mammogram classification, suspicious region segmentation and micro-calcification identification are compared across several parameters. The following subsection checks through two main aspects: the data augmentation impact on the dataset sizes, and the pre and post-augmentation performances of the approaches as depicted in Table 2 and Table 3.

Data Augmentation Impact

Although several methods achieve decent accuracy rates over several datasets, three main points are to be highlighted:
1.
Oyelade and Ezugwu [8] achieved a 93.75% accuracy rate on anomaly detection from mammograms using a CNN-based technique and basic data augmentation techniques (rotation by 90, 180 and 270 degrees, mirroring and additive Poisson noise).
2.
Conditional infilling GANs for data augmentation in mammogram classification by Dhivya et al. [10] averagely scored 94% accuracy over three different datasets, respectively, MIAS, INBreast and DDSM, which include images having heterogeneous spatial resolution and acquiring device properties. The same method gets to an 88% accuracy rate when only basic data augmentation techniques are adopted.
3.
Razali et al. [84] reached an excellent 99% accuracy rate on InBreast and DDSM with basic augmentation techniques on two datasets. However, it would be worth investigating any further improvement with advanced data augmentation techniques. However, after surveying all methods in Table 2, it is noticeable how advanced mammogram augmentation impacts the accuracy rate improvement by 6% over three different datasets. Investigating all elements causing an increase in accuracy on a specific task is not trivial.
Table 3 allows comparing several methods according to the data augmentation impact on dataset size and performances over different datasets. The table consists of three columns for pre-augmentation dataset size, post-augmentation dataset size, and post-augmentation model performance. Overall, regardless of the specific augmentation technique employed, the increasing factor for the datasets is remarkably high. Spanning all methods in Table 3, the dataset in [81] got its size increased by a factor of around 5, while in [26] the same factor goes up to 7. Adegiba et al. [47] successfully got the number of patches up to 29 times the original size. Each image in [10] remarkably turns into 546 new samples, and the post-augmentation performances improved by almost 25%. Fourteen is the dataset size growing factor in [89], moving from 215 to 3006 ROIs (Regions of Interest). Muduli et al. [35] were able to extend MIAS, DDSM and INbreast, respectively, by 10, 19.2 and 5.4 times their original sizes.
As far as it concerns the data augmentation impact on the methods’ performances, the third column in Table 3 provides details on the accuracy, AUC, sensitivity, and auROC. Unfortunately, due to the lack of experimental results description in the original articles, only some methods can be discussed here. Starting from the fourth row in Table 3, the image co-registration augmentation technique allowed Domingues et al. [81] to gain 33% on accuracy rates. Fundamental transformations such as flipping, rotation, shifting and zooming proved effective in Shi et al.’s method [28], with a gain of 32.3% on the validation accuracy rate over MIAS. Razali et al. [84] obtained post-augmentation performances in breast cancer classification 26% higher. Only rotation, flipping and shearing were applied to INBreast and CBIS-DDSM. The GAN-based data augmentation adopted by Dhiva et al. [10] turned out to be reliable over three datasets: MIAS, DDSM, and INBreast. On the other side, the geometric transformations tested in [30] only slightly hit the target with a small improvement of 1.6% in mammogram classification with VGGNet, GoogleNet and ResNet. Lu et al. [88] reported improvements in Inception V3 sensitivity rates in mass detection over the INBreast dataset that was augmented with geometric transformations plus contrast and brightness adjustments. Rises in sensitivity rates may correspond to drops in false negatives in the detection system. The post-augmentation AUC of the classification system proposed in [31] is higher than the pre-augmentation AUC by 0.21. However, the experimental dataset is private, and the size increase factor obtained with geometric transformations is unknown. If the two advanced techniques adopted in [56,66] struggle to get remarkable classification accuracy rate improvements, the GAN-based method in [55] proved its reliability over two main tasks: mammogram image synthesis and suspicious region detection.

6. Conclusions

This paper aims to provide insights into the broader area of mammogram image analysis from a data augmentation perspective. Although some deep learning methods’ performances are excellent, further investigations are necessary to draw a line on the impact of data augmentation on the information generalisation capabilities of supervised deep learning paradigms. Some evidence shows a decisive increase in accuracy rates from basic to advanced augmentation techniques, especially the GANs-based ones. The first sections introduce the main theoretical concepts about the most widely adopted data augmentation techniques in a broader sense. In this work, the techniques mentioned above are surveyed to check out their pros and cons in a specific topic, that is, mammogram image analysis.
The advent of deep learning approaches and the increasingly sophisticated architectures to extract hidden properties from data play a critical role in various computer vision tasks. The main goal here is to discuss specifically the impact of data augmentation techniques on deep learning methods in tasks such as suspicious region detection and classification.
In Table 2 and Table 3, methods from the scientific literature are listed and compared across factors such as task performed, model, dataset, model performance, data augmentation approach. Thanks to the undisputed knowledge inference capabilities of deep learning architectures, most methods reviewed in this paper reach high accuracy rates on their tasks and over some specific datasets. Overall, most DL methods score high in architectural distortion detection, mass detection, density classification, and more generic suspicious region detection. Moreover, the contribution of data augmentation techniques is remarkable, especially to the dataset size increase and accuracy rates improvement. For instance, the tumour classification method proposed in [10] benefits from GAN-based data augmentation, an increased factor of 546 for the dataset. Furthermore, as described in Table 3, the method’s accuracy rate goes up to 94% with an improvement of almost 25% compared to the pre-augmentation performance. Apart from that, our discussion needs to consider that the latter works out mass classification over three different datasets: MIAS, INBreast and DDSM. Moreover, the so-called conventional data augmentation techniques (geometric transformations) allow up to 88% of accuracy, while the more advanced GAN-based techniques outperform them by 6%.
Current and future trends in computer vision see new methods building on self-supervised and semi-self supervised paradigms competing with supervised learning approaches. Purely supervised learning approaches combined with advanced data augmentation should, then, run against self-supervised and semi self-supervised learning methods to balance computational costs, accuracy rates, and information generalisation capabilities.

Author Contributions

Conceptualization, P.O.; Formal analysis, P.O.; Investigation, P.O.; Writing—original draft preparation, P.O.; Writing—review and editing, P.O. and A.B.; Supervision, P.S., S.P., F.A., A.B. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
PRISMAPreferred Reporting Items for Systematic Reviews and Meta-Analyses
DLDeep Learning
CNNConvolutional Neural Network
TLTransfer Learning
CADComputer-Aided Diagnosis
BI-RADSBreast Imaging Reporting and Database System
ROIRegion of Interest
NSTNeural Style Transfer
GANGenerative Adversarial Network
AUCArea Under Curve
ROCReceiver Operating Characteristic
ADArchitectural Distortion
TTATest Time Augmentation
VATVirtual Adversarial Training
UIHUnited Imaging Healthcare

References

  1. Shorten, C.; Khoshgoftaar, T.M. A survey on image data augmentation for deep learning. J. Big Data 2019, 6, 60. [Google Scholar] [CrossRef]
  2. Oza, P.; Sharma, P.; Patel, S. Machine learning applications for computer-aided medical diagnostics. In Proceedings of Second International Conference on Computing, Communications, and Cyber-Security; Springer: Singapore, 2021; pp. 377–392. [Google Scholar]
  3. Pillai, R.; Oza, P.; Sharma, P. Review of machine learning techniques in health care. In Proceedings of ICRIC 2019; Springer: Cham, Switzerland, 2020; pp. 103–111. [Google Scholar]
  4. Oza, P.; Sharma, P.; Patel, S.; Kumar, P. Deep convolutional neural networks for computer-aided breast cancer diagnostic: A survey. Neural Comput. Appl. 2022, 34, 1815–1836. [Google Scholar] [CrossRef]
  5. Chlap, P.; Min, H.; Vandenberg, N.; Dowling, J.; Holloway, L.; Haworth, A. A review of medical image data augmentation techniques for deep learning applications. J. Med. Imaging Radiat. Oncol. 2021, 65, 545–563. [Google Scholar] [CrossRef]
  6. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  7. Patel, H.J.; Oza, P.; Agrawal, S. AI Approaches for Breast Cancer Diagnosis: A Comprehensive Study. In International Conference on Innovative Computing and Communications; Springer: Singapore, 2022; pp. 393–419. [Google Scholar]
  8. Oyelade, O.N.; Ezugwu, A.E. A deep learning model using data augmentation for detection of architectural distortion in whole and patches of images. Biomed. Signal Process. Control 2021, 65, 102366. [Google Scholar] [CrossRef]
  9. Wu, E.; Wu, K.; Cox, D.; Lotter, W. Conditional infilling GANs for data augmentation in mammogram classification. In Image Analysis for Moving Organ, Breast, and Thoracic Images; Springer: Cham, Switzerland, 2018; pp. 98–106. [Google Scholar]
  10. Dhivya, S.; Mohanavalli, S.; Karthika, S.; Shivani, S.; Mageswari, R. GAN based Data Augmentation for Enhanced Tumor Classification. In Proceedings of the 2020 4th International Conference on Computer, Communication and Signal Processing (ICCCSP), Chennai, India, 28–29 September 2020; pp. 1–5. [Google Scholar] [CrossRef]
  11. Oza, P.; Shah, Y.; Vegda, M. A Comprehensive Study of Mammogram Classification Techniques. In Tracking and Preventing Diseases with Artificial Intelligence; Springer: Cham, Switzerland, 2022; pp. 217–238. [Google Scholar]
  12. Bruno, A.; Ardizzone, E.; Vitabile, S.; Midiri, M. A novel solution based on scale invariant feature transform descriptors and deep learning for the detection of suspicious regions in mammogram images. J. Med. Signals Sens. 2020, 10, 158. [Google Scholar]
  13. Nalepa, J.; Marcinkiewicz, M.; Kawulok, M. Data augmentation for brain-tumor segmentation: A review. Front. Comput. Neurosci. 2019, 13, 83. [Google Scholar] [CrossRef] [Green Version]
  14. Bakas, S.; Reyes, M.; Jakab, A.; Bauer, S.; Rempfler, M.; Crimi, A.; Shinohara, R.T.; Berger, C.; Ha, S.M.; Rozycki, M.; et al. Identifying the best machine learning algorithms for brain tumor segmentation, progression assessment, and overall survival prediction in the BRATS challenge. arXiv 2018, arXiv:1811.02629. [Google Scholar]
  15. Moher, D.; Liberati, A.; Tetzlaff, J.; Altman, D.G.; the PRISMA Group. Preferred reporting items for systematic reviews and meta-analyses: The PRISMA statement. Ann. Intern. Med. 2009, 151, 264–269. [Google Scholar] [CrossRef] [Green Version]
  16. Wong, S.C.; Gatt, A.; Stamatescu, V.; McDonnell, M.D. Understanding data augmentation for classification: When to warp? In Proceedings of the 2016 International Conference on Digital Image Computing: Techniques and Applications (DICTA), Gold Coast, QLD, Australia, 30 November–2 December 2016; pp. 1–6. [Google Scholar]
  17. Alzubaidi, L.; Zhang, J.; Humaidi, A.J.; Al-Dujaili, A.; Duan, Y.; Al-Shamma, O.; Santamaría, J.; Fadhel, M.A.; Al-Amidie, M.; Farhan, L. Review of deep learning: Concepts, CNN architectures, challenges, applications, future directions. J. Big Data 2021, 8, 1–74. [Google Scholar] [CrossRef]
  18. Rosebrock, A. Keras ImageDataGenerator and Data Augmentation. Available online: https://pyimagesearch.com/2019/07/08/keras-imagedatagenerator-and-data-augmentation/ (accessed on 20 February 2022).
  19. Shin, H.C.; Tenenholtz, N.A.; Rogers, J.K.; Schwarz, C.G.; Senjem, M.L.; Gunter, J.L.; Andriole, K.P.; Michalski, M. Medical image synthesis for data augmentation and anonymization using generative adversarial networks. In Proceedings of the International Workshop on Simulation and Synthesis in Medical Imaging, Granada, Spain, 16 September 2018; Springer: Cham, Switzerland, 2018; pp. 1–11. [Google Scholar]
  20. Heath, M.; Bowyer, K.; Kopans, D.; Kegelmeyer, P.; Moore, R.; Chang, K.; Munishkumaran, S. Current status of the digital database for screening mammography. In Digital Mammography; Springer: Dordrecht, The Netherlands, 1998; pp. 457–460. [Google Scholar]
  21. Lee, R.S.; Gimenez, F.; Hoogi, A.; Miyake, K.K.; Gorovoy, M.; Rubin, D.L. A curated mammography data set for use in computer-aided detection and diagnosis research. Sci. Data 2017, 4, 170177. [Google Scholar] [CrossRef] [PubMed]
  22. Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Fei-Fei, L. ImageNet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar] [CrossRef] [Green Version]
  23. Suckling, J.; Parker, J.; Dance, D.; Astley, S.; Hutt, I.; Boggis, C.; Ricketts, I.; Stamatakis, E.; Cerneaz, N.; Kok, S.; et al. Mammographic Image Analysis Society (mias) Database v1. 21. International Congress Series. 2015. Available online: https://www.repository.cam.ac.uk/handle/1810/250394 (accessed on 1 January 2022).
  24. Costa, A.C.; Oliveira, H.C.; Vieira, M.A. Data augmentation: Effect in deep convolutional neural network for the detection of architectural distortion in digital mammography. In Proceedings of the Brazilian Congress on Medical Physics, Santos, Brazil, 21–24 August 2019. [Google Scholar]
  25. Cha, K.H.; Petrick, N.A.; Pezeshk, A.X.; Graff, C.G.; Sharma, D.; Badal, A.; Sahiner, B. Evaluation of data augmentation via synthetic images for improved breast mass detection on mammograms using deep learning. J. Med. Imaging 2019, 7, 012703. [Google Scholar] [CrossRef] [PubMed]
  26. Omonigho, E.L.; David, M.; Adejo, A.; Aliyu, S. Breast cancer: Tumor detection in mammogram images using modified alexnet deep convolution neural network. In Proceedings of the 2020 International Conference in Mathematics, Computer Engineering and Computer Science (ICMCECS), Ayobo, Nigeria, 18–21 March 2020; pp. 1–6. [Google Scholar]
  27. Rahman, A.S.A.; Belhaouari, S.B.; Bouzerdoum, A.; Baali, H.; Alam, T.; Eldaraa, A.M. Breast mass tumor classification using deep learning. In Proceedings of the 2020 IEEE International Conference on Informatics, IoT, and Enabling Technologies (ICIoT), Doha, Qatar, 2–5 February 2020; pp. 271–276. [Google Scholar]
  28. Shi, P.; Wu, C.; Zhong, J.; Wang, H. Deep learning from small dataset for BI-RADS density classification of mammography images. In Proceedings of the 2019 10th International Conference on Information Technology in Medicine and Education (ITME), Qingdao, China, 23–25 August 2019; pp. 102–109. [Google Scholar]
  29. Magny, S.J.; Shikhman, R.; Keppke, A.L. Breast, Imaging, Reporting and Data System (BI RADS); StatPearls: Treasure Island, FL, USA, 2020. [Google Scholar]
  30. Khan, H.N.; Shahid, A.R.; Raza, B.; Dar, A.H.; Alquhayz, H. Multi-view feature fusion based four views model for mammogram classification using convolutional neural network. IEEE Access 2019, 7, 165724–165733. [Google Scholar] [CrossRef]
  31. Zhang, X.; Zhang, Y.; Han, E.Y.; Jacobs, N.; Han, Q.; Wang, X.; Liu, J. Whole mammogram image classification with convolutional neural networks. In Proceedings of the 2017 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Kansas City, MO, USA, 13–16 November 2017; pp. 700–704. [Google Scholar]
  32. Suckling, J.P. The mammographic image analysis society digital mammogram database. Digital Mammo 1994, 375–386. Available online: http://peipa.essex.ac.uk/info/mias.html (accessed on 1 January 2022).
  33. Assari, Z.; Mahloojifar, A.; Ahmadinejad, N. A bimodal BI-RADS-guided GoogLeNet-based CAD system for solid breast masses discrimination using transfer learning. Comput. Biol. Med. 2022, 142, 105160. [Google Scholar] [CrossRef]
  34. Aly, G.H.; Marey, M.; El-Sayed, S.A.; Tolba, M.F. YOLO based breast masses detection and classification in full-field digital mammograms. Comput. Methods Programs Biomed. 2021, 200, 105823. [Google Scholar] [CrossRef]
  35. Muduli, D.; Dash, R.; Majhi, B. Automated diagnosis of breast cancer using multi-modal datasets: A deep convolution neural network based approach. Biomed. Signal Process. Control 2022, 71, 102825. [Google Scholar] [CrossRef]
  36. Li, Z.; Cui, Z.; Wang, S.; Qi, Y.; Ouyang, X.; Chen, Q.; Yang, Y.; Xue, Z.; Shen, D.; Cheng, J.Z. Domain Generalization for Mammography Detection via Multi-style and Multi-view Contrastive Learning. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Strasbourg, France, 27 September–1 October 2021; Springer: Cham, Switzerland, 2021; pp. 98–108. [Google Scholar]
  37. Agarwal, M.; Mahajan, R. Medical images contrast enhancement using quad weighted histogram equalization with adaptive gama correction and homomorphic filtering. Procedia Comput. Sci. 2017, 115, 509–517. [Google Scholar] [CrossRef]
  38. Sahnoun, M.; Kallel, F.; Dammak, M.; Mhiri, C.; Mahfoudh, K.B.; Hamida, A.B. A comparative study of MRI contrast enhancement techniques based on Traditional Gamma Correction and Adaptive Gamma Correction: Case of multiple sclerosis pathology. In Proceedings of the 2018 4th International Conference on Advanced Technologies for Signal and Image Processing (ATSIP), Sousse, Tunisia, 21–24 March 2018; pp. 1–7. [Google Scholar]
  39. Galdran, A.; Alvarez-Gila, A.; Meyer, M.I.; Saratxaga, C.L.; Araújo, T.; Garrote, E.; Aresta, G.; Costa, P.; Mendonça, A.M.; Campilho, A. Data-driven color augmentation techniques for deep skin image analysis. arXiv 2017, arXiv:1703.03702. [Google Scholar]
  40. Min, H.; Wilson, D.; Huang, Y.; Liu, S.; Crozier, S.; Bradley, A.P.; Chandra, S.S. Fully automatic computer-aided mass detection and segmentation via pseudo-color mammograms and mask r-cnn. In Proceedings of the 2020 IEEE 17th International Symposium on Biomedical Imaging (ISBI), Iowa City, IA, USA, 3–7 April 2020; pp. 1111–1115. [Google Scholar]
  41. Zhong, Z.; Zheng, L.; Kang, G.; Li, S.; Yang, Y. Random erasing data augmentation. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 13001–13008. [Google Scholar]
  42. Kang, G.; Dong, X.; Zheng, L.; Yang, Y. Patchshuffle regularization. arXiv 2017, arXiv:1707.07103. [Google Scholar]
  43. Netzer, Y.; Wang, T.; Coates, A.; Bissacco, A.; Wu, B.; Ng, A.Y. Reading Digits in Natural Images with Unsupervised Feature Learning. 2011. Available online: http://ufldl.stanford.edu/housenumbers/nips2011_housenumbers.pdf (accessed on 14 March 2022).
  44. LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef] [Green Version]
  45. Krizhevsky, A.; Hinton, G. Learning Multiple Layers of Features from Tiny Images. 2009. Available online: http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.222.9220&rep=rep1&type=pdf (accessed on 15 December 2021).
  46. Coates, A.; Ng, A.; Lee, H. An analysis of single-layer networks in unsupervised feature learning. In Proceedings of the Fourteenth International Conference on Artificial Intelligence and Statistics, Fort Lauderdale, FL, USA, 11–13 April 2011; pp. 215–223. [Google Scholar]
  47. Adedigba, A.P.; Adeshinat, S.A.; Aibinu, A.M. Deep learning-based mammogram classification using small dataset. In Proceedings of the 2019 15th International Conference on Electronics, Computer and Computation (ICECCO), Abuja, Nigeria, 10–12 December 2019; pp. 1–6. [Google Scholar]
  48. Yemini, M.; Zigel, Y.; Lederman, D. Detecting masses in mammograms using convolutional neural networks and transfer learning. In Proceedings of the 2018 IEEE International Conference on the Science of Electrical Engineering in Israel (ICSEE), Eilat, Israel, 12–14 December 2018; pp. 1–4. [Google Scholar]
  49. Oza, P.; Sharma, P.; Patel, S.; Bruno, A. A Bottom-Up Review of Image Analysis Methods for Suspicious Region Detection in Mammograms. J. Imaging 2021, 7, 190. [Google Scholar] [CrossRef] [PubMed]
  50. Gao, M.; Fessler, J.A.; Chan, H.P. Deep Convolutional Neural Network with Adversarial Training for Denoising Digital Breast Tomosynthesis Images. IEEE Trans. Med. Imaging 2021, 40, 1805–1816. [Google Scholar] [CrossRef] [PubMed]
  51. Shen, R.; Yao, J.; Yan, K.; Tian, K.; Jiang, C.; Zhou, K. Unsupervised domain adaptation with adversarial learning for mass detection in mammogram. Neurocomputing 2020, 393, 27–37. [Google Scholar] [CrossRef]
  52. Shen, L.; Margolies, L.R.; Rothstein, J.H.; Fluder, E.; McBride, R.; Sieh, W. Deep learning to improve breast cancer detection on screening mammography. Sci. Rep. 2019, 9, 12495. [Google Scholar] [CrossRef]
  53. Desai, S.D.; Giraddi, S.; Verma, N.; Gupta, P.; Ramya, S. Breast Cancer Detection Using GAN for Limited Labeled Dataset. In Proceedings of the 2020 12th International Conference on Computational Intelligence and Communication Networks (CICN), Bhimtal, India, 25–26 September 2020; pp. 34–39. [Google Scholar] [CrossRef]
  54. Singh, V.K.; Rashwan, H.A.; Romani, S.; Akram, F.; Pandey, N.; Sarker, M.M.K.; Saleh, A.; Arenas, M.; Arquez, M.; Puig, D.; et al. Breast tumor segmentation and shape classification in mammograms using generative adversarial and convolutional neural network. Expert Syst. Appl. 2020, 139, 112855. [Google Scholar] [CrossRef]
  55. Shen, T.; Hao, K.; Gou, C.; Wang, F.Y. Mass Image Synthesis in Mammogram with Contextual Information Based on GANs. Comput. Methods Programs Biomed. 2021, 202, 106019. [Google Scholar] [CrossRef]
  56. Muramatsu, C.; Nishio, M.; Goto, T.; Oiwa, M.; Morita, T.; Yakami, M.; Kubo, T.; Togashi, K.; Fujita, H. Improving breast mass classification by shared data with domain transformation using a generative adversarial network. Comput. Biol. Med. 2020, 119, 103698. [Google Scholar] [CrossRef]
  57. Swiderski, B.; Gielata, L.; Olszewski, P.; Osowski, S.; Kołodziej, M. Deep neural system for supporting tumor recognition of mammograms using modified GAN. Expert Syst. Appl. 2021, 164, 113968. [Google Scholar] [CrossRef]
  58. Korkinof, D.; Heindl, A.; Rijken, T.; Harvey, H.; Glocker, B. MammoGAN: High-resolution synthesis of realistic mammograms. In Proceedings of the International Conference on Medical Imaging with Deep Learning–Extended Abstract Track, London, UK, 8–10 July 2019. [Google Scholar]
  59. Pezeshk, A.; Petrick, N.; Sahiner, B. Seamless lesion insertion in digital mammography: Methodology and reader study. In Medical Imaging 2016: Computer-Aided Diagnosis; Tourassi, G.D., Armato, S.G., III, Eds.; International Society for Optics and Photonics, SPIE: Bellingham, WA, USA, 2016; Volume 9785, pp. 134–139. [Google Scholar] [CrossRef]
  60. Jing, Y.; Yang, Y.; Feng, Z.; Ye, J.; Yu, Y.; Song, M. Neural Style Transfer: A Review. IEEE Trans. Vis. Comput. Graph. 2020, 26, 3365–3385. [Google Scholar] [CrossRef] [Green Version]
  61. Gatys, L.A.; Ecker, A.S.; Bethge, M. A neural algorithm of artistic style. arXiv 2015, arXiv:1508.06576. [Google Scholar] [CrossRef]
  62. Gatys, L.A.; Ecker, A.S.; Bethge, M. Image style transfer using convolutional neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2414–2423. [Google Scholar]
  63. Wang, S.; Huo, J.; Ouyang, X.; Che, J.; Xue, Z.; Shen, D.; Wang, Q.; Cheng, J.Z. mr2 NST: Multi-resolution and Multi-reference Neural Style Transfer for Mammography. In Proceedings of the International Workshop on Predictive Intelligence in Medicine, Lima, Peru, 8 October 2020; Springer: Cham, Switzerland, 2020; pp. 169–177. [Google Scholar]
  64. Ramadan, S.Z. Using convolutional neural network with cheat sheet and data augmentation to detect breast cancer in mammograms. Comput. Math. Methods Med. 2020, 2020, 9523404. [Google Scholar] [CrossRef] [PubMed]
  65. Zeiser, F.A.; da Costa, C.A.; Zonta, T.; Marques, N.; Roehe, A.V.; Moreno, M.; da Rosa Righi, R. Segmentation of masses on mammograms using data augmentation and deep learning. J. Digit. Imaging 2020, 33, 858–868. [Google Scholar] [CrossRef] [PubMed]
  66. Raj, R.; Mathew, J.; Kannath, S.K.; Rajan, J. Crossover based technique for data augmentation. Comput. Methods Programs Biomed. 2022, 218, 106716. [Google Scholar] [CrossRef] [PubMed]
  67. Cao, H.; Pu, S.; Tan, W.; Tong, J. Breast mass detection in digital mammography based on anchor-free architecture. Comput. Methods Programs Biomed. 2021, 205, 106033. [Google Scholar] [CrossRef] [PubMed]
  68. Chen, X.; Wang, X.; Zhang, K.; Fung, K.M.; Thai, T.C.; Moore, K.; Mannel, R.S.; Liu, H.; Zheng, B.; Qiu, Y. Virtual Adversarial Training for Semi-supervised Breast Mass Classification. arXiv 2022, arXiv:2201.10675. [Google Scholar]
  69. DeVries, T.; Taylor, G.W. Improved regularization of convolutional neural networks with cutout. arXiv 2017, arXiv:1708.04552. [Google Scholar]
  70. Xu, Z.; Liu, D.; Yang, J.; Raffel, C.; Niethammer, M. Robust and generalizable visual representation learning via random convolutions. arXiv 2020, arXiv:2007.13003. [Google Scholar]
  71. Garrucho, L.; Kushibar, K.; Jouide, S.; Diaz, O.; Igual, L.; Lekadir, K. Domain generalization in deep learning-based mass detection in mammography: A large-scale multi-center study. arXiv 2022, arXiv:2201.11620. [Google Scholar]
  72. Halling-Brown, M.D.; Warren, L.M.; Ward, D.; Lewis, E.; Mackenzie, A.; Wallis, M.G.; Wilkinson, L.S.; Given-Wilson, R.M.; McAvinchey, R.; Young, K.C. OPTIMAM Mammography image database: A large-scale resource of mammography images and clinical data. Radiol. Artif. Intell. 2020, 3, e200103. [Google Scholar] [CrossRef]
  73. Moreira, I.C.; Amaral, I.; Domingues, I.; Cardoso, A.; Cardoso, M.J.; Cardoso, J.S. Inbreast: Toward a full-field digital mammographic database. Acad. Radiol. 2012, 19, 236–248. [Google Scholar] [CrossRef] [Green Version]
  74. Moura, D.C.; López, M.A.G.; Cunha, P.; Posada, N.G.d.; Pollan, R.R.; Ramos, I.; Loureiro, J.P.; Moreira, I.C.; Araújo, B.M.; Fernandes, T.C. Benchmarking datasets for breast cancer computer-aided diagnosis (CADx). In Proceedings of the Iberoamerican Congress on Pattern Recognition, Havana, Cuba, 20–23 November 2013; Springer: Berlin/Heidelberg, Germany, 2013; pp. 326–333. [Google Scholar]
  75. Tran, S.B.; Nguyen, H.T.; Pham, H.H.; Nguyen, H.Q. Transparency strategy-based data augmentation for BI-RADS classification of mammograms. arXiv 2022, arXiv:2203.10609. [Google Scholar]
  76. Yun, S.; Han, D.; Oh, S.J.; Chun, S.; Choe, J.; Yoo, Y. Cutmix: Regularization strategy to train strong classifiers with localizable features. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27–28 October 2019; pp. 6023–6032. [Google Scholar]
  77. Moreno-Barea, F.J.; Strazzera, F.; Jerez, J.M.; Urda, D.; Franco, L. Forward noise adjustment scheme for data augmentation. In Proceedings of the 2018 IEEE Symposium Series on Computational Intelligence (SSCI), Bangalore, India, 18–21 November 2018; pp. 728–734. [Google Scholar]
  78. Naveed, H. Survey: Image mixing and deleting for data augmentation. arXiv 2021, arXiv:2106.07085. [Google Scholar]
  79. Baek, F.; Park, S.; Kim, H. Data augmentation using adversarial training for construction-equipment classification. arXiv 2019, arXiv:1911.11916. [Google Scholar]
  80. Antoniou, A.; Storkey, A.; Edwards, H. Data augmentation generative adversarial networks. arXiv 2017, arXiv:1711.04340. [Google Scholar]
  81. Domingues, I.; Abreu, P.H.; Santos, J. Bi-rads classification of breast cancer: A new pre-processing pipeline for deep models training. In Proceedings of the 2018 25th IEEE International Conference on Image Processing (ICIP), Athens, Greece, 7–10 October 2018; pp. 1378–1382. [Google Scholar]
  82. Zhang, X.; Zhang, Y.; Han, E.Y.; Jacobs, N.; Han, Q.; Wang, X.; Liu, J. Classification of whole mammogram and tomosynthesis images using deep convolutional neural networks. IEEE Trans. Nanobiosci. 2018, 17, 237–242. [Google Scholar] [CrossRef] [PubMed]
  83. Castro, E.; Cardoso, J.S.; Pereira, J.C. Elastic deformations for data augmentation in breast cancer mass detection. 2018 IEEE EMBS International Conference on Biomedical & Health Informatics (BHI), Las Vegas, NV, USA, 4–7 March 2018; pp. 230–234. [Google Scholar]
  84. Razali, N.F.; Isa, I.S.; Sulaiman, S.N.; Karim, N.K.A.; Osman, M.K. High-level Features in Deeper Deep Learning Layers for Breast Cancer Classification. In Proceedings of the 2021 11th IEEE International Conference on Control System, Computing and Engineering (ICCSCE), Penang, Malaysia, 27–28 August 2021; pp. 170–175. [Google Scholar]
  85. Alkhaleefah, M.; Chittem, P.K.; Achhannagari, V.P.; Ma, S.C.; Chang, Y.L. The influence of image augmentation on breast lesion classification using transfer learning. In Proceedings of the 2020 International Conference on Artificial Intelligence and Signal Processing (AISP), Amaravati, India, 10–12 January 2020; pp. 1–5. [Google Scholar]
  86. Wang, Y.; Song, M.; Tian, X. Meta-learning Based Breast Abnormality Classification on Screening Mammograms. In Proceedings of the 2021 International Conference on Computer Engineering and Application (ICCEA), Kunming, China, 25–27 June 2021; pp. 153–156. [Google Scholar]
  87. Bagchi, S.; Mohd, M.N.H.; Debnath, S.K.; Nafea, M.; Suriani, N.S.; Nizam, Y. Performance Comparison of Pre-trained Residual Networks for Classification of the Whole Mammograms with Smaller Dataset. In Proceedings of the 2020 IEEE Student Conference on Research and Development (SCOReD), Batu Pahat, Malaysia, 27–29 September 2020; pp. 368–373. [Google Scholar]
  88. Lu, H.C.; Loh, E.W.; Huang, S.C. The classification of mammogram using convolutional neural network with specific image preprocessing for breast cancer detection. In Proceedings of the 2019 2nd International Conference on Artificial Intelligence and Big Data (ICAIBD), Chengdu, China, 25–28 May 2019; pp. 9–12. [Google Scholar]
  89. Vedalankar, A.V.; Gupta, S.S.; Manthalkar, R.R. Addressing architectural distortion in mammogram using AlexNet and support vector machine. Inform. Med. Unlocked 2021, 23, 100551. [Google Scholar] [CrossRef]
  90. Huang, M.L.; Lin, T.Y. Dataset of breast mammography images with masses. Data Brief 2020, 31, 105928. [Google Scholar] [CrossRef]
  91. Agarwal, R.; Díaz, O.; Yap, M.H.; Llado, X.; Marti, R. Deep learning for mass detection in Full Field Digital Mammograms. Comput. Biol. Med. 2020, 121, 103774. [Google Scholar] [CrossRef]
  92. Chougrad, H.; Zouaki, H.; Alheyane, O. Multi-label transfer learning for the early diagnosis of breast cancer. Neurocomputing 2020, 392, 168–180. [Google Scholar] [CrossRef]
  93. Ting, F.F.; Tan, Y.J.; Sim, K.S. Convolutional neural network improvement for breast cancer classification. Expert Syst. Appl. 2019, 120, 103–115. [Google Scholar] [CrossRef]
  94. End-to-end object detection with transformers. In European Conference on Computer Vision; Springer: Cham, Switzerland, 2020.
  95. Azevedo, V.; Silva, C.; Dutra, I. Quantum transfer learning for breast cancer detection. Quantum Mach. Intell. 2022, 4, 5. [Google Scholar] [CrossRef] [PubMed]
  96. Kolchev, A.; Pasynkov, D.; Egoshin, I.; Kliouchkin, I.; Pasynkova, O.; Tumakov, D. YOLOv4-Based CNN Model versus Nested Contours Algorithm in the Suspicious Lesion Detection on the Mammography Image: A Direct Comparison in the Real Clinical Settings. J. Imaging 2022, 8, 88. [Google Scholar] [CrossRef] [PubMed]
  97. Sadad, T.; Khan, A.R.; Hussain, A.; Tariq, U.; Fati, S.M.; Bahaj, S.A.; Munir, A. Internet of medical things embedding deep learning with data augmentation for mammogram density classification. Microsc. Res. Tech. 2021, 84, 2186–2194. [Google Scholar] [CrossRef] [PubMed]
  98. Susilo, A.B.; Sugiharti, E. Accuracy Enhancement in Early Detection of Breast Cancer on Mammogram Images with Convolutional Neural Network (CNN) Methods using Data Augmentation and Transfer Learning. J. Adv. Inf. Syst. Technol. 2021, 3, 9–16. [Google Scholar]
  99. Wang, G.; Li, W.; Aertsen, M.; Deprest, J.; Ourselin, S.; Vercauteren, T. Aleatoric uncertainty estimation with test-time augmentation for medical image segmentation with convolutional neural networks. Neurocomputing 2019, 338, 34–45. [Google Scholar] [CrossRef] [PubMed]
  100. Kim, I.; Kim, Y.; Kim, S. Learning loss for test-time augmentation. arXiv 2020, arXiv:2010.11422. [Google Scholar]
  101. Wang, G.; Li, W.; Ourselin, S.; Vercauteren, T. Automatic brain tumor segmentation using convolutional neural networks with test-time augmentation. In Proceedings of the International MICCAI Brainlesion Workshop, Granada, Spain, 16 September 2018; Springer: Cham, Switzerland, 2018; pp. 61–72. [Google Scholar]
  102. Isensee, F.; Petersen, J.; Klein, A.; Zimmerer, D.; Jaeger, P.F.; Kohl, S.; Wasserthal, J.; Koehler, G.; Norajitra, T.; Wirkert, S.; et al. nnu-net: Self-adapting framework for u-net-based medical image segmentation. arXiv 2018, arXiv:1809.10486. [Google Scholar]
  103. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; Springer: Cham, Switzerland, 2015; pp. 234–241. [Google Scholar]
  104. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25, 1097–1105. [Google Scholar] [CrossRef]
  105. Pang, T.; Xu, K.; Zhu, J. Mixup inference: Better exploiting mixup to defend adversarial attacks. arXiv 2019, arXiv:1909.11515. [Google Scholar]
Figure 1. (a) Shows the ideal trend of the model with training and validation error functions decreasing almost simultaneously. (b) Shows the undesired effect of overfitting, having the training error decrease and, conversely, validation error increases suddenly.
Figure 1. (a) Shows the ideal trend of the model with training and validation error functions decreasing almost simultaneously. (b) Shows the undesired effect of overfitting, having the training error decrease and, conversely, validation error increases suddenly.
Jimaging 08 00141 g001
Figure 2. Methods to tackle overfitting.
Figure 2. Methods to tackle overfitting.
Jimaging 08 00141 g002
Figure 3. The PRISMA flow diagram and the selection method.
Figure 3. The PRISMA flow diagram and the selection method.
Jimaging 08 00141 g003
Figure 4. Example of images after applying geometric transformation.
Figure 4. Example of images after applying geometric transformation.
Jimaging 08 00141 g004
Figure 5. An example of patch of Mammogram and a sample of patches generated with geometric transformation [12].
Figure 5. An example of patch of Mammogram and a sample of patches generated with geometric transformation [12].
Jimaging 08 00141 g005
Figure 6. Mammograms after applying random erasing.
Figure 6. Mammograms after applying random erasing.
Jimaging 08 00141 g006
Figure 7. Example of data augmentation based on various filters and noise.
Figure 7. Example of data augmentation based on various filters and noise.
Jimaging 08 00141 g007
Figure 8. Given mask image (A); normal mammogram image (B); generated mammogram image with synthetic mask (C) [55].
Figure 8. Given mask image (A); normal mammogram image (B); generated mammogram image with synthetic mask (C) [55].
Jimaging 08 00141 g008
Figure 9. Private Dataset: Given mask image (A); normal mammogram image (B); generated mammogram image with synthetic mask (C) [55].
Figure 9. Private Dataset: Given mask image (A); normal mammogram image (B); generated mammogram image with synthetic mask (C) [55].
Jimaging 08 00141 g009
Figure 10. Randomly sampled examples of original and synthetic mammograms [58].
Figure 10. Randomly sampled examples of original and synthetic mammograms [58].
Jimaging 08 00141 g010
Figure 11. Train and test-time data augmentation.
Figure 11. Train and test-time data augmentation.
Jimaging 08 00141 g011
Figure 12. Test-time data augmentation framework.
Figure 12. Test-time data augmentation framework.
Jimaging 08 00141 g012
Table 1. Summary of basic and advanced image augmentation techniques.
Table 1. Summary of basic and advanced image augmentation techniques.
Sr No.DA TechniqueSub CategoryLabel PreservingStrengthLimitation
1Geometric
Transformation [1,5]
FlippingNoGood solutions for positional bias present
in training data. Easy implementation
Additional memory, Transformation
compute cost, Additional training
time, Manual observation
CroppingNot always
RotationNot always
TranslationYes
2Noise Injection [77]-YesAllows model to learn more robustDifficult to decide amount of noise
to be added
3Kernel Filters [1]-YesGood to generate sharpen and
blurred images
Similar to CNN mechanism
4Mixing Images [78]-No-Makes not much sense from human
perspective. Not suitable for medical
images
5Random Erasing [41]-Not alwaysAnalogous to dropout regularization.
Designed to combat image recognition
challenges due to occlusion, A promising
technique to guarantee a network pays
attention to the entire image, not a subset
of it
Some manual intervention may be
necessary depending on the dataset
and application
6Adversarial Training [79]-YesHelp to illustrate weak decision boundaries
better than standard classification metrics
Less explored
7Generative Adversarial Network [80]-YesGANs generate data that looks similar to
original data
Harder to train, Generating results
from text or speech is very complex.
8Neural Style Transfer [60]--Improves the generalization ability of
simulated datasets
Efforts needed to select style,
Additional memory,
transformation cost
Table 2. Summary of articles using Image augmentation.
Table 2. Summary of articles using Image augmentation.
Ref.Task PerformedModelDatasetModel PerformanceData Augmentation Approach
[24]AD detectionDeep CNN
(Augmented CNN-SW+)
PrivateAUC: 0.83 ± 0.14Rotation by 90, 180 and 270 degrees,
mirroring and adding Poisson noise
[8]AD detectionDeep CNNMIAS, DDSM, INBreastAccuracy: 93.75%Rotation, flipping, shear, scaling, etc.
[25]Mass detectionFaster R-CNNCBIS-DDSMSensitivity: 0.833 ± 0.038Horizontal and Vertical Flipping
[63]Mass detectionmr2NSTmammograms from
GE and UIH
-Neural Style Transfer
[81]BI-RADS ClassificationAlexNetINBreastAccuracy: 83.4Image co-registration
[26]Tumor detectionModified AlexNetMIAS95.70%Scaling, horizontal flip,
rotation (90, 180, 270)
[27]Mass ClassificationInceptionV3 and ResNet50DDSMAccuracy:
InceptionV3: 79.6
ResNet50-85.71
Geometric Transformation
[82]Mammogram classificationPre-trained CNN ArchitecturesPrivate-Reflection and Rotation
[28]BI-RADS classificationCNNMIASAccuracy: 83.6%Flip, rotation, shift and zoom
[47]Mammogram
Classification
Pre-trained CNN ArchitecturesMIASAccuracy: 99.01%Gaussian blurring, horizontal flipping,
internal refection and mild addition of
white noise
[48]Mass detectionGoogle Inception-V3INBreastROC: 0.86Gaussian noise, Flipping,
Changing image brightness
[83]Mass ClassificationVGG based DCNNINBreast, CBIS, BCRP-elastic deformations
[10]Mass ClassificationDCNNMIAS, INBreast, DDSMConventional DA techniques: 88%
GAN: 94%
GAN
[84]Mass ClassificationAlexNet, InceptionV3INBreast, CBIS-DDSMAccuracy:
INBreast:
Alexnet: 0.9892,
InceptionV3: 0.9919
CBIS-DDSM:
Alexnet: 0.6138,
InceptionV3: 0.8142
rotation, flipping, shearing
[85]Lesion
Classification
ResNet50, VGG16, VGG19CBIS-DDSMAccuracy: 90.4%Geometric transformation,
Contrast and brightness adjustment
[86]Abnormality
Classification
Meta Learning, REsnet101CBIS-DDSMAccuracy:
Meta Learning: 76%,
Resnet101: 71%
Geometric transformations
[30]Mammogram
Classification
VGGNet, GoogleNet, ResnetCBIS-DDSM, MIASAUC: 0.932Geometric transformations
[87]Mammogram
Classification
Residual NetworksINBreastSpecificity: 0.89Rotation, Translation
[88]Mass detectionInceptionV3INBreastROC: 0.91Geometric transformations,
Contrast and brightness adjustment,
[31]Mammogram ClassificationAlexnet, ResnetPrivate-Geometric transformations
[89]AD detectionAlexnet, SVMCBIS-DDSM, DDSM, MIASAccuracy: 92Geometric transformations, TTA
[34]Mammogram detection and
classification
YOLOINBreastAccuracy: 89.6Rotation, Flipping
[90]Build datasets of breast
mammography
Alexnet, Densenet,
Shufflenet
INBreast-Rotation, Flipping
[91]Mass DetectionFaster R-CNNOMI-DBTPR:
0.99 ± 0.03 at 1.17 FPI—malignant
0.85 ± 0.08 at 1.0 FPI—benign
Horizontal Flipping
[92]Breast cancer diagnosisPre-trained CNN
Architectures
CBIS- DDSM, BCDR,
INBreast, MIAS
F1 Score for MIAS 0.907 ± 0.150-
[93]Breast cancer classificationDCNNMIASAccuracy: 90.50Feature wise data augmentation
[56]Mass ClassificationCNNDDSM-cycle GAN
[33]Masses DiscriminationGoogleNetDDSMAccuracy: 90.38%Flipping, Cropped-ROI, Gaussian noise
[66]Image ClassificationVGG-16/19Mini MIAS-Crossover technique
[55]Mass Image SynthesisGANDDSM, Private-Contextual Information Based on GANs
[67]Mass DetectionOne-Stage Object Detection
Architecture (BMassDNet)
INBreast
DDSM
Recall: INBreast: 0.93
DDSM:0.943
Elastic Deformation
[51]Mass DetectionFully Convolutional NetworkCBIS-DDSM
Inbreast
0.8040 PAUC
0.8787 [email protected]
Adversarial Learning
[35]Breast Cancer ClassificationDeep CNNMIAS, DDSM,
Inbreast
Accuracy:
MIAS: 96.55%,
DDSM: 90.68%,
INbreast: 91.28%,
Geometric Transformations,
Gaussian noise
[36]Mass DetectionContrastive Learning,
CycleGAN
Inbreast, Private-Geometric Transformations
[68]Mass ClassificationDeep CNNPrivate0.760 ± 0.015 for 80% labeled dataVirtual Adversarial Training
[71]Mass DetectionEight Object Detection ModelsOPTIMAM, Inbreast,
BCDR
Out of eight models, DETR [94]
could perform well
Cutout and RandConv
[75]BI-RADS ClassificationEfficientNet-B2PrivateMacro F1 score: 0.595Transparency Strategy
[95]Mass DetectionPre-trained CNNs,
DenseNet, ResNet, ResNeXt
BCDRAccuracy: 84%Geometric Transformations
[96]Lesion DetectionYOLOv4
Nested Contours Algorithm
INBreastSensitivity: 93% by NCAGeometric Transformations
[97]Mammogram Density ClassificationDenseNet201, ResNet50MIASAccuracy:
DenseNet201: 90.47%
Geometric Transformations
[65]Mass SegmentationU-NetDDSMSensitivity: 92.32%Geometric Transformations
[98]Breast Cancer DetectionPre-trained CNNs
VGG-16, VGG-19, ResNet-50
MIASAccuracy:
ResNet-50: 71%
Geometric Transformations
Table 3. Articles with pre and post augmentation dataset size and model performance.
Table 3. Articles with pre and post augmentation dataset size and model performance.
Ref.Pre-Augmentation Dataset SizePost-Augmentation Dataset SizePost-Augmentation Model Performance
[24]280 (Mammograms)345,000 ROIs-
[8]5136 ROIs (MIAS),
410 whole images (Inbreast),
322 whole images (MIAS),
55,890 ROIs (DDSM, CBIS)
49,724 ROIs (MIAS),
7914 whole images (MIAS),
1688 whole images (Inbreast),
179,447 ROIs (DDSM, CBIS)
-
[25]-8 new labels per image-
[81]3741560 samplesAccuracy improved by more than 33%
[26]3222576-
[82]329026,320-
[28]--Rise in validation accuracy from
51.3% to 83.6%
[47]3229000-
[48]--Increased AUC from 0.78 to 0.86
[83]--Improved FPI 3.509 (CBIS), 1.864 (BCRP)
[84]--Rise in accuracy from 0.6026 to 0.8670
[10]1798Single image to be augmented into 546 imagesRise in accuracy from 69.85% to 94%
[85]5257104,795-
[30]--Rise in accuracy from 78.92% to 80.56%
[88]--Improvement in sensitivity from 0.786 to 0.913
[31]--Improvement in auROC from 0.62 to 0.73
[89]215 ROI3006 ROI-
[34]107428-
[90]1067632-
[93]221 (Patches)1768 Patches-
[56]--Improvement in accuracy by 1.4 %
[33]-Dataset is expanded by 24 times-
[66]--Improvement in accuracy by 1.47%,
[55]--Improvement in detection rate by 5.03%
[35]322 (MIAS), 1500 (DDSM), 410 (Inbreast)3200 (MIAS), 28,800 (DDSM), 2240 (Inbreast)-
[75]25,373 (Training Samples)28,000 (Training Samples)-
[96]1061080-
[65]798948,659 ROI-
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Oza, P.; Sharma, P.; Patel, S.; Adedoyin, F.; Bruno, A. Image Augmentation Techniques for Mammogram Analysis. J. Imaging 2022, 8, 141. https://doi.org/10.3390/jimaging8050141

AMA Style

Oza P, Sharma P, Patel S, Adedoyin F, Bruno A. Image Augmentation Techniques for Mammogram Analysis. Journal of Imaging. 2022; 8(5):141. https://doi.org/10.3390/jimaging8050141

Chicago/Turabian Style

Oza, Parita, Paawan Sharma, Samir Patel, Festus Adedoyin, and Alessandro Bruno. 2022. "Image Augmentation Techniques for Mammogram Analysis" Journal of Imaging 8, no. 5: 141. https://doi.org/10.3390/jimaging8050141

APA Style

Oza, P., Sharma, P., Patel, S., Adedoyin, F., & Bruno, A. (2022). Image Augmentation Techniques for Mammogram Analysis. Journal of Imaging, 8(5), 141. https://doi.org/10.3390/jimaging8050141

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop