Next Article in Journal
Analysis of Light Obstruction from Street Lighting in Road Scenes
Next Article in Special Issue
A Lightweight Arbitrarily Oriented Detector Based on Transformers and Deformable Features for Ship Detection in SAR Images
Previous Article in Journal
Aboveground Forest Biomass Estimation Using Tent Mapping Atom Search Optimized Backpropagation Neural Network with Landsat 8 and Sentinel-1A Data
Previous Article in Special Issue
Faster and Lightweight: An Improved YOLOv5 Object Detector for Remote Sensing Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Semantic-Layout-Guided Image Synthesis for High-Quality Synthetic-Aperature Radar Detection Sample Generation

1
College of Information Science and Technology, Beijing University of Chemical Technology, Beijing 100013, China
2
Aerospace Information Research Institute, Chinese Academy of Sciences, Beijing 100045, China
3
Shandong Key Laboratory of Low-Altitude Airspace Surveillance Network Technology, Heze 274201, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(24), 5654; https://doi.org/10.3390/rs15245654
Submission received: 20 October 2023 / Revised: 29 November 2023 / Accepted: 4 December 2023 / Published: 7 December 2023

Abstract

:
With the widespread application and functional complexity of deep neural networks (DNNs), the demand for training samples is increasing. This elevated requirement also extends to DNN-based SAR object detection. Most public SAR object detection datasets are oriented to marine targets such as ships, while data sets oriented to land targets are relatively rare, though they are an effective way to improve the land object detection capability of deep models through SAR sample generation. In this paper, a synthesis generation collaborative SAR sample augmentation framework is proposed to achieve flexible and diverse high-quality sample augmentation. First, a semantic-layout-guided image synthesis strategy is proposed to generate diverse detection samples. The issues of object location rationality and object layout diversity are also addressed. Meanwhile, a pix2pixGAN network guided by layout maps is utilized to achieve diverse background augmentation. Second, a progressive training strategy of diffusion models is proposed to achieve semantically controllable SAR sample generation to further improve the diversity of scene clutter. Finally, a sample cleaning method considering distribution migration and network filtering is employed to further improve the quality of detection samples. The experimental results show that this semantic synthesis generation method can outperform existing sample augmentation methods, leading to a comprehensive improvement in the accuracy metrics of classical detection networks.

1. Introduction

Synthetic-Aperture Radar (SAR) imaging technology, known for its high resolution, extensive coverage area, and independence from weather conditions, is extensively used in geographic information systems, resource exploration, and civil applications. With the development of deep learning, data-driven deep models are widely used in SAR processing tasks such as object detection and recognition [1,2,3,4].
In the realm of object detection tasks, data-driven deep learning techniques have showcased remarkable advancements in performance, surpassing traditional approaches such as template matching, image processing, and machine learning methods. Through extensive training on large-scale datasets, DNN-based methods have realized heightened precision and robustness in object detection capabilities. High-quality training samples constitute the foundational basis for achieving high-performance outcomes with deep learning approaches. However, due to challenges in acquiring SAR samples and annotating samples, SAR object detection encounters difficulties in matching the sample advantage seen in natural scene object detection. As a solution to this challenge, two technological approaches have emerged, namely few-shot learning and sample augmentation. In contrast to few-shot learning, SAR sample augmentation represents a direct approach by augmenting and generating high-quality samples to fulfill the learning requirements of deep models.
Common techniques for SAR sample augmentation can be categorized into three groups: basic methods, image synthesis, and intelligent generation. The central premise focuses on augmenting object diversity while retaining the intrinsic characteristics of the samples. Basic methods include fundamental sample augmentation operations such as random transformations [5,6,7,8,9] and image erasure [10]. These basic methods enhance the diversity of SAR samples by simulating variations within them. While foundational methods enhance model robustness against diverse deformations in tasks like detection and recognition, their impact on the semantic distribution in SAR images is often marked by low background diversity and high semantic information may be limited. In such scenarios, these methods may hardly influence the object’s semantic distribution and could inadvertently introduce noise into the augmented samples, leading to adverse impacts on model inference [11].
The image synthesis method enhances the semantic information of the sample by merging elements from various synthesized SAR images. Element synthesis techniques expand the SAR dataset by generating augmented samples through whole-image fusion [12,13,14,15] or element composition [16,17,18,19,20]. On the one hand, the semantic complexity of the samples has been increased by introducing auxiliary SAR images for fusion. For instance, Hiroshi Inoue et al. [21] randomly selected images from the dataset and fused them pairwise. The resulting synthetic samples significantly improved the performance of the classification network. Jiang et al. [12] adopted a different approach. They fused SAR scene images under various spatiotemporal conditions and trained a model using these synthetic samples. The model was able to effectively capture the changes in land cover in the images. On the other hand, by separating the objects and backgrounds within SAR images and placing sliced objects into backgrounds, new samples can be synthesized for semantic enhancement. This approach provides advantages in terms of both adaptability and cost efficiency. However, these synthesized images often overlook the diversity and complexity of real images, failing to account for factors like imaging direction and lighting angles. Moreover, abrupt changes in object edge intensities contribute to a lack of realism in synthesized images.
The intelligent generation method enhances SAR samples using deep generative models. This approach [22,23,24,25,26] exploits semantic constraints set by input conditions to guide sample generation, resulting in SAR images with crucial semantic information. Generative models are categorized into two types: likelihood-based generative models [27] and implicit generative models [28]. The former employ maximum likelihood criteria to learn the probability distribution of a sample, such as autoregressive models [29], flow models [30], and variational autoencoders [31]. In contrast, the latter do not explicitly represent probability distributions and include notable examples like generative adversarial networks (GANs) [32] and denoising diffusion implicit models (DDIMs). GANs generate realistic images via adversarial training between the generator and discriminator, while DDIM samples images iteratively from noise distributions. In recent years, generative models have exhibited remarkable performance in the field of natural image augmentation. However, generating high-quality SAR samples is challenging due to the limited quantity of SAR detection samples, their complex distribution characteristics, and the lack of annotated samples.
Despite the success of the above methods in achieving effective data augmentation for recognition tasks, they all have certain limitations that result in suboptimal performance in detection tasks. An inherent issue is that synthetic methods often overlook SAR images’ distinct imaging mechanisms and traits, compromising the diversity and logical consistency of object positioning in synthesized images. Furthermore, generative models inherently possess randomness. In the context of object detection tasks, ensuring the accuracy of object categories in generated images to avoid label confusion adds complexity to sample augmentation. Additionally, generative methods often require rich conditional information for guidance, and the creation of paired training samples demands significant human resources. In summary, SAR detection sample augmentation faces the following challenges:
  • The synthesis of SAR images often overlooks the correlation between key elements, such as ensuring that objects do not overlap and maintaining logical layouts between objects and backgrounds. This lack of semantic information integration leads to a performance drop in real-world scenarios.
  • The generation process of SAR detection samples lacks clear constraints on the range of semantic image transformations. As a result, the quality of generated images is unstable, making it difficult to ensure consistency in object categories. This renders the generated sample less suitable for high-reliability tasks.
To address the mentioned challenges, this study takes a semantic constraint approach from the SAR background and combines basic methods with intelligent generation techniques to optimize traditional image synthesis methods. Prior to sample synthesis, we performed semantic layout of backgrounds to ensure the diversity and logical coherence of the object distribution, and significantly improved synthesis efficiency. However, since image elements are derived from the original SAR sample library, there is no substantial variation in objects and backgrounds in the synthesized images. As a result, the diversity of augmented samples becomes limited in scope. Consequently, we further utilize generated synthetic samples to enhance the original dataset. Firstly, we leverage semantic layout maps to guide the training of a pix2pixGAN model, enabling semantic-controllable SAR scene augmentation. Moreover, we aim to feed the generative model with an increased volume of synthetic samples via progressive training to achieve higher-quality generated objects. At the same time, it is essential to clean the generated sample to acquire a high-quality augmented sample. We devised a sample cleaning strategy that combines distribution transfer and network-based filtering. This approach takes into consideration the statistical distribution characteristics of SAR samples and their corresponding labels, thereby effectively obtaining high-quality SAR samples.
In general, the main contributions of our research are as follows:
(1) A synthesis generation collaborative SAR sample augmentation framework is proposed. The diffusion model and SAR sample synthesis method are combined to achieve flexible and controllable SAR sample augmentation.
(2) A sample synthesis method based on a semantic layout is proposed. Prior to synthesizing objects into backgrounds, we pre-layout each background to avoid the inefficiency caused by repetitive evaluations of the same background during the synthesis process. Meanwhile, we employ a pix2pixGAN network trained on layout maps to achieve semantic-controllable background augmentation, further enhancing the diversity of the background in synthesized images.
(3) In order to obtain high-quality SAR-generated samples, a progressively trained conditional diffusion model is proposed, utilizing synthetic samples. By integrating the sample cleaning strategy of distribution transfer and network filtering, high-quality SAR samples are efficiently acquired.
The rest of this article is organized as follows. Section 2 provides a brief overview of related work on SAR sample augmentation. Section 3 presents a detailed description of our efficient sample augmentation strategy. In Section 4, we conduct extensive sample augmentation experiments and compare the results with other classical methods. Finally, Section 5 presents the conclusions of the experiments.

2. Related Works

2.1. Basic Methods of Image Augmentation

Fundamental image enhancement methods primarily include overall image transformations. These include information transformation and information fusion between images. These methods address image enhancement from various perspectives using direct and practical techniques. As summarized in Table 1, these methods offer diverse approaches to image enhancement.

2.2. Sample Synthesis

Sample synthesis merges objects and backgrounds from varied sources, producing images with distinct distribution semantics [33]. The process of natural image synthesis can be broken down into two crucial steps: feature decomposition and feature recombination. Feature decomposition research is primarily anchored in image segmentation and matting [34,35], targeting enhanced segmentation accuracy. Conversely, feature re-combination research focuses on rectifying semantic inconsistencies during image synthesis, aiming for superior synthesized image quality. These semantic inconsistencies include appearance disparities (image incoherence) [36,37], geometric incongruities (unreasonable positions and shapes) [38,39], and logical incongruities (unreasonable element associations) [33].
Unlike natural imaging, SAR imaging remains impervious to lighting and atmospheric disturbances. Consequently, objects in SAR images typically exhibit a relatively consistent appearance, with minimal variations under different observational conditions. Furthermore, SAR objects typically manifest as reflection intensities. These differences are subtler compared to features like color and texture in visible images. Thus, during sample synthesis, an excessive focus on appearance consistency is not necessary. Given that SAR image backgrounds, often terrains or static objects, exhibit minimal temporal or weather-induced variations, the geometric consistency and logical interaction between backgrounds and objects are primarily manifested through their spatial positions. Therefore, SAR sample synthesis prioritizes bolstering the spatial relationship between objects and backgrounds for object detection, ensuring greater sample relevance and stability.

2.3. Sample Generation

The generation of detection samples refers to the production of diverse images under given conditional constraints [40]. The majority of SAR image generation is based on GAN networks. For sample augmentation, CGAN [41] integrates added conditional constraints into GANs, facilitating the generation of category-specific images. However, this approach suffers from issues of limited diversity in generated images and recurring problems of pattern collapse [42,43]. Following this, seminal models like Pix2Pix [44] and CycleGAN [45] were developed. Pix2Pix, as the earliest image-to-image transformation model, is founded on the principle of learning mapping relationships between paired input and output images. Conversely, CycleGAN tackles the issue of obtaining paired training samples. In recent years, diffusion models have gained great success in the field of natural image Artificial Intelligence-Generated Content (AIGC) due to their powerful creativity. Diffusion models consist of two processes: forward denoising and reverse inference. Jascha et al. introduced the reverse diffusion process in 2015, laying the groundwork for diffusion models [46]. In 2020, the debut of the denoising diffusion probability model, which integrates the denoising fraction matching function, brought diffusion models into prominence [47]. In the same year, Song et al. proposed the denoising diffusion implicit model [48], greatly improving the speed of reverse denoising. The distinct imaging techniques and inherent traits of SAR samples have led to the infrequent application of diffusion models in SAR sample augmentation. However, Yuan et al. [49] have recently introduced an efficient, controllable training strategy tailored for conditional diffusion models.

3. Proposed Method

Figure 1 presents the schematic diagram of the proposed method. The augmented dataset comprises three subsets: I S 1 , I S 2 , and I G . I S 1 denotes synthetic samples generated using a semantic layout strategy. I S 2 indicates synthetic samples augmented for background diversity, and I G corresponds to samples generated via the diffusion model.

3.1. Semantic Layout Strategy

Our collection consists of n images, denoted as I O 1 , I O 1 ,…, I O n , sampled from the object library O . Additionally, we select a background image I B , from the background library B . The process of image synthesis is represented as shown in Equation (1).
I S = i = 1 n λ i I O i + I B
where I S represents the synthesized image, and λ i is the positional parameter when synthesizing the objects I O i .
Due to the unique imaging mechanism of SAR images, choosing appropriate positional parameters λ i can significantly enhance the visual realism of synthesized images. The arbitrary selection of positions using land classification masks can result in over-regularization of λ i , inefficiencies from redundant background determination, and potential time wastage due to inadequate positioning. To address this, we initially rasterize the mask to obtain refined positional parameters λ i . Then, guided by our semantic layout strategy, we allocate selection probabilities p i to each λ i , adjusting the intensity for object placements across various positions as g i . Here, g i is the ratio of the mean of pixel value within a rectangular region (Mean(region)) to the mean pixel value of the entire background (Mean(global)). Consequently, our enhanced image synthesis approach is described as follows:
I S 1 = i = 1 n p i λ i I O i g i + I B
g i = M e a n ( r e g i o n ) M e a n ( g l o b a l )

3.1.1. Mask Rasterization

Layout grids are created via sliding window operations on predefined boxes. The effectiveness of these layout boxes is determined by assessing the area proportion of valid regions and their degree of dispersion, ultimately producing the scene’s layout map.
In object detection tasks, sample augmentation often focuses on small objects or those with limited representation in specific classes. As a result, variations in object size are usually constrained. The dimensions of the layout boxes are set to match the largest object instance. The mask of the background image B , denoted as M b , is also considered. Additionally, preset position boxes are defined as a i . A column-major sliding window traversal is executed on the background image, as shown in Figure 2. M i is defined as the mapping region of a i on M b , and the effective occupancy ratio F i of this region is calculated as shown in Equation (4):
F i = a r e a ( m i ) / a r e a ( M i )
where F i quantifies the proportion of the effective region m i within the position box M i .
When F i exceeds the threshold ε, indicating the object placement condition is met, we proceed to calculate the dispersion D of the effective region within M i . As described by Equation (5), we count the pixels in the effective region, denoted as n, and compute the centroid (X0, Y0) of this region. We calculate the mean squared Euclidean distances l i between these pixels and the centroid, determining the dispersion D of M i . An appropriate dispersion threshold allows us to generate a refined set of layout position boxes.
D = i = 1 n l i 2 n

3.1.2. Semantic Constraints

Employed mask rasterization not only improves synthesis efficiency but also enhances the quality of the resultant images by imposing constraints on key semantic dimensions. The semantic layout strategy incorporates constraints in three key dimensions: object positioning, distribution density, and scattering intensity. Specifically, it includes the following strategies.
To boost the robustness of the synthesized sample, edge mirroring augmentation is applied before initiating scene layout. Objects positioned at the edges are then cropped back, resulting in what we term ‘incomplete objects’. To mitigate the adverse impact of too many ‘incomplete objects’, placement boxes are classified into two categories during scene layout: edge boxes (marked in red) and interior boxes (marked in green), as shown in Figure 2. The likelihood of successful object placement within edge regions is intentionally reduced. Furthermore, to prevent the clustering of small objects, the placement probability p i for each layout box is modulated. Specifically, the probability of placing an object successfully in edge regions is to be 0.5 × p i .
p i = e d i t
where, d i represents the distance from the center of each rectangle to the discrete center P, which is the center of the rectangle in the layout closest to the region center (X0, Y0). The hyperparameter t is used to adjust p i .
A greater distance between the ith rectangle and point P increases the likelihood of successful object placement, thus promoting object dispersion. Lastly, the scattering intensity for SAR objects is calibrated using the ratio p l i . This ratio is derived from the mean pixel value R m e a n of the background for that region relative to the global background mean B m e a n . The complete algorithm is shown in Algorithm 1.
Algorithm 1: The Semantic-Layout Strategy.
Input: background B and mask M , object slice I
Output: layout map, position box information
Process:
1: The size of M is (w0, h0), perform edge mirroring augmentation to obtain M1(w, h). Preset the size of position boxes as (m, n) and perform grid-based layout on M1:
2: for j in 1, 2, …, w do
3:   for k in 1, 2, …, h do
4:      Calculate the effective occupancy ratio F i .
5:       if F i > ε do
6:        Calculate the center point (X0, Y0) of the effective region.
7:        Calculate the dispersion D according to Equation (5).
8:       end if
9:     Record the rectangular position information λ i , including center (x, y), width–height (h, w), and category (whether it is located on the edge)
10:  end for
11: end for
12: Count the number of position boxes (num), calculate the probability p i based on Equation (2).
13: start composing:
14:  for each B in the background dataset do
15:    Perform edge mirror augmentation on B to match the same augmentation as M .
16:    Randomly select I O i from the object library, where i ∈ [1,num]
17:    Place the objects in the position boxes at random positions based on p i .
18:    Synthesize samples according to Equation (2).
19:    Restore B by cropping the edges.
20:  end for

3.2. Diverse Scene Generation

The Pix2pixGAN [44] model, built upon the foundation of Conditional Generative Adversarial Networks (CGANs) [41], is used for image-to-image translation. Its network structure is shown in Figure 3. Its core objective is to perform supervised translations between different image domains. Training Pix2pixGAN networks can be challenging, particularly with a scarcity of training samples. To overcome this, a series of strategies are employed to simplify the complexity of cross-domain transformation tasks. One key strategy involves replacing the input mask with a layout map, streamlining the task’s complexity. This approach helps alleviate the training difficulty, especially when dealing with limited samples. Additionally, the grid-based layout map provides extra spatial structure, aiding the network in learning the texture information within object regions.
To better control background generation, we introduced region-wise gradient loss ( L l a p l a c i a n ) and region position loss ( L p o s i t i o n ) into the generator’s training, as shown in Equations (7) and (8).
L p o s i t i o n = 1 N i = 1 N | G ( A ) i B i |
L l a p l a c i a n = 1 N i = 1 N ( L ( G ( A ) i ) L ( B i ) ) 2
where N represents the total number of elements, G ( A ) i and B i denote the ith elements of the generated and the real images, and L (   . ) is the Laplacian operator, used for computing Laplacian gradients of images.
Equation (8) calculates the Mean Square Error (MSE) between the Laplacian gradients of the generated and real images. Integrating these dual loss functions has proven effective in optimizing the generator’s performance. These loss functions more accurately capture local structural nuances in the generated images, enhancing background generation quality. The refined background set is then used for sample synthesis, as shown in Equations (9) and (10).
I B A = L G ( I l m )
I S 2 = i = 1 n p i λ i I O i g i + I B A
where L G ( i n p u t ) represents the pix2pixGAN model output, I B A is the augmented background, and I S 2 the synthesized sample with background augmentation.

3.3. Progressive Training of Synthetic Data-Driven Diffusion Models

3.3.1. Progressive Training of a Diffusion Model

In contrast to natural scene images, SAR scene clutter features a more intricate and diverse array of characteristics, encompassing various forms of interference, spatial correlations, and non-Gaussian statistical distributions. The Denoising Diffusion Probabilistic Model (DDPM) is a state-of-the-art generative model effective in capturing the complex clutter features of SAR scenes. It preserves spatial correlations and offers flexibility and controllability.
Noise addition in the diffusion model follows a Markovian procedure. In this model, α t is the weight parameter for noise addition, and Z represents Gaussian noise sampled from a standard normal distribution. In each successive step of the forward noise addition process, denoted as X t , the computation involves noise addition weight α t and Gaussian noise Z t .
X t = α t X t 1 + 1 α t Z = δ t X 0 + 1 δ t Z ¯ t
where δ t is the product of the noise addition weights from previous steps ( α t · α t 1 · · · α 2 · α 1 ). In each forward diffusion step, standard Gaussian noise is added until the image is full noisy.
Conversely, the denoising process functions as a contrasting operation to the noise addition process, aiming to systematically eliminate noise. This process iteratively removes noise to generate a new image. The noise addition process derives X t from X t 1 , while the denoising process estimates X ~ t 1 from X ~ t . Moreover, the conditional diffusion model includes a conditional input c for constrained denoising. In our experiment, the multi-class mask and distribution map serve as the condition c , the model f θ ( X ~ t ,   c ,   δ t ) estimates the added noise δ t in the forward diffusion. Therefore, the conditional diffusion model can be expressed as Equation (12).
X ~ t 1 = 1 α t X ~ t 1 α t 1 δ t   f θ X ~ t ,   c ,   δ t + 1 α Z ¯ t
Although diffusion models have shown excellent performance in image generation tasks, they require significant data support for training. Drawing inspiration from semi-supervised deep matting networks, we synthesis large datasets using masks instead of trimaps for model training, achieving high-quality matting masks [13]. Guided by this approach, we first use a synthesis method to obtain the constraint condition c, and then apply a diffusion model for image generation.
Our methodology begins with using synthesis techniques to create multi-class masks and distribution maps, as shown in Figure 4. In the multi-class mask, different pixel values represent various object categories. And the distribution map includes real object information, not just masked regions. We start with the multi-class mask for pre-training, focusing on constraining position and resolution information. Subsequently, the distribution map is used for transfer learning to further constrain the texture and category information. Through a progressive conditional learning approach, our model becomes capable of generating SAR samples with predetermined positions and categories.

3.3.2. Cleaning Generated SAR Samples

Outputs from the generative model can have issues like excessive background noise, object distortions, and semantic inconsistencies, reducing the utility of generated samples for augmentation tasks. Therefore, cleaning these generated SAR samples is essential to ensuring their quality and usability. SAR images often deviate from the conventional Gaussian distribution in pixel values with their intricate terrains and complex scattering mechanisms. Instead, their pixel values typically follow the K-distribution (Kappa distribution). The K-distribution effectively models these complex scattering mechanisms and noise characteristics, as illustrated in Equation (13).
f x ; k , θ = 2 k | x | 2 k 1 2 Γ k θ k   e x p ( | x | k θ )
The K-distribution is defined by two parameters: k , representing degrees of freedom, and θ , the scale parameter. Based on the labels of the synthetic sample, the slices of objects from the generated images are extracted. As shown in Figure 5, we fit these object slices to the K-distribution and adjust their pixel values based on discrepancies in the k parameters. Next, we use a yolov5 model trained on the original samples to identify the generated samples. The detection results are cross-verified with actual labels to facilitate sample cleaning.

4. Experimental Results and Analysis

This section introduces the SAR dataset used and compares our method with other SAR sample augmentation methods.

4.1. Experimental Data and Evaluation Indexes

(1) Experimental Data. To establish a robust baseline for sample augmentation, we used the MSTAR dataset to generate synthetic samples and created more representative and diverse samples for training and testing. We focused on ensuring the coherence of background and object resolution and the semantic authenticity of small object distribution. We selected 10 object classes, including 2S1, bmp2, BRDM_2, among others, and employed various image enhancement strategies. This process results in a dataset comprising 1000 high-resolution SAR images, serving as the baseline for our experiments. Moreover, we prepared a test set of 311 SAR images with corresponding masks, matching the specifications of the baseline dataset. This test set covered 10 different object types and 8 land cover types, such as barren land, grassland, farmland, shrubbery, forest, roads, buildings, and ocean.
(2) Evaluation Metrics. This experiment employed two primary sets of evaluation metrics to assess the performance of the generated images and the effectiveness of the augmented datasets for the given task. Initially, well-established quality evaluation metrics were used, such as the Inception Score (IS) and Fréchet Inception Distance (FID). They were utilized to assess image quality and quantify the quality of images generated by the models. Secondly, the precision (P), recall (R), and mean average precision (mAP) metrics were employed to evaluate the performance of the augmented datasets generated by different methods for a specific task. By comprehensively using these metrics, the experiment provides a more comprehensive assessment and comparison of various augmentation samples, ensuring more reasonable and accurate evaluation results.
  • IS↑: Inception Score. IS consists of two components: the first value is the mean of the KL divergence, where a larger value indicates higher data quality. The second value is the standard deviation of the KL divergence, indicating richer sample diversity as the value increases.
  • FID↓: Fréchet Inception Distance. FID is a metric used to measure the difference between generated images and real images. The lower the FID indicates the smaller difference between generated images and real images.
  • P↑: Precision. The precision rate refers to the ratio of the truly positive samples among the positive samples detected by the model. The higher the precision rate indicates the lower the probability of the model predicting negative samples as positive samples.
  • R↑: Recall. The recall rate refers to the proportion of truly positive samples detected by the model among all positive samples. A higher recall rate indicates a stronger ability of the model to correctly detect positive samples.
  • mAP↑: mean Average Precision. The overall performance evaluation index of the model. A higher mAP value indicates superior model performance in object detection tasks.
  • F1-score↑: F1-score is a performance metric that provides a comprehensive evaluation of the accuracy of a classification model by considering both precision and recall.
(3) Experimental environment. All experiments were executed on a uniform hardware configuration, with detailed environmental settings enumerated in Table 2.

4.2. Comparison with Other Advanced Methods

To evaluate the quality of the augmented samples, we performed qualitative and quantitative comparisons with samples produced by leading generative networks. The models for comparison are as follows:
  • Pix2pixGAN [44]: An improved generative adversarial network that achieves real-time transformation from input images to object images by learning the mapping relationship between them.
  • CycleGAN [45]: An unsupervised generative adversarial network that maintains the original image’s content using cycle consistency loss.
  • BicycleGAN: By introducing additional consistency loss and mutual information loss, it enhances controllability and diversity in image generation tasks.
  • DDPM [47]: A generative model that transforms random noise into a coherent image by simulating the reverse of diffusion, providing a unique way to create high-quality and diverse samples.
To evaluate the quality and label alignment of the generated sample, we conducted both qualitative and quantitative comparisons. The yolov5 model was employed to evaluate label alignment in the generated samples, with results detailed in Figure 6 and Table 3. The original training dataset consisted of 1000 images, and the test set contained 311 images. We generated 1000 images separately using three different GAN networks. The generated samples originated from three datasets: one synthesized with a semantic layout strategy (Ours-SL, 500 images), one blending synthetic backgrounds with real objects (Ours-BA, 500 images), and one created through progressive training of a diffusion model (Ours-GE, 1000 images). These datasets collectively form our augmented dataset (Ours-all).
Our qualitative and quantitative analyses revealed key insights: CycleGAN-generated images exhibit noticeable artifacts, particularly around object areas and suffer from limited diversity and low realism. In contrast, pix2pixGAN showed superior performance in image-to-image translation compared to CycleGAN. However, it tends to produce images with more background noise, lower resolution, and limited background diversity, evidenced by a KL (std) of only 0.038. While BicycleGAN excels in generating realistic backgrounds, it faces significant object distortions, resulting in a high FID score of 32.07. It also encounters issues of mode collapse in background generation. Compared to other GAN-based methods, DDPM generates higher-quality images, achieving a KL (mean) of 1.871 and a significantly lower FID score of 25.93. Nonetheless, it faces challenges in object generation distortions and does not meet the label matching requirements for data augmentation. Our generative model (Ours-GE) outperformed the benchmarks, achieving an FID score of 25.54, and produced images with higher resolution and fewer object distortions. However, it does not match pix2pixGAN in terms of background complexity.
To further validate the effectiveness of our augmented data, we trained a yolov5 network on the original dataset to assess the alignment between generated sample and object labels. In our study, we utilized the YOLOv5 model to perform detection on generated samples. Subsequently, we compared the detection results with ground truth labels, as depicted in Table 3. The term “Matching degree” is employed to indicate the degree of correspondence between the test results and the ground truth labels, represented as a percentage. As shown in the last column of Table 3, GAN networks perform poorly in generating specified-category small objects, with matching rates ranging from 10% to 20%. DDPM demonstrates improved performance in generating specified-class targets, though its matching percentage remains below 50%. In comparison, datasets synthesized using real objects and real backgrounds have nearly identical label matching rates to the original dataset. However, with the introduction of generated backgrounds in Ours-BA, there is a slight decrease in the alignment between synthesized samples and labels. Notably, samples generated by Ours-GE achieve a label matching rate of 78.2%, indicating that they retain crucial feature information from real samples. This characteristic renders them highly suitable for sample augmentation purposes. Finally, our augmented dataset, which amalgamates three distinct enhancement techniques, exhibited improved sample diversity compared to the original dataset when evaluated using the KL divergence metric.

4.3. Ablation Analysis of Different Sample Augmentation Strategies

Throughout the experiments, we maintained consistent training parameters across all network models and utilized a uniform validation dataset. We evaluated the effectiveness of different augmentation schemes proposed in classic detection models Yolov5, RetinaNet, and VarifocalNet. This involved contrasting the detection results of different augmentation schemes on these classic networks.
(1) Semantic Layout Strategy. Our dataset, synthesized via a semantic layout strategy, demonstrated superior performance in multiple metrics when compared to a dataset generated through random placement across various parameters of RetinaNet and VarifocalNet, particularly achieving a notable 7.3% improvement in accuracy P for VarifocalNet. This is attributed to the fact that the semantic layout strategy rigorously constrains the distribution of objects on the background during instance placement, enhancing the association between objects and backgrounds. Consequently, the synthesized images demonstrate better semantic consistency between objects and backgrounds, enabling the model to distinguish between objects and backgrounds more effectively. However, for the yolov5 model, while we observed gains in Precision (P) and mean Average Precision (mAP), there was a decline in Recall. We speculate that this could be related to the dense prediction nature of the yolov5 network, where sparse object distribution may have adversely affected detection performance.
(2) Diverse Background Generation. Our comparative experiments (SL vs. SL + BA, as detailed in Table 4) indicated that background enhancement led to significant improvements across multiple detection metrics. As shown in Figure 7, the diverse background images generated from layout maps relax strict spatial constraints while ensuring accurate generation of land cover types within the position boxes. This flexibility allows for the generation of various terrains outside the position boxes. Consequently, the introduction of two additional loss terms, L l a p l a c i a n and L p o s i t i o n , enhances the background details beyond the position boxes. This enhancement, specifically in texture and lighting, is depicted in Figure 7. After augmenting the background diversity, the detection network’s feature map outputs during training are illustrated in Figure 8. By synthesizing diverse backgrounds with real objects, the model has been exposed to samples from various environmental contexts. This exposure has contributed to the enhancement of its robustness and its capacity to successfully differentiate objects from backgrounds in various environmental settings.
(3) Progressive Training Strategy. Utilizing solely the multi-class mask as a constraining condition, the diffusion-model-generated images demonstrated commendable results, especially in object placement and overall scene composition in terms of object placement and scene generation. However, these images fell short in constraining object scattering intensity and preserving texture details, resulting in severe deformation and distortion issues in the generated samples, as shown in Figure 9a. Conversely, leveraging the distribution map as a guiding condition during transfer learning resulted in generated samples that more closely mirrored the appearance of genuine objects, as illustrated in Figure 9b, and possess power spectra closer to real SAR samples (Figure 10).
We simultaneously observed a significant difference in frequency information between our generated samples and their real counterparts. Consequently, we applied K-distribution fitting to both real objects and generated objects separately. We then adjusted the pixel values of the generated objects in accordance with the K-distribution of real objects, aiming to retain more frequency information from the original data. Lastly, we utilized the yolov5 model for sample cleaning, inputting generated samples, comparing object detection results with real labels, and discarding samples showing significant discrepancies.
The study results indicate that generated samples, refined through progressive training and sample cleaning, surpass those from mere background augmentation in augmentation tasks. Our approach, Ours-SL + BA, resulted in an improvement of model accuracy by 0.5% to 1.8%, recall by 0.4% to 2.1%, mAP by 0.6% to 1.9%, and F1-score by 0.45% to 1.59%. Additionally, incorporating GE samples further improved model accuracy, with increases ranging from 1.3% to 3.4%, recall from 1.2% to 3.8%, mAP from 1.2% to 3.9%, and F1-score from 1.22% to 3.71%. These findings suggest that the generated images exhibit significant semantic diversity, highlighting the overall improvement achieved through data augmentation in the context of the detection task.

4.4. Discussion

The proposed synthesis-generation collaborative method presents a novel framework for SAR sample augmentation. The method synthesizes samples effectively, ensuring visual consistency and logical coherence through semantic layout strategies. The integration of synthetic samples and generative models further enhances the semantic diversity of objects and backgrounds within the samples. However, the method faces challenges, particularly in improving its processing efficiency and versatility. On one hand, due to the method being primarily designed for MSTAR objects, it is less effective in generating large objects with complex structures and rich textures. On the other hand, the introduction of diffusion models brings relatively high computational costs. Therefore, developing lightweight approaches for this framework is crucial and will be our future research focus. Nonetheless, the sample augmentation method based on the mutual promotion of synthesis and generation still provides a new and meaningful research perspective for the remote sensing image processing field.

5. Conclusions

We propose a synthesis-generation collaborative SAR sample augmentation framework. Firstly, we introduce a semantic-layout strategy that enhances the correlation between objects and backgrounds through semantic constraints on layout boxes, resulting in semantically enhanced synthetic sample. To increase background diversity, we train a pix2pixGAN network to generate semantically controlled backgrounds. Finally, we progressively train a conditional diffusion model, coupled with distribution transfer and sample cleaning strategies, to obtain high-quality SAR samples. Comprehensive experiments indicate that our image synthesis intelligent generation framework outperforms existing sample augmentation approaches.

Author Contributions

Conceptualization, F.Z.; Methodology, Y.K.; Software, Y.K.; Validation, F.L. and Y.L.; Formal analysis, F.M.; Resources, F.M.; Data curation, F.L. and Y.L.; Writing—review & editing, F.M.; Supervision, F.Z.; Funding acquisition, F.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Natural Science Foundation of China, grant number 62271034.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ma, F.; Sun, X.; Zhang, F.; Zhou, Y.; Li, H.C. What Catch Your Attention in SAR Images: Saliency Detection Based on Soft-Superpixel Lacunarity Cue. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5200817. [Google Scholar] [CrossRef]
  2. Zhou, Y.; Liu, H.; Ma, F.; Pan, Z.; Zhang, F. A Sidelobe-Aware Small Ship Detection Network for Synthetic Aperture Radar Imagery. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5205516. [Google Scholar] [CrossRef]
  3. Zhang, F.; Yao, X.; Tang, H.; Yin, Q.; Hu, Y.; Lei, B. Multiple Mode SAR Raw Data Simulation and Parallel Acceleration for Gaofen-3 Mission. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 2115–2126. [Google Scholar] [CrossRef]
  4. Lyu, Y.; Wang, Z.; Wang, P.; Li, S.; Tan, H.; Chen, K.; Zhao, L.; Sun, X. Scattering information and meta-learning based SAR images interpretation for aircraft target recognition. J. Radars 2022, 11, 652–665. [Google Scholar]
  5. Wangiyana, S.; Samczyński, P.; Gromek, A. Data augmentation for Building Footprint Segmentation in SAR Images: An Empirical Study. Remote Sens. 2022, 14, 2012. [Google Scholar] [CrossRef]
  6. Ma, Q. Improving SAR Object Recognition Performance Using Multiple Preprocessing Techniques. Comput. Intell. Neurosci. 2021, 2021, 6572362. [Google Scholar] [CrossRef] [PubMed]
  7. Yang, R. Rethinking the Random Cropping Data augmentation Method Used in the Training of CNN-Based SAR Image Ship Detector. Sensors 2021, 13, 34. [Google Scholar] [CrossRef]
  8. Yun, S.; Han, D.; Oh, S.J.; Chun, S.; Choe, J.; Yoo, Y. Cutmix: Regularization strategy to train strong classifiers with localizable features. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 6023–6032. [Google Scholar]
  9. Jiang, T.; Cui, Z.; Zhou, Z.; Cao, Z. Data augmentation with Gabor Filter in Deep Convolutional Neural Networks for Sar Object Recognition. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium, Valencia, Spain, 22–27 July 2018; pp. 689–692. [Google Scholar] [CrossRef]
  10. Tang, G.; Zhao, H.; Claramunt, C.; Men, S. FLNet: A Near-shore Ship Detection Method Based on Image Enhancement Technology. Remote Sens. 2022, 14, 4857. [Google Scholar] [CrossRef]
  11. Gong, C.; Wang, D.; Li, M.; Chandra, V.; Liu, Q. Keep augment: A simple information-preserving data augmentation approach. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 19–21 June 2021; pp. 1055–1064. [Google Scholar]
  12. Jiang, X.; Huang, B. Unmixing-Based Spatiotemporal Image Fusion Accounting for Complex Land Cover Changes. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5623010. [Google Scholar] [CrossRef]
  13. Sengupta, S.; Jayaram, V.; Curless, B.; Seitz, S.M.; Kemelmacher-Shlizerman, I. Background Matting: The World Is Your Green Screen. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 2291–2300. [Google Scholar]
  14. Wang, J.; Li, J.; Sun, B.; Zuo, Z. SAR image synthesis based on conditional generative adversarial networks. J. Eng. 2019, 2019, 8093–8097. [Google Scholar] [CrossRef]
  15. Wang, R.; Zhang, H.; Han, B.; Zhang, Y.; Guo, J.; Hong, W.; Sun, W.; Hu, W. Multiangle SAR dataset construction of aircraft targets based on angle interpolation simulation. J. Radars 2022, 11, 637–651. [Google Scholar]
  16. Geng, Z.; Xu, Y.; Wang, B.-N.; Yu, X.; Zhu, D.Y.; Zhang, G. Object Recognition in SAR Images by Deep Learning with Training Data augmentation. Sensors 2023, 23, 941. [Google Scholar] [CrossRef]
  17. Bao, W.; Hu, J.; Huang, M.; Xu, Y.; Ji, N.; Xiang, X. Detecting Fine-Grained Airplanes in SAR Images with Sparse Attention-Guided Pyramid and Class-Balanced Data augmentation. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 8586–8599. [Google Scholar] [CrossRef]
  18. Suo, Z.; Zhao, Y.; Chen, S.; Hu, Y. BoxPaste: An Effective Data augmentation Method for SAR Ship Detection. Remote Sens. 2022, 14, 5761. [Google Scholar] [CrossRef]
  19. Lv, J.T.; Liu, Y. Data augmentation Based on Attributed Scattering Centers to Train Robust CNN for SAR ATR. IEEE Access 2019, 7, 25459–25473. [Google Scholar] [CrossRef]
  20. Ding, B.; Wen, G.; Huang, X.; Huang, X.; Ma, C.; Yang, X. Data augmentation by Multilevel Reconstruction Using Attributed Scattering Center for SAR Object Recognition. IEEE Geosci. Remote Sens. Lett. 2017, 14, 979–983. [Google Scholar] [CrossRef]
  21. Inoue, H. Data augmentation by pairing samples for images classification. arXiv 2018, arXiv:1801.02929. [Google Scholar]
  22. Sun, Y.; Wang, Y.; Hu, L.; Huang, Y.; Liu, H.; Wang, S.; Zhang, C. Attribute-Guided Generative Adversarial Network with Improved Episode Training Strategy for Few-Shot SAR Image Generation. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2023, 16, 1785–1801. [Google Scholar] [CrossRef]
  23. Wang, C.; Pei, J.; Liu, X.; Huang, Y.; Mao, D.; Zhang, Y.; Yang, J. SAR Object Image Generation Method Using Azimuth-Controllable Generative Adversarial Network. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 9381–9397. [Google Scholar] [CrossRef]
  24. Baier, G.; Deschemps, A.; Schmitt, M.; Yokoya, N. Synthesizing Optical and SAR Imagery from Land Cover Maps and Auxiliary Raster Data. IEEE Trans. Geosci. Remote Sens. 2022, 60, 4701312. [Google Scholar] [CrossRef]
  25. Lou, X.; Liu, Y.; Xiong, Z.; Wang, H. Generative knowledge transfer for ship detection in SAR images. Comput. Electr. Eng. 2022, 101, 108041. [Google Scholar] [CrossRef]
  26. Huang, H.; Zhang, F.; Zhou, Y.; Yin, Q.; Hu, W. High Resolution SAR Image Synthesis with Hierarchical Generative Adversarial Networks. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium, Yokohama, Japan, 28 July–2 August 2019; pp. 2782–2785. [Google Scholar] [CrossRef]
  27. Serrà, J.; Álvarez, D.; Gómez, V.; Slizovskaia, O.; Núñez, J.F.; Luque, J. Input complexity and out-of-distribution detection with likelihood-based generative models. arXiv 2019, arXiv:1909.11480. [Google Scholar]
  28. Mohamed, S.; Lakshminarayanan, B. Learning in implicit generative models. arXiv 2016, arXiv:1610.03483. [Google Scholar]
  29. Chung, Y.A.; Hsu, W.N.; Tang, H.; Glass, J. An unsupervised autoregressive model for speech representation learning. arXiv 2019, arXiv:1904.03240. [Google Scholar]
  30. Boutsidis, C.; Woodruff, D.P.; Zhong, P. Optimal principal component analysis in distributed and streaming models. In Proceedings of the 48th Annual ACM Symposium on Theory of Computing, Cambridge, MA, USA, 19–21 June 2016; pp. 236–249. [Google Scholar]
  31. Kingma, D.P.; Welling, M. Auto-encoding variational bayes. arXiv 2013, arXiv:1312.6114. [Google Scholar]
  32. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial networks. arXiv 2014, arXiv:1406.2661. [Google Scholar] [CrossRef]
  33. Niu, L.; Cong, W.; Liu, L.; Hong, Y.; Zhang, B.; Liang, J.; Zhang, L. Making Images Real Again: A Comprehensive Survey on Deep Image Composition. arXiv 2021, arXiv:2106.14490. [Google Scholar]
  34. Fu, Y.; Hu, Y. A survey on image matting methods based on deep learning. Front. Econ. Manag. 2022, 3, 497–502. [Google Scholar]
  35. Xu, N.; Price, B.; Cohen, S.; Huang, T. Deep image matting. In Proceedings of the EEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 2970–2979. [Google Scholar]
  36. Cong, W.; Zhang, J.; Niu, L.; Liu, L.; Ling, Z.; Li, W.; Zhang, L. DoveNet: Deep image harmonization via domain verification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 8394–8403. [Google Scholar]
  37. Cun, X.; Pun, C.-M. Improving the harmony of the composite image by spatial-separated attention module. IEEE Trans. Image Process. 2020, 29, 4759–4771. [Google Scholar] [CrossRef]
  38. Azadi, S.; Pathak, D.; Ebrahimi, S.; Darrell, T. Compositional GAN: Learning image conditional binary composition. Int. J. Comput. Vis. 2020, 128, 2570–2585. [Google Scholar] [CrossRef]
  39. Zhang, L.; Wen, T.; Min, J.; Wang, J.; Han, D.; Shi, J. Learning object placement by inpainting for compositional data augmentation. In Proceedings of the Computer Vision—ECCV 2020, Glasgow, UK, 23–28 August 2020; pp. 566–581. [Google Scholar]
  40. Katsumata, K.; Nakayama, H. Semantic image synthesis from inaccurate and coarse masks. In Proceedings of the IEEE International Conference on Acoustics Speech and Signal Processing (ICASSP), Toronto, ON, Canada, 6–11 June 2021; pp. 2285–2289. [Google Scholar]
  41. Mirza, M.; Osindero, S. Conditional generative adversarial nets. arXiv 2014, arXiv:1411.1784. [Google Scholar]
  42. Audebert, N.; Le Saux, B.; Lefevre, S. Generative Adversarial Networks for Realistic Synthesis of Hyperspectral Samples. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium, Valencia, Spain, 22–27 July 2018; pp. 4359–4362. [Google Scholar] [CrossRef]
  43. Wang, G.; Dong, G.; Li, H.; Han, L.; Tao, X.; Ren, P. Remote Sensing Image Synthesis via Graphical Generative Adversarial Networks. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium, Yokohama, Japan, 28 July–2 August 2019; pp. 10027–10030. [Google Scholar] [CrossRef]
  44. Isola, P.; Zhu, J.Y.; Zhou, T. Image-to-Image Translation with Conditional Adversarial Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 5967–5976. [Google Scholar]
  45. Zhu, J.-Y.; Park, T.; Isola, P.; Efros, A.A. Unpaired Image-to-Image Translation Using Cycle-Consistent Adversarial Networks. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2242–2251. [Google Scholar]
  46. Sohl-Dickstein, J.; Weiss, E.; Maheswaranathan, N.; Ganguli, S. Deep unsupervised learning using nonequilibrium thermodynamics. In Proceedings of the International Conference on Machine Learning 2015, Lille, France, 6–11 July 2015; Volume 19, pp. 2256–2265. [Google Scholar]
  47. Ho, J.; Jain, A.; Abbeel, P. Denoising diffusion probabilistic models. Adv. Neural Inf. Process. Syst. 2020, 33, 6840–6851. [Google Scholar]
  48. Song, J.; Meng, C.; Ermon, S. Denoising diffusion implicit models. arXiv 2020, arXiv:2010.02502. [Google Scholar]
  49. Yuan, Z.; Hao, C.; Zhou, R.; Chen, J.; Yu, M.; Zhang, W.; Wang, H.; Sun, X. Efficient and Controllable Remote Sensing Fake Sample Generation Based on Diffusion Model. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5615012. [Google Scholar] [CrossRef]
Figure 1. Proposed synthesis generation collaborative SAR sample augmentation framework. (a) The designed semantic layout strategy. (b) Diverse scene generation based on layout map. (c) Progressive training strategy for preserving more object details.
Figure 1. Proposed synthesis generation collaborative SAR sample augmentation framework. (a) The designed semantic layout strategy. (b) Diverse scene generation based on layout map. (c) Progressive training strategy for preserving more object details.
Remotesensing 15 05654 g001
Figure 2. The process of mask rasterization.
Figure 2. The process of mask rasterization.
Remotesensing 15 05654 g002
Figure 3. The architecture diagram of pix2pixGAN. G stands for the generator, and D is the discriminator.
Figure 3. The architecture diagram of pix2pixGAN. G stands for the generator, and D is the discriminator.
Remotesensing 15 05654 g003
Figure 4. Input images for the progressive training strategy of the diffusion model. (a) Synthetic image, (b) multi-class mask, different colors indicate different categories, (c) distribution map.
Figure 4. Input images for the progressive training strategy of the diffusion model. (a) Synthetic image, (b) multi-class mask, different colors indicate different categories, (c) distribution map.
Remotesensing 15 05654 g004
Figure 5. Transfer of distribution characteristics. (a) Original slice; (b) the generate slice; (c) distribution transfer.
Figure 5. Transfer of distribution characteristics. (a) Original slice; (b) the generate slice; (c) distribution transfer.
Remotesensing 15 05654 g005
Figure 6. Qualitative comparison between Ours-GE and other generative models. There are four sets of example images, including five classes of MSTAR objects: 2S1, BMP2, BRDM_2, BTR_60, and BTR_70. The red boxes highlight severely distorted portions in the generated objects.
Figure 6. Qualitative comparison between Ours-GE and other generative models. There are four sets of example images, including five classes of MSTAR objects: 2S1, BMP2, BRDM_2, BTR_60, and BTR_70. The red boxes highlight severely distorted portions in the generated objects.
Remotesensing 15 05654 g006
Figure 7. Examples of diverse background augmentation. (a) Displays five reference images selected from real backgrounds, encompassing various land cover types. (b) Represents our layout map. (c) Shows the results generated before improving the loss using pix2pixGAN, and (d) presents the generated results after improving the loss.
Figure 7. Examples of diverse background augmentation. (a) Displays five reference images selected from real backgrounds, encompassing various land cover types. (b) Represents our layout map. (c) Shows the results generated before improving the loss using pix2pixGAN, and (d) presents the generated results after improving the loss.
Remotesensing 15 05654 g007
Figure 8. Comparation of the output feature maps before the SPP layer of the yolov5 network. The yellow pixels represent the most attended regions, followed by green and light green. The SL method reduces the network’s focus on the background, the BA method further reduces the attention to the background and more accurately concentrates on the object regions. (a) Baseline dataset, (b) SL, (c) SL + BA.
Figure 8. Comparation of the output feature maps before the SPP layer of the yolov5 network. The yellow pixels represent the most attended regions, followed by green and light green. The SL method reduces the network’s focus on the background, the BA method further reduces the attention to the background and more accurately concentrates on the object regions. (a) Baseline dataset, (b) SL, (c) SL + BA.
Remotesensing 15 05654 g008
Figure 9. Visualization of the generated results. (a) Represents guidance using a multi-class mask, and (b) represents guidance using the distribution map. Compared to (a), the results generated using (b) contain more information about real objects and reduce deformation and distortion of objects.
Figure 9. Visualization of the generated results. (a) Represents guidance using a multi-class mask, and (b) represents guidance using the distribution map. Compared to (a), the results generated using (b) contain more information about real objects and reduce deformation and distortion of objects.
Remotesensing 15 05654 g009
Figure 10. Power spectrum comparison. Compared to the power spectrum distribution of SAR sample images (a), the power spectrum of a sample generated by the single-step diffusion model (b) lacks side-lobe distribution. However, the power spectrum of our multi-condition progressive generated sample (c) is more concentrated in the low-frequency region, with more pronounced side-lobes.
Figure 10. Power spectrum comparison. Compared to the power spectrum distribution of SAR sample images (a), the power spectrum of a sample generated by the single-step diffusion model (b) lacks side-lobe distribution. However, the power spectrum of our multi-condition progressive generated sample (c) is more concentrated in the low-frequency region, with more pronounced side-lobes.
Remotesensing 15 05654 g010
Table 1. Basic methods of image enhancement and concise introduction.
Table 1. Basic methods of image enhancement and concise introduction.
MethodsTransformationsIntroduction
Information TransformationTranslationMoving the image horizontally or vertically
RotationRotating the image at a certain angle
ScalingStretching or compressing the image in a ratio
MirroringFlipping the image horizontally or vertically
Noise AdditionAdding noise to the image
CroppingCropping different regions from the image
OverlayOverlaying multiple images together
IlluminationAdjusting the brightness, contrast, and color of the image
Information
Fusion
Image StitchingSeamlessly merging multiple images together
Image BlendingBlending two or more different images together
Image FusionFusing two or more different images together
Table 2. Basic experimental environment settings.
Table 2. Basic experimental environment settings.
PlatformWindows 11
CPUIntel Core i7-12700 k
Memory16 G
GPUNvidia GeForce RTX 3070
Video memory8 G
Table 3. Quantitative comparison of various augmented samples. Bold words represent the best results. ↑ and ↓ indicate the optimization direction of the indicator.
Table 3. Quantitative comparison of various augmented samples. Bold words represent the best results. ↑ and ↓ indicate the optimization direction of the indicator.
MethodKL (Mean)↑FID↓KL (Std)↑Matching Degree (%)↑
Original1.866-0.16392.9
CycleGAN1.34928.420.07018.5
Pix2pixGAN1.25228.470.03815.1
BicycleGAN1.12632.070.05610.2
DDPM1.87125.930.08448.6
Ours-SL1.782-0.01092.7
Ours-BA1.30030.170.04287.8
Ours-GE1.77425.540.09078.2
Ours-all1.88418.830.18081.8
Table 4. Comparison of detection results on classical networks with different augmentation strategies.
Table 4. Comparison of detection results on classical networks with different augmentation strategies.
NetworkMethodP (%)R (%)mAP (%)F1 (%)
Yolov5Baseline78.879.866.679.30
Random8080.367.380.15
Ours-SL81.180.168.780.60
Ours-SL+ BA82.981.569.782.19
Ours-all85.586.373.685.90
RetinaNetBaseline69.278.954.373.73
Random70.279.554.874.56
Ours-SL70.579.855.274.86
Ours-SL+ BA71.981.957.176.57
Ours-all73.28358.377.79
VarifocalNetBaseline74.481.959.577.97
Random76.782.661.679.54
Ours-SL81.783.466.482.54
Ours-SL+ BA82.283.86782.99
Ours-all85.684.670.485.10
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kuang, Y.; Ma, F.; Li, F.; Liu, Y.; Zhang, F. Semantic-Layout-Guided Image Synthesis for High-Quality Synthetic-Aperature Radar Detection Sample Generation. Remote Sens. 2023, 15, 5654. https://doi.org/10.3390/rs15245654

AMA Style

Kuang Y, Ma F, Li F, Liu Y, Zhang F. Semantic-Layout-Guided Image Synthesis for High-Quality Synthetic-Aperature Radar Detection Sample Generation. Remote Sensing. 2023; 15(24):5654. https://doi.org/10.3390/rs15245654

Chicago/Turabian Style

Kuang, Yi, Fei Ma, Fangfang Li, Yingbing Liu, and Fan Zhang. 2023. "Semantic-Layout-Guided Image Synthesis for High-Quality Synthetic-Aperature Radar Detection Sample Generation" Remote Sensing 15, no. 24: 5654. https://doi.org/10.3390/rs15245654

APA Style

Kuang, Y., Ma, F., Li, F., Liu, Y., & Zhang, F. (2023). Semantic-Layout-Guided Image Synthesis for High-Quality Synthetic-Aperature Radar Detection Sample Generation. Remote Sensing, 15(24), 5654. https://doi.org/10.3390/rs15245654

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop