Next Article in Journal
Deep Learning Investigation of Mercury’s Explosive Volcanism
Next Article in Special Issue
A Multicomponent Linear Frequency Modulation Signal-Separation Network for Multi-Moving-Target Imaging in the SAR-Ground-Moving-Target Indication System
Previous Article in Journal
Actual Evapotranspiration Dominates Drought in Central Asia
Previous Article in Special Issue
A Deep-Learning-Facilitated, Detection-First Strategy for Operationally Monitoring Localized Deformation with Large-Scale InSAR
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Mix MSTAR: A Synthetic Benchmark Dataset for Multi-Class Rotation Vehicle Detection in Large-Scale SAR Images

College of Nuclear Engineering, Rocket Force University of Engineering, Xi’an 710000, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(18), 4558; https://doi.org/10.3390/rs15184558
Submission received: 8 August 2023 / Revised: 7 September 2023 / Accepted: 11 September 2023 / Published: 16 September 2023
(This article belongs to the Special Issue Exploitation of SAR Data Using Deep Learning Approaches)

Abstract

:
Because of the counterintuitive imaging and confusing interpretation dilemma in Synthetic Aperture Radar (SAR) images, the application of deep learning in the detection of SAR targets has been primarily limited to large objects in simple backgrounds, such as ships and airplanes, with much less popularity in detecting SAR vehicles. The complexities of SAR imaging make it difficult to distinguish small vehicles from the background clutter, creating a barrier to data interpretation and the development of Automatic Target Recognition (ATR) in SAR vehicles. The scarcity of datasets has inhibited progress in SAR vehicle detection in the data-driven era. To address this, we introduce a new synthetic dataset called Mix MSTAR, which mixes target chips and clutter backgrounds with original radar data at the pixel level. Mix MSTAR contains 5392 objects of 20 fine-grained categories in 100 high-resolution images, predominantly 1478 × 1784 pixels. The dataset includes various landscapes such as woods, grasslands, urban buildings, lakes, and tightly arranged vehicles, each labeled with an Oriented Bounding Box (OBB). Notably, Mix MSTAR presents fine-grained object detection challenges by using the Extended Operating Condition (EOC) as a basis for dividing the dataset. Furthermore, we evaluate nine benchmark rotated detectors on Mix MSTAR and demonstrate the fidelity and effectiveness of the synthetic dataset. To the best of our knowledge, Mix MSTAR represents the first public multi-class SAR vehicle dataset designed for rotated object detection in large-scale scenes with complex backgrounds.

Graphical Abstract

1. Introduction

Thanks to its unique advantages, such as all-time, all-weather, high-resolution, and long-range detection, SAR has been widely used in various fields, such as land analysis and target detection. Vehicle detection in SAR-ATR is of great significance in urban traffic, hotspot target focusing, and other aspects.
In recent years, with the development of artificial intelligence, deep learning-based object detection algorithms [1,2] have dominated the field with their powerful capabilities in automatic feature extraction. Deep learning is a subject of data hunger. Historical experience has shown that big data is an important driver for the flourishing development of deep learning technology in various fields. With the rapid development of aerospace and sensor technology, an increasing number of high-resolution remote sensing images can be obtained. In the field of remote sensing, visible light object detection has experienced vigorous development since the release of the DOTA [3]. As the first publicly available SAR ship dataset, the introduction of the SSDD [4] has directly promoted the application of deep learning in SAR object detection and has led to the emergence of more SAR ship datasets [5,6,7,8,9], which are still one of the detection benchmarks and exhibit strong vitality to this day.
However, due to the imaging mechanism of SAR images being distinct from visible light, their interpretation is unintuitive for the human eye. Ground clutter and scattering caused by object corner points can seriously interfere with human interpretation. This leads to the fact that the detection objects of the current SAR datasets are mainly large targets such as ships and planes in relatively pure backgrounds. In contrast, SAR datasets for vehicles are very rare. The community has long relied on the Moving and Stationary Target Acquisition and Recognition (MSTAR) [10] released by the Sandia National Laboratory in the last century. However, the vehicle images in MSTAR are also separated from large-scene images and appear in the form of small patches. Due to its lack of complex background, it is only suitable for classification tasks, and its classification accuracy has reached more than 99%. Up until now, in the SAR-ATR field, MSTAR has been more widely used in few-shot learning and semi-supervised learning [11,12]. Meanwhile, the volume of the SAR dataset containing vehicle images with large scenes is quite small. The reason for this is that the small area of the vehicle requires higher resolution for SAR-ATR than aircrafts and ships, which leads to higher data acquisition costs. Moreover, vehicles exist in more complex clutter backgrounds, which increases the difficulty of manual interpretation and reduces the accuracy of target annotation. Table 1 provides detailed information on existing public SAR vehicle datasets with large scenes. Unfortunately, in these datasets, there is no official localization annotation that can be obtained, so manual identification of annotations is required. Due to strong noise interference, the FARAD X BAND [13] and FARAD KA BAND [14] make it too difficult for humans to identify the position of vehicles, so the annotations cannot meet the accuracy requirements. The Spotlight SAR [15] has only a very small number of vehicles, and pairs of pictures were taken at different time periods at the same location. The Mini SAR [16] includes more vehicles, but it only contains 20 pictures, and it has the same problem as the Spotlight SAR in having duplicate scenes. The subsequent experiments proved that the small size of the Mini SAR caused a large standard error in the results. These reasons make the above datasets difficult to use as a reliable benchmark for SAR-ATR algorithms. In addition, there is the GOTCHA [17], which contains vehicles and large scenes, but it is a fully polarized circular SAR dataset that is significantly different from the commonly used single-polarized linear SAR. It only contains one scene and is mainly used for the classification of calibrated vehicles in the SAR-ATR field. The size of GOTCHA makes it difficult to meet the requirements of object detection, so it is not included in the table here for comparison.
In view of the scarcity of vehicle datasets in the SAR-ATR field, people have conducted a series of data generation efforts around MSTAR, which can be mainly divided into the following three methods: The first method is based on generative adversarial nets (GANs) [18]. The generator network transforms the input noise into generative images that can deceive discriminator networks by fitting the distribution of real images. In theory, GANs [19,20,21] can generate an infinite number of generative images (see Figure 1a), thereby solving the problem of scarce real samples. However, unlike optical images, SAR imaging is strictly based on radar scattering mechanisms, and the black box properties of neural networks cannot prove that the generative samples comply with SAR imaging mechanisms. Moreover, due to the limitations of real samples, it is difficult to generate large-scene images. The second method is based on computer-aided design (CAD) 3D modeling and electromagnetic calculation simulation [22,23,24,25]. Among them is the SAMPLE [25] dataset released by Lewis et al. from the same institution as MSTAR, and it has advantages in terms of model errors, as shown in Figure 1b. The advantage of this method is that the imaging of synthetic samples is based on physical mechanisms, and the imaging under different conditions can be easily obtained by changing the simulation environment parameters. Compared with the original images, the simulation images can also remove the correlation between the targets and the background by setting random background noise, which prevents overfitting of the detection model. However, both of these methods have background limitations, and it is currently difficult to simulate vehicles located in complex, large-scale backgrounds. The third method is background transfer [26,27,28]. Chen et al. [26] believe that since the acquisition conditions of the chip image (Chip for short) and the clutter image (Clutter for short) in MSTAR are similar, Chips can be embedded in Clutters to generate vehicle images with large scenes, as shown in Figure 1c. Like the first method, the synthetic images cannot strictly comply with SAR imaging mechanisms, and the current use of such methods is to directly paste Chips with their backgrounds onto Clutters, which looks quite abrupt visually while maintaining an association between the target and the background.
To generate large-scale SAR images with complex backgrounds, we constructed the Mix MSTAR using an improved background transfer method. Unlike the previous works, we overcame the abrupt visual appearance of synthetic images and demonstrated the fidelity and effectiveness of synthetic data. Our key contributions are as follows:
  • We improved the method of background transfer and generated realistic synthetic data by linearly fusing vehicle masks in Chips and Clutters, resulting in the fusion of 20 types of vehicles (5392 in total) into 100 large background images. The dataset adopts rotation bounding box annotation and includes one Standard Operating Condition (SOC) and two EOC partitioning strategies, making it a challenging and diverse dataset;
  • Based on the Mix MSTAR, we evaluated nine benchmark models for general remote sensing object detection and analyzed their strengths and weaknesses for SAR-ATR;
  • To address potential artificial traces and data variance in synthetic images, we designed two experiments to demonstrate the fidelity and effectiveness of Mix MSTAR in SAR image features, demonstrating that Mix MSTAR can serve as a benchmark dataset for evaluating deep learning-based SAR-ATR algorithms.
The remaining article is composed of four sections. Section 2 presents the detailed methodology employed to construct the synthetic dataset as well as an extensive analysis of the dataset itself. In Section 3, we introduce and evaluate nine rotated object detectors using the synthetic dataset as the benchmark. Sequentially, a comprehensive analysis of the results is conducted. Section 4 focuses specifically on the analysis and validation of two vital problems related to the dataset, namely, artificial traces and data variance. Moreover, we provide an outlook on the potential future of the synthetic dataset. Section 5 concludes our work.

2. Materials and Methods

2.1. Preliminary Feasibility Assessment

We first evaluated the feasibility of merging Clutters and Chips. Since the sensor’s depression was 15° when collecting Clutters, we chose Chips with the same depression as the target images. As shown in Table 2, both Clutters and Chips use the same STARLOS sensor based on an airborne platform and maintain consistency in terms of radar center frequency, bandwidth, polarization, and depression. Although the radar mode is different, the final imaging resolution and pixel spacing are the same. Therefore, we assume that if the working parameters of Clutters are used for imaging vehicles, the visual effect will be approximately the same as that of Chips. So, it is feasible to transfer the vehicles in Chips to the Clutters’ backgrounds, and the final effect is in line with the human observation mechanism. Of course, we must acknowledge that due to the differences in the operating modes, the two have significant differences in the raw radar data (especially phase). This means that synthetic data generated by background transfer cannot strictly conform to the scattering mechanism of the radar. However, what we pursue is the consistency between synthetic data and real data in terms of 8-bit image features, which is crucial for current deep learning models based on image feature extraction in the computer vision field.
Unlike previous attempts that involved crude background transfers, Mix MSTAR aims to be a visually realistic synthetic dataset. To achieve this goal, we conducted extensive research into domain transfer and imaging algorithms to harmoniously blend two different radar datasets and developed a paradigm for creating synthetic datasets, as shown in Figure 2. Next, we will describe in detail the process of constructing the dataset.

2.2. Mask Extraction

In order to make vehicles fit seamlessly into the Clutters’ backgrounds, we used labelme [29] to mask the outlines of the vehicles on Chips. Since the shadow in the radar blind area is also an important feature of SAR targets, the shadow of the vehicle was included in the mask. We also labeled the OBBs of the vehicle on Chips to be used as the label for the final synthetic dataset. The four points of the OBBs are labeled in a clockwise direction, and the first point is on the left side of the vehicle’s front. It is worth noting that according to the principle of electromagnetic wave scattering, there will always be a part of the vehicle in the shadow area at any angle with weak but not negligible reflected signals. This ambiguity can cause interference in manual annotation. Therefore, to unify the standard, the strategy followed for annotating the OBBs is based on human visual perception. We only label the salient areas that can attract the attention of the human rather than including the entire actual occupation of the vehicle based on prior knowledge according to the object resolution and vehicle size, as shown in Figure 3b. This conforms to the annotation rules of computer vision and ensures that the model trained on this dataset focuses on features that are in line with human perception.

2.3. Data Harmonization

In fact, after extracting the masks of the vehicles in Chips, we can already embed the mask in Clutters as the foreground. However, prior to this step, it is necessary to harmonize the two kinds of data for the visual harmony of the synthetic image. In the field of image composition, image harmonization aims to adjust the foreground to make it compatible with the background in the composite image [30]. In visible light image harmonization, traditional methods [31,32,33] or deep learning-based methods [30,34,35,36] can already perfectly combine the foreground and background visually. However, in the strict imaging mechanism of SAR, pixel brightness corresponds to the intensity of radar echoes, which requires synthetic images to not only look visually harmonious but also conform to the physical mechanism. Therefore, we propose a domain transfer method that uses the same ground objects as prior information to harmonize the synthetic images, conforming to the SAR imaging mechanism as much as possible. Notably, in the following two steps, we apply data harmonization to the raw radar data with high bit depth to obtain more accurate results.

2.3.1. Domain Transfer

Since Chips and Clutters are two different types of data, their distribution and threshold values are different, so it is necessary to unify them reasonably based on their relationship before merging. Since the background is the main body in the synthetic images, we choose to transfer masks from domain Chips to domain Clutters. Based on the satellite map and information in the source files, we noticed that the background of Chips is a dry grassland, and Clutters also contains a large amount of grassland. Both were collected in Huntsville City, less than 26 km apart, and in the autumn season, so it can be assumed that the vegetation in the grasslands of the two places is similar. To validate this assumption, we annotated the grassland in nine Clutters and conducted data analysis with the grassland in seven kinds of Chips, collection dates of which were close to Clutters, as the region of interest (RoI). As shown in Table 3, it can be seen that the coefficient of variation calculated based on Formula (1) for both data is around 0.6, indicating similar data dispersion levels.
C V = σ μ
Based on the above analysis and given the similar data distribution of both data after being dimensionless, we linearly mapped the data of Chips to the data space of Clutters. According to Formula (2), we multiplied the data for Chips by the ratio coefficient K (K = 1371.8) of the mean value of the grassland in both RoIs and then rounded it. Following the pipeline shown in Figure 4a, we calculated the histograms of the grassland in transformed Chips and Clutter and calculated their cosine similarity (CSIM) according to Formula (3). From Figure 4b, it can be seen that the data distribution of the two datasets is very similar. In Table 3, the CSIM values for the two grasslands are all above 0.99. Therefore, K can be used as the mapping coefficient from domain Chips to domain Clutters, and the whole data of Chips can be harmonized by multiplying it by K.
K = Mean ( Clutter ) 1 c c Mean ( Chip )
CSIM = cos ( θ ) = a · b a · b

2.3.2. Brightness Uniformity

Schumacher et al. pointed out that the background and the targets of Chips are highly relevant [37,38]. Geng et al. conducted experiments and indicated that the SAR-ATR model recognizes vehicles by treating the brightness of the background as an important feature [39]. For instance, the background of BRDM2 is brighter than that of other types of vehicles, which causes the neural network to learn from the training data that “the brighter ones are more likely to be BRDM2” [39]. Thus, the SAR-ATR model cheats by recognizing the associated background to classify the vehicles. We discovered that this phenomenon is due to the nonlinear mapping of the official imaging algorithm, as seen in the left column of Table 4. ScaleAdj in the 11th step of the original algorithm is determined by the value of the most and least appearing pixels in each image. We found that the mean of ScaleAdj in BRDM2 is higher than that of other vehicles. Additionally, the non-uniform ScaleAdj results in different gray-level transformations for each category of vehicles and even for each image. Furthermore, for Clutters, the original algorithm produces very dark images. The reason for this lies in the high dynamic range of Clutters radar data, with most data in low values, and the maximum-minimum value stretching in the 3rd step, which results in most data being assigned low gray values.
Therefore, we believe that applying a uniform brightness transformation to the imaging algorithm is an effective way to avoid the aforementioned two problems, as shown in the right column of Table 4. The improved imaging algorithm maps the radar amplitude values linearly to the image gray values by setting a threshold and a linear transformation. Too high a threshold pools the low-value signals, while too low a threshold causes the loss of information from the high-value signals. Therefore, to preserve most of the image details while minimizing the loss of high-value signals, we set the threshold to 511, as 99.8% of the radar amplitudes in Clutters are less than this threshold and 95.5% for the mask of the vehicle in Chips. This approach linearly images the low-value signals and preserves most of the image details without significant loss of the high-value signals.

2.4. Embedded Synthesis

Based on OpenCV, our laboratory developed an interactive software that can conveniently embed vehicle masks of the specific category or specified azimuth angles at designated positions in the Clutters background. We follow the basic logic of radar scattering to select the embedding positions. First, we prevent the overlap of vehicle masks through logical settings at the code level. Second, we avoid placing vehicles above tall objects (such as trees or buildings) or their shadow areas. To achieve a seamless transition between the mask and the background at the edges, a 5 × 5 Gaussian operator is applied for smoothing filtering on the inner and outer circles of the mask edges. To investigate the impact of background objects and corner reflectors on SAR-ATR, we mark the recognition difficulty of vehicles near objects with strong reflection echoes, such as trees or buildings, as 1. Additionally, we embed corner reflectors with a 15° depression in Clutters and set the recognition difficulty of vehicles near them to 2. For other vehicle positions, the recognition difficulty is set by default to 0. As shown in Equation (4), the final label format follows the DOTA format [3], with each ground truth including the position of the four vertices of the rectangle, category, and difficulty. The position of the vertices of each rectangle is obtained from the rotated bounding box (shown in Figure 3) after coordinate transformation.
( x 1 , y 1 , x 2 , y 2 , x 3 , y 3 , x 4 , y 4 , category , difficult )

2.5. Analysis of the Dataset

In order to meet the requirements of large scene detection tasks, we selected 34 out of 100 Clutters that can be stitched together as the test set, and the remaining 66 Clutters serve as the train set. For the Chips partition, To create a challenging dataset, we combined one SOC and two EOC division strategies. As shown in Table 5, the first EOC strategy is based on version variants using BMP2sn-9563 as the train set and BMP2sn-9566 and BMP2sn-c21 as the test sets. The second EOC strategy is based on configuration variants using a 7:3 fine-grained partitioning of T72’s 11 subtypes. The rest of the eight vehicle categories are partitioned based on a 7:3 SOC strategy. Similarly, as described in Section 2.4, corner reflectors are embedded in a 7:3 ratio but are not used as detection objects. Finally, the proportion of total vehicles in the training set and the test set also remained at 66:34, which matches the partition of the Clutters. After the partitioning of the dataset, we fused Chips and Clutters according to the method described in Figure 2, resulting in 100 images. To simulate a realistic remote sensing application scenario, we stitched together the geographically contiguous images in the test set into four large images.
In summary, Mix MSTAR consists of 100 large images with 5392 vehicles in 20 fine-grained categories. The geographically contiguous test set can be stitched into four large images, as shown in Figure 5. The arrangement of vehicles is diverse, with both tight and sparse groupings and various scenes such as urban, highway, grassland, and forest.
As shown in the data analysis in Figure 6, the vehicle orientations are relatively uniformly distributed between [0–2π], and the vehicle areas fluctuate due to changes in azimuth angles, with different vehicles having different sizes. The aspect ratio of the vehicle ranges from 1 to over 3. According to the definition of object sizes in the COCO regulation [40], over 98% of the vehicles are small objects, which requires detection algorithms to have good small object detection capabilities. The number of vehicles in each Clutter is also uneven, ranging from 1 to over 90 vehicles, indicating the need for detection algorithms to be more robust to the issue of uneven sample distribution.

3. Results

After constructing Mix MSTAR, in order to further evaluate the dataset, nine benchmark models are selected in this section to evaluate the performance of mainstream rotated object detection algorithms on Mix MSTAR.

3.1. Models Selected

In the field of deep learning, the types of detectors can be roughly divided into single-stage, refinement-stage, two-stage, and anchor-free algorithms.
The single-stage algorithm directly predicts the class and bounding box coordinates for objects from the feature maps. It tends to be computationally more efficient, albeit at the potential cost of less precise localization.
The refinement stage algorithm is typically a supplementary step incorporated within an object detection process to enhance the precision of the detected bounding box coordinates, as proposed initially. It refines the spatial dimensions of bounding boxes via a series of regressors learning to make small iterative corrections towards the ground truth box, thereby improving the performance of object localization.
The two-stage algorithm operates on the principle of segregation between object localization and its classification. First, it generates region proposals through its region proposal network (RPN) stage based on the input images. Then, these proposals are run through the second stage, where the actual object detection takes place, discerning the object class and refining the bounding boxes. Due to this two-step process, these algorithms tend to be more accurate but slower.
Unlike traditional algorithms, which leverage anchor boxes as prior knowledge for object detection, anchor-free algorithms operate by directly predicting the object’s bounding box without relying on predetermined anchor boxes. They circumvent drawbacks such as choosing the optimal scale, ratio, and number of anchor boxes for different datasets and tasks. Furthermore, they simplify the pipeline of object detection models and have been successful in certain contexts on both efficiency and accuracy fronts.
To make the evaluation results more convincing, the nine algorithms cover the four kinds of algorithms mentioned above.

3.1.1. RotatedRetinanet

Retinanet [41] argues that the core reason why single-stage detectors underperform compared to two-stage models is due to the extreme foreground and background imbalance during training. To address this, the Focal Loss was proposed, which adds two weights to the binary cross-entropy loss to balance the importance of positive and negative samples. It also reduces the emphasis on easy samples so that the focus of training is on hard negatives. Retinanet is the first single-stage model with accuracy surpassing that of two-stage models. Based on it, Rotated Retinanet predicts an additional angle in the regression branch (x, y, w, h) without other modifications. The network architecture can be seen in Figure 7.

3.1.2. S2A-Net

S2A-Net [42] is a refinement stage model that proposes the Feature Alignment Module (FAM) on the basis of improving deformable convolution (DCN) [43]. The network architecture is shown in Figure 8. In the refinement stage, the horizontal anchor is refined to a rotated anchor by the Anchor Refinement Network (ARN), which is a learnable offset field module that is directly supervised by box annotations. Next, the feature map within the anchor is aligned and then convolved with the Alignment Convolution. This method eliminates the low-quality, heuristically defined anchors and addresses the uncorrelated problem between anchor boxes and the axis-aligned features they cause.

3.1.3. R3Det

R3Det [44] is a refinement stage model that proposes the Feature Refinement Module (FRM) for reconstructing the feature map according to the refined bounding box. Each point in the reconstructed feature map is obtained by adding five feature vectors consisting of five points (four corner points and the center point in the refined bounding box) after interpolation. FRM can alleviate the feature misalignment problems that exist in refined single-stage detectors and can be added multiple times for better performance. Additionally, an approximate SkewIoU loss is proposed, which can better reflect the real loss of SkewIoU while maintaining differentiability. The network architecture is depicted in Figure 9.

3.1.4. ROI Transformer

ROI Transformer [45] is a two-stage model that adds a learnable module from horizontal RoI (HRoI) to rotated RoI (RRoI). It generates HRoI based on a small number of horizontal anchors and proposes RRoI via the offset of the rotated ground truth relative to HRoI. This operation eliminates the need to preset a large number of rotated anchors with different angles for directly generating RRoI. In the next step, the proposed Rotated Position Sensitive RoI Align extracts rotation-invariant features from the feature map and RRoI to enhance subsequent classification and regression. This study also examines the advantages of retaining appropriate context in RRoI for enhancing the detector’s performance. The network architecture of ROI Transformer is shown in Figure 10.

3.1.5. Oriented RCNN

Oriented RCNN [46] is built upon the Faster RCNN [2] and proposes an efficient oriented RPN network. The overall framework is shown in Figure 11. The oriented RPN uses a novel six-parameter mid-point offset representation to represent the offsets of the rotated ground truth relative to the horizontal anchor box and generate a quadrilateral proposal. Compared with RRPN [47], it avoids the huge amount of calculation caused by presetting a large number of rotating anchor boxes. Compared to ROI Transformer, it converts horizontal anchor boxes into oriented proposals in a single step, greatly reducing the parameter amount of the RPN network. Efficient and high-quality oriented proposal networks make Oriented RCNN both high-accuracy and high-speed.

3.1.6. Gliding Vertex

Gliding Vertex [48] introduces a robust OBB representation that addresses the limitations of predicting vertices and angles. Specifically, on the regression branch of RCNN, four extra length ratio parameters are used to slide the corresponding vertex on each side of the horizontal bounding box. This approach avoids the problem of order confusion when directly predicting the position of the four vertices. It also mitigates the high sensitivity issue caused by predicting the angle. Additionally, with the idea of divide and conquer, an area ratio parameter r is used to predict the obliquity of the bounding box. This parameter can guide the regression in the Horizontal Bounding Box method, or OBB method, resolving the confusion issue of nearly-horizontal objects. The network architecture is shown in Figure 12.

3.1.7. ReDet

ReDet [49] argues that the regular CNNs are not equivariant to the rotation and that rotated data augmentation, or RRoI Align, can only approximate rotation invariance. To address this issue, ReDet uses e2cnn theory [50] to design a new rotational equivariant backbone called ReResNet, which is based on ResNet [1]. The new backbone features a higher degree of rotation weight sharing, allowing it to extract rotation-equivariant features. Additionally, the paper proposes Rotation-Invariant RoI Align, which performs warping on the spatial dimension and then circularly switches channels to interpolate and align on the orientation dimension to produce completely rotation-invariant features. The overall network architecture can be seen in Figure 13.

3.1.8. Rotated FCOS

FCOS [51] is an anchor-free, one-stage detector that employs a full convolution structural design. Unlike traditional detectors, FCOS eliminates the need for presetting anchors, thereby avoiding complex anchor operations, sensitive and heuristic hyperparameter settings, and the large number of parameters and calculations associated with anchors. FCOS employs the four distances (l, r, t, and b) between the feature point and the four sides of the bounding box as the prediction format. The distance between the center point and the feature point is used to measure the bounding box’s center-ness, which is then multiplied by the classification score to obtain the final confidence. The multi-level prediction based on FPN [52] alleviates the influence of overlapping ambiguous samples. Rotated FCOS is a re-implementation of FCOS in rotated object detection that adds an additional angle branch parallel to the regression branch. The network architecture is shown in Figure 14.

3.1.9. Oriented RepPoints

Based on RepPoints [53], Oriented RepPoints [54] summarizes three ways of converting a point set to an OBB, making it suitable for detecting aerial objects. Inherited from RepPoints, Oriented RepPoints combines DCN [43] with anchor-free key-point detection, enabling the model to extract non-axis-aligned features from an aerial perspective. To constrain the spatial distribution of point sets, the proposed spatially constrained loss constrains the vulnerable outliers within their instance owner and uses GIOU [55] to quantify localization loss. Additionally, the proposed Adaptive Points Assessment and Assignment adopts four metrics to evaluate the quality of learning point sets and use them to determine positive samples. The network architecture is shown in Figure 15.

3.2. Evaluation Metrics

In rotated object detection, the ground truth of the object’s position and the bounding box predicted by the model are oriented bounding boxes. Similar to generic target detection, rotated target detection uses Intersection over Union (IoU) to measure the quality of the predicted bounding box:
IoU OBB = area ( OBB pd OBB gt ) area ( OBB pd OBB gt )
In the classification stage, based on the combination of the prediction bounding box and the ground truth, four results are produced: True Positives (TP), True Negatives (TN), False Negatives (FN), and False Positives (FP). Judging whether a prediction bounding box is TP is based on its IoU with the ground truth. If the IoU exceeds the threshold, it is TP; otherwise, it is FP. In the object detection of remote sensing, the IoU threshold is generally set to 0.5, as in this paper. Precision and recall are formulated as follows:
precision = TP TP + FP
recall = TP TP + FN
Based on precision and recall, AP is defined as the area under the precision-recall (P-R) curve, while Mean Average Precision (mAP) is defined as the mean of AP values across all classes:
AP = 0 1 p ( r )   dr
mAP = 1 c i = 1 c AP i
The F1 score is the harmonic mean of precision and recall, which is defined as:
F 1 = 2 precision recall precision + recall

3.3. Environment and Details

All experiments were implemented in Ubuntu 20.04.4 with Python3.8.10, Pytorch1.11.0, and Cuda11.3. The CPU is an Intel i9 11900k @3.5GHz with 32 GB of RAM, and the GPU is an Nvidia GeForce RTX 3090 (24GB) with driver version 470.103.01.
All models involved in this article are implemented through the MMRotate [56] framework. For fair comparison, the backbone network of each detector is ResNet50 [1], pretrained on ImageNet [57] by default, and the neck is FPN [52]. Each image in the train set was cropped into 4 pieces of 1024 ∗ 1024, and the four large-scene images in the test set were split into a series of 1024 ∗ 1024 patches with a stride of 824, which follows the setting of DOTA [3]. To display the performance of each detector on Mix MSTAR as fairly as possible, we simply follow these settings without additional embellishments: data augmentation used a random flip with a probability of 0.25 on horizontal, vertical, or diagonal axes. Each model was trained for 180 epochs with 2 images per batch. The optimizer was SGD with an initial learning rate of 0.005, momentum of 0.9, and weight decay of 1 × 10−4. The L2 norm was adopted for gradient clipping, with the maximum gradient set to 35. The learning rate decayed by a factor of 10 at the 162nd and 177th epochs. Linear preheating was used for the first 500 iterations, with the initial preheating learning rate set to 1/3 of the initial learning rate. The mAP and its standard error for all models in this article were obtained by training the network with three different random seeds. The final result is obtained by mapping the prediction of each small picture to the big picture and applying NMS. More details can be found in our log files.

3.4. Result Analysis on Mix MSTAR

The evaluation results for nine models on Mix MSTAR are shown in Table 6, while the class-wise AP results are shown in Table 7. It is important to note that in Table 6, Precision, Recall, and F1-score are calculated based on the statistics of all categories of TP, FP, and FN.
Combining the results and the previous analysis of the model and the dataset, we can draw the following conclusions:
  • In terms of the mAP metric, Oriented RepPoints achieved the best accuracy, which we attribute to its unique proposal approach based on sampling points. This approach successfully combines deformation convolution and non-axis-aligned feature extraction. Additionally, being a two-stage model, its feature extraction is more accurate. Compared to other refine-stage models, it has more sampling points, up to 9, which makes the extracted features more comprehensive. However, the heavy use of deformation convolution has made its training speed slow. The two-stage model performs better than the single-stage network due to the initial screening of the RPN network. However, the performance of Gliding Vertex is average, which may be due to its failure to use directed proposals in the first stage, resulting in inaccurate feature extraction. ReDet has poor performance, possibly because the rotation-invariant network used in ReDet is not suitable for SAR images with low depression. Mix MSTAR is simulated at a depression of 15°, and the shadow area is quite large, leading to significant imaging differences for the same object under different azimuth angles. For example, rotating a vehicle image at an angle of θ by α degrees would produce an image that is significantly different from the image of the same vehicle captured at (θ + α) degrees, which may cause ReResNet to extract incorrect rotation-invariant features. Compared to single-stage models, refined-stage models demonstrate a significant performance improvement, suggesting that refined-stage models are more accurate in extracting non-axis-aligned features of rotated objects, which can reduce the gap between refined-stage models and two-stage models. While the performance of R3Det is slightly inferior, it is similar to ReDet, and its reason may lie in the sampling points in its refined stage, which are fixed at the four corners and the center point. In low-pitch-angle SAR images, one vertex far from the radar sensor is necessarily shaded, which means that the feature extraction of the sampling point interferes with the overall feature expression. S2A-Net uses deformation convolution, with the position of the sampling point being learnable. Although there is still a probability of collecting data from the shaded vertexes, there are nine sampling points, which dilutes the influence of features from the shaded vertexes;
  • In terms of speed, Rotated FCOS performs the best, benefiting from its anchor-free design and full convolution structure. Its parameters and computation are both lower than those of Rotated Retinanet. In contrast, other models use deformation convolution, non-conventional feature Alignment Convolution, or non-full convolution structures, making network speed relatively slow. Due to its special rotation-equivariant convolution, ReDet has the slowest inference speed, even though its parameters and computation are the lowest. In terms of parameter quantities, the two anchor-free models and the single-stage model have fewer parameters than other models. The RPN of the ROI Transformer requires two stages to extract the rotation ROI, so it has the most parameters. In terms of computation, due to its multi-head design, the detection head of the single-stage model is too cumbersome, making its computation not significantly lower than that of the two-stage model. However, Mix MSTAR is a small target data set, with most of its ground truth width being below 32. After five downsamplings, its localization information has been lost. A better balance may be obtained by optimizing the regression subnetwork of layers with downsample sizes greater than 32;
  • In terms of precision and recall metrics, all networks tend to maintain high recall. As using inter-class NMS limits the Recall integration range of mAP, like the DOTA, inter-class NMS is disabled. But this resulted in lower accuracy. Among them, ROI Transformer achieved a balance between accuracy and recall and obtained the highest F1 score;
  • From the results presented in Table 7, it is evident that the fine-grained classification result of the T72 tank is poor and has a significant impact on all detectors. Figure 16a further illustrates this point, as the confusion matrix of Oriented RepPoints indicates a considerable number of FP assigned to wrong subtypes of the T72 tank, which is also observed in cross-category confusion intervals such as BTR70-BTR60, 2S1-T62, and T72-T62. Another notable observation is the poor detection effect of BMP2 under EOC, as indicated in the confusion matrix. Many BMP2 subtypes that did not appear in the train set were mistaken for other vehicles in testing. Figure 16b depicts the P-R curves of all detectors;
  • Figure 17 presents the detection results of three detectors on the same picture. The results showed that the localization of the vehicles was accurate, but the recognition accuracy was not high, with a small number of false positives and misses. Additionally, we discovered two unknown vehicles in the scene, which were initially hidden among the Clutters and did not belong to the Chips. One vehicle was recognized as T62 by all three models, while the other vehicle was classified as background, possibly because its area was significantly larger than the vehicles in the Mix MSTAR. This indicates that the model trained by Mix MSTAR has the ability to recognize real vehicles.

4. Discussion

For a synthetic dataset that aims to become a detection benchmark, both fidelity and effectiveness are essential. However, in the production of Mix MSTAR, it is necessary to manually extract vehicles from Chips and fuse radar data collected in different modes before generating the final image. Thus, there are two potential problems in this process that will affect the visual effectiveness of the synthetic images:
  • Artificial traces: The vehicle masks manually extracted can alter the contour features of the vehicles and leave artificial traces in the synthetic images. Even though Gaussian smoothing was applied to reduce this effect on the vehicle edges, theoretically, these traces could still be utilized by overfitting models to identify targets;
  • Data variance: The vehicle and background data in Mix MSTAR were collected under different operating modes. Although we harmonized the data amplitude based on reasonable assumptions, Chips was collected using spotlight mode, while Clutters used strip mode. The two different scanning modes of radar can cause variances in the image style (particularly spatial distribution) of the foreground and background in the synthetic images. This could lead detection models to find some cheating shortcuts due to the non-realistic effects of the synthetic images, failing to extract common image features.
To address these concerns, we designed two separate experiments to demonstrate the reliability of the synthetic dataset.

4.1. The Artificial Traces Problem

To address the potential problem of artificial traces and to prove the fidelity of the synthetic dataset, our approach was to use a model trained in Mix MSTAR to detect intact vehicle images. We randomly selected 25 images from the Chips and expanded them to 204 × 204 to maintain their original size. These images were then stitched into a 1024 × 1024 large image, which was input into the ROI Transformer trained on Mix MSTAR. As shown in Figure 18a, all these intact vehicles were accurately localized, with a classification accuracy of 80%.
However, an accuracy of 80% is not an ideal result, as the background in Chips is quite simple and the five misidentified vehicles were all subtypes of T72. As a comparison experiment, we trained and tested ResNet18 as a classification model on the 20 classes of Chips of MSTAR, following the same partition strategy as Mix MSTAR, and the classifier easily achieved 92.22% accuracy. However, we found through class activation maps [58] that since each type of vehicle in MSTAR was captured at different angles but at the same location, the high correlation between the backgrounds in Chips causes the classifier to focus more on the terrain than the vehicles themselves. As shown in Figure 19, the two subtypes of T72 were identified based on their tracks and unusual vegetation, with recognition rates of 98.77% and 100%, respectively. However, the accuracy of the two T72 subtypes that did not benefit from background correlation was only 73.17% and 66.67%, respectively. This phenomenon also existed in other types of vehicles, indicating that the training results of using background-correlated Chips are actually unreliable.
Through the detection of intact vehicles in real images, we have proven that the artificial traces generated in the process of mask extraction did not affect the models. On the contrary, benefiting from mask extraction and background transfer, Mix MSTAR eliminated background correlation, allowing models trained on the high-fidelity synthetic images to focus on vehicle features such as shadows and bright spots, as shown in Figure 18b.

4.2. The Data Variance Problem

To address the potential data variance problem and demonstrate the authentic detection capability of models obtained from Mix MSTAR, we designed the following experiment to prove the effectiveness of Mix MSTAR. The real dataset, Mini SAR, was used to train and evaluate models pretrained on Mix MSTAR and those not pretrained on Mix MSTAR. For the pretrained models, we froze the weights of the first stage of the backbone, forcing the network to extract features in the same way as it does with synthetic images. The non-pretrained models were loaded from ImageNet weights as a regular setting. We selected nine images containing vehicles as the dataset, and seven were used for training and two for validation. The images were divided into 1024 × 1024 images with a stride of 824. Since the dataset was very small, the training process for each network was unstable. Therefore, we extended the number of iterations to 240 epochs, recorded the mAP of the model on the validation set after each epoch, and set the learning rate to reduce 10 folds at the 160th epoch and the 220th epoch, with all other settings consistent with those in the Mix MSTAR experiments. It is worth noting that there is no perfect unified training setting that can fit all detectors due to their different feature extraction capabilities and the propensity for overfitting on the small dataset. Thus, we record the best results of the validation set during training in Table 8.
Firstly, as shown in Table 8, all models obtained improvements after being pretrained on Mix MSTAR. Since the weights of the first layer are frozen after pretraining, this indicates that the models effectively learn how to extract general underlying features from SAR images. Secondly, since the validation set contains only two images, the results of non-pretrained models were very unstable, but the standard errors of all models were significantly reduced after pretraining on Mix MSTAR. Additionally, as shown in Figure 20, the pretrained models had very rapid loss reduction during the training process. After a few epochs, their accuracy on the validation set increased significantly and ultimately reached a relatively stable result (see Figure 21). However, the loss and mAP of the non-pretrained models were unstable.
We noticed that Rotated RetinaNet and Rotated FCOS are very sensitive to the random seed initialization, making them prone to training failure. This may be due to the weak ability of single-stage detectors in feature extraction, which makes it difficult for them to learn effective feature extraction capabilities from a small quantity of data. Therefore, we conducted a comparison experiment in which we added the Mix MSTAR train set to the Mini SAR train set to increase the data size when training the non-pretrained models. As shown in Table 9, both single-stage models obtained significant improvements after mixed training with the two datasets. As seen in Figure 22, pretraining on Mix MSTAR or mixed training with Mix MSTAR both resulted in increased recall and precision of the models, achieving more accurate bounding box regression.
Based on the above comparison experiments using real data, we have demonstrated the effectiveness of Mix MSTAR, indicating that synthetic data can also help networks learn how to extract features from real SAR images, thereby proving the effectiveness and transferability ability of Mix MSTAR. In addition, the experiment shows that the unstable Mini SAR is not suitable as the benchmark dataset for algorithm comparison, especially for the single-stage model, and also verifies that the Mix MSTAR is effective in addressing the problem of insufficient real data for SAR vehicle detection.

4.3. Potential Application

As more and more creative work leverages synthetic data to advance human understanding of the real world, Mix MSTAR, as the first public SAR vehicle multi-class detection dataset, has many potential applications. Here, we envision two potential use cases:
  • SAR image generation. While mutual conversion between optical and SAR imagery is no longer a groundbreaking achievement, current style transfer methods between visible light and SAR are primarily used for low-resolution terrain classification [59]. Given the scarcity of high-resolution SAR images and the abundance of high-resolution labeled visible light images, a promising avenue is to combine the two to generate more synthetic SAR images to address the lack of labeled SAR data and ultimately improve real SAR object detection. Although the synthetic image obtained in this way cannot be used for model evaluation, it can help the detection model obtain stronger positioning ability when detecting real SAR objects through pre-training or mixed training. Figure 23 demonstrates an example of using CycleGAN [60] to transfer vehicle images from the DOTA domain to the Mix MSTAR domain;
  • Out-of-distribution detection. Out-of-distribution detection, or OOD detection, aims to detect test samples that are drawn from a distribution that is different from the training distribution [61]. Using the model trained by synthetic images to classify real images was regarded as a challenging problem in SAMPLE [25]. Unlike visible-light imagery, SAR imaging is heavily influenced by sensor operating parameters, resulting in significant stylistic differences between images captured under different conditions. Our experiments found that current models’ performance on different SAR datasets is poorly generalizable. If reannotation and retraining are required for every new dataset, the cost will increase significantly, exacerbating the scarcity of SAR imagery and limiting the application scenarios of SAR-ATR. Therefore, it is an important research direction to use the limited labeled datasets to detect more unlabeled data. We used the Redet model trained on Mix MSTAR to detect real vehicles in an image from FARAD KA BAND. Due to resolution differences, three vehicles were detected after applying multi-scale test techniques, as shown in Figure 24.

5. Conclusions

This research released a large-scale SAR image synthesis dataset for multi-class rotated vehicle detection and proposed a paradigm for realistically fusing SAR data from different domains. Upon evaluating nine different benchmark detectors, we found that fine-grained classification makes Mix MSTAR highly challenging, with considerable room for improving object detection performance. Additionally, to address concerns over potential artificial traces and data variance in synthetic data, we conducted two experiments to demonstrate the fidelity and effectiveness of Mix MSTAR. Finally, we summarized two potential applications of Mix MSTAR and called on the community to enhance communication and cooperation in SAR data sharing to alleviate the scarcity of data and promote the development of SAR.

Author Contributions

Conceptualization, Z.L. and S.L.; methodology, Z.L.; software, S.L.; validation, Z.L., S.L. and Y.W.; formal analysis, Z.L.; investigation, Z.L.; resources, Z.L.; data curation, S.L.; writing—original draft preparation, S.L.; writing—review and editing, Z.L.; visualization, Z.L.; supervision, Z.L.; project administration, Z.L.; funding acquisition, Y.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded in part by the National Natural Science Foundation of China under Grant 61905285 and in part by the Young Talent Fund of the University Association for Science and Technology in Shaanxi, China, under Grant 20200704.

Data Availability Statement

The dataset and the training log files can be obtained by contacting us through [email protected], which those are welcome to do so.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  2. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. Adv. Neural Inf. Process. Syst. 2015, 28, 1137–11493. [Google Scholar] [CrossRef] [PubMed]
  3. Xia, G.-S.; Bai, X.; Ding, J.; Zhu, Z.; Belongie, S.; Luo, J.; Datcu, M.; Pelillo, M.; Zhang, L. DOTA: A large-scale dataset for object detection in aerial images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3974–3983. [Google Scholar]
  4. Li, J.; Qu, C.; Shao, J. Ship detection in SAR images based on an improved faster R-CNN. In Proceedings of the 2017 SAR in Big Data Era: Models, Methods and Applications (BIGSARDATA), Beijing, China, 13–14 November 2017; pp. 1–6. [Google Scholar]
  5. Wang, Y.; Wang, C.; Zhang, H.; Dong, Y.; Wei, S. A SAR dataset of ship detection for deep learning under complex backgrounds. Remote Sens. 2019, 11, 765. [Google Scholar] [CrossRef]
  6. Xian, S.; Zhirui, W.; Yuanrui, S.; Wenhui, D.; Yue, Z.; Kun, F. AIR-SARShip-1.0: High-resolution SAR ship detection dataset. J. Radars 2019, 8, 852–863. [Google Scholar] [CrossRef]
  7. Wei, S.; Zeng, X.; Qu, Q.; Wang, M.; Su, H.; Shi, J. HRSID: A high-resolution SAR images dataset for ship detection and instance segmentation. IEEE Access 2020, 8, 120234–120254. [Google Scholar] [CrossRef]
  8. Zhang, T.; Zhang, X.; Ke, X.; Zhan, X.; Shi, J.; Wei, S.; Pan, D.; Li, J.; Su, H.; Zhou, Y. LS-SSDD-v1. 0: A deep learning dataset dedicated to small ship detection from large-scale Sentinel-1 SAR images. Remote Sens. 2020, 12, 2997. [Google Scholar] [CrossRef]
  9. Lei, S.; Lu, D.; Qiu, X.; Ding, C. SRSDD-v1. 0: A high-resolution SAR rotation ship detection dataset. Remote Sens. 2021, 13, 5104. [Google Scholar] [CrossRef]
  10. The Air Force Moving and Stationary Target Recognition Database. Available online: https://www.sdms.afrl.af.mil/datasets/mstar/ (accessed on 10 March 2011).
  11. Zhang, L.; Leng, X.; Feng, S.; Ma, X.; Ji, K.; Kuang, G.; Liu, L. Domain knowledge powered two-stream deep network for few-shot SAR vehicle recognition. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–15. [Google Scholar] [CrossRef]
  12. Zhang, L.; Leng, X.; Feng, S.; Ma, X.; Ji, K.; Kuang, G.; Liu, L. Azimuth-Aware Discriminative Representation Learning for Semi-Supervised Few-Shot SAR Vehicle Recognition. Remote Sens. 2023, 15, 331. [Google Scholar] [CrossRef]
  13. SANDIA FARAD SAR DATA COLLECTION—X BAND—4” RESOLUTION. Available online: https://www.sandia.gov/files/radar/complex-data/FARAD_X_BAND.zip (accessed on 30 April 2023).
  14. SANDIA FARAD SAR DATA COLLECTION—KA BAND—4” RESOLUTION. Available online: https://www.sandia.gov/files/radar/complex-data/FARAD_KA_BAND.zip (accessed on 30 April 2023).
  15. SANDIA Spotlight SAR. Available online: https://www.sandia.gov/files/radar/complex-data/20060214.zip (accessed on 30 April 2023).
  16. SANDIA Mini SAR Complex Imagery. Available online: https://www.sandia.gov/files/radar/complex-data/MiniSAR20050519p0009image003.zip (accessed on 30 April 2023).
  17. Casteel, C.H., Jr.; Gorham, L.A.; Minardi, M.J.; Scarborough, S.M.; Naidu, K.D.; Majumder, U.K. A challenge problem for 2D/3D imaging of targets from a volumetric data set in an urban environment. In Proceedings of the Algorithms for Synthetic Aperture Radar Imagery XIV, Orlando, FL, USA, 9–13 April 2007; pp. 97–103. [Google Scholar]
  18. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial networks. Commun. ACM 2020, 63, 139–144. [Google Scholar] [CrossRef]
  19. Gao, F.; Yang, Y.; Wang, J.; Sun, J.; Yang, E.; Zhou, H. A deep convolutional generative adversarial networks (DCGANs)-based semi-supervised method for object recognition in synthetic aperture radar (SAR) images. Remote Sens. 2018, 10, 846. [Google Scholar] [CrossRef]
  20. Cui, Z.; Zhang, M.; Cao, Z.; Cao, C. Image data augmentation for SAR sensor via generative adversarial nets. IEEE Access 2019, 7, 42255–42268. [Google Scholar] [CrossRef]
  21. Vignaud, L. GAN4SAR: Generative Adversarial Networks for Synthetic Aperture Radar imaging of targets signature. In Proceedings of the SET-273 Specialists Meeting on Multidimensional Radar Imaging and ATR-CfP, Marseille, France, 25–26 October 2021. [Google Scholar]
  22. Auer, S.; Bamler, R.; Reinartz, P. RaySAR-3D SAR simulator: Now open source. In Proceedings of the 2016 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Beijing, China, 10–15 July 2016; pp. 6730–6733. [Google Scholar]
  23. Malmgren-Hansen, D.; Kusk, A.; Dall, J.; Nielsen, A.A.; Engholm, R.; Skriver, H. Improving SAR automatic target recognition models with transfer learning from simulated data. IEEE Geosci. Remote Sens. Lett. 2017, 14, 1484–1488. [Google Scholar] [CrossRef]
  24. Cha, M.; Majumdar, A.; Kung, H.; Barber, J. Improving SAR automatic target recognition using simulated images under deep residual refinements. In Proceedings of the 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; pp. 2606–2610. [Google Scholar]
  25. Lewis, B.; Scarnati, T.; Sudkamp, E.; Nehrbass, J.; Rosencrantz, S.; Zelnio, E. A SAR dataset for ATR development: The Synthetic and Measured Paired Labeled Experiment (SAMPLE). In Proceedings of the Algorithms for Synthetic Aperture Radar Imagery XXVI, Baltimore, MD, USA, 14–18 April 2019; pp. 39–54. [Google Scholar]
  26. Chen, S.; Wang, H.; Xu, F.; Jin, Y.-Q. Target classification using the deep convolutional networks for SAR images. IEEE Trans. Geosci. Remote Sens. 2016, 54, 4806–4817. [Google Scholar] [CrossRef]
  27. Han, Z.-S.; Wang, C.-P.; Fu, Q. Arbitrary-oriented target detection in large scene sar images. Def. Technol. 2020, 16, 933–946. [Google Scholar] [CrossRef]
  28. Sun, Y.; Wang, W.; Zhang, Q.; Ni, H.; Zhang, X. Improved YOLOv5 with transformer for large scene military vehicle detection on SAR image. In Proceedings of the 2022 7th International Conference on Image, Vision and Computing (ICIVC), Xi’an, China, 26–28 July 2022; pp. 87–93. [Google Scholar]
  29. labelme: Image Polygonal Annotation with Python (Polygon, Rectangle, Circle, Line, Point and Image-Level Flag Annotation). Available online: https://github.com/wkentaro/labelme (accessed on 30 April 2023).
  30. Cong, W.; Zhang, J.; Niu, L.; Liu, L.; Ling, Z.; Li, W.; Zhang, L. Dovenet: Deep image harmonization via domain verification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 8394–8403. [Google Scholar]
  31. Reinhard, E.; Adhikhmin, M.; Gooch, B.; Shirley, P. Color transfer between images. IEEE Comput. Graph. Appl. 2001, 21, 34–41. [Google Scholar] [CrossRef]
  32. Pérez, P.; Gangnet, M.; Blake, A. Poisson image editing. ACM Siggraph 2003, 2, 313–318. [Google Scholar] [CrossRef]
  33. Sunkavalli, K.; Johnson, M.K.; Matusik, W.; Pfister, H. Multi-scale image harmonization. ACM Trans. Graph. (TOG) 2010, 29, 1–10. [Google Scholar] [CrossRef]
  34. Tsai, Y.-H.; Shen, X.; Lin, Z.; Sunkavalli, K.; Lu, X.; Yang, M.-H. Deep image harmonization. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 3789–3797. [Google Scholar]
  35. Zhang, L.; Wen, T.; Shi, J. Deep image blending. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–8 January 2022; pp. 231–240. [Google Scholar]
  36. Ling, J.; Xue, H.; Song, L.; Xie, R.; Gu, X. Region-aware adaptive instance normalization for image harmonization. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, Nashville, TN, USA, 20–25 June 2021; pp. 9361–9370. [Google Scholar]
  37. Schumacher, R.; Rosenbach, K. ATR of battlefield targets by SAR classification results using the public MSTAR dataset compared with a dataset by QinetiQ UK. In Proceedings of the RTO SET Symposium on Target Identification and Recognition Using RF Systems, Oslo, Norway, 11–13 October 2004. [Google Scholar]
  38. Schumacher, R.; Schiller, J. Non-cooperative target identification of battlefield targets-classification results based on SAR images. In Proceedings of the IEEE International Radar Conference, Arlington, VA, USA, 9–12 May 2005; pp. 167–172. [Google Scholar]
  39. Geng, Z.; Xu, Y.; Wang, B.-N.; Yu, X.; Zhu, D.-Y.; Zhang, G. Target Recognition in SAR Images by Deep Learning with Training Data Augmentation. Sensors 2023, 23, 941. [Google Scholar] [CrossRef]
  40. Lin, T.-Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft coco: Common objects in context. In Proceedings of the Computer Vision–ECCV 2014: 13th European Conference Part V 13, Zurich, Switzerland, 6–12 September 2014; pp. 740–755. [Google Scholar]
  41. Lin, T.-Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the IEEE international conference on computer vision, Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  42. Han, J.; Ding, J.; Li, J.; Xia, G.-S. Align deep features for oriented object detection. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–11. [Google Scholar] [CrossRef]
  43. Dai, J.; Qi, H.; Xiong, Y.; Li, Y.; Zhang, G.; Hu, H.; Wei, Y. Deformable convolutional networks. In Proceedings of the IEEE international conference on computer vision, Venice, Italy, 22–29 October 2017; pp. 764–773. [Google Scholar]
  44. Yang, X.; Yan, J.; Feng, Z.; He, T. R3det: Refined single-stage detector with feature refinement for rotating object. In Proceedings of the AAAI conference on artificial intelligence, Washington, DC, USA, 7–14 February 2023; pp. 3163–3171. [Google Scholar]
  45. Ding, J.; Xue, N.; Long, Y.; Xia, G.-S.; Lu, Q. Learning roi transformer for oriented object detection in aerial images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2019; pp. 2849–2858. [Google Scholar]
  46. Xie, X.; Cheng, G.; Wang, J.; Yao, X.; Han, J. Oriented R-CNN for object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 3520–3529. [Google Scholar]
  47. Ma, J.; Shao, W.; Ye, H.; Wang, L.; Wang, H.; Zheng, Y.; Xue, X. Arbitrary-oriented scene text detection via rotation proposals. IEEE Trans. Multimed. 2018, 20, 3111–3122. [Google Scholar] [CrossRef]
  48. Xu, Y.; Fu, M.; Wang, Q.; Wang, Y.; Chen, K.; Xia, G.-S.; Bai, X. Gliding vertex on the horizontal bounding box for multi-oriented object detection. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 43, 1452–1459. [Google Scholar] [CrossRef] [PubMed]
  49. Han, J.; Ding, J.; Xue, N.; Xia, G.-S. Redet: A rotation-equivariant detector for aerial object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 2786–2795. [Google Scholar]
  50. Weiler, M.; Cesa, G. General e (2)-equivariant steerable cnns. arXiv 2019, arXiv:1911.08251. [Google Scholar]
  51. Tian, Z.; Shen, C.; Chen, H.; He, T. Fcos: Fully convolutional one-stage object detection. In Proceedings of the IEEE/CVF international conference on computer vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 9627–9636. [Google Scholar]
  52. Lin, T.-Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE conference on computer vision and pattern recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  53. Yang, Z.; Liu, S.; Hu, H.; Wang, L.; Lin, S. Reppoints: Point set representation for object detection. In Proceedings of the IEEE/CVF international conference on computer vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 9657–9666. [Google Scholar]
  54. Li, W.; Chen, Y.; Hu, K.; Zhu, J. Oriented reppoints for aerial object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 1829–1838. [Google Scholar]
  55. Rezatofighi, H.; Tsoi, N.; Gwak, J.; Sadeghian, A.; Reid, I.; Savarese, S. Generalized intersection over union: A metric and a loss for bounding box regression. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 658–666. [Google Scholar]
  56. Zhou, Y.; Yang, X.; Zhang, G.; Wang, J.; Liu, Y.; Hou, L.; Jiang, X.; Liu, X.; Yan, J.; Lyu, C. MMRotate: A Rotated Object Detection Benchmark using PyTorch. In Proceedings of the 30th ACM International Conference on Multimedia, Lisboa, Portugal, 10–14 October 2022; pp. 7331–7334. [Google Scholar]
  57. Deng, J.; Dong, W.; Socher, R.; Li, L.-J.; Li, K.; Fei-Fei, L. Imagenet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE conference on computer vision and pattern recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
  58. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-cam: Visual explanations from deep networks via gradient-based localization. In Proceedings of the IEEE international conference on computer vision, Las Vegas, NV, USA, 27–30 June 2016; pp. 618–626. [Google Scholar]
  59. Yang, X.; Zhao, J.; Wei, Z.; Wang, N.; Gao, X. SAR-to-optical image translation based on improved CGAN. Pattern Recognit. 2022, 121, 108208. [Google Scholar] [CrossRef]
  60. Zhu, J.-Y.; Park, T.; Isola, P.; Efros, A.A. Unpaired image-to-image translation using cycle-consistent adversarial networks. In Proceedings of the IEEE international conference on computer vision, Venice, Italy, 22–29 October 2017; pp. 2223–2232. [Google Scholar]
  61. Yang, J.; Zhou, K.; Li, Y.; Liu, Z. Generalized out-of-distribution detection: A survey. arXiv 2021, arXiv:2110.11334. [Google Scholar]
Figure 1. Three data generation methods around MSTAR. (a) Some sample pictures based on GANs; (b) Some sample pictures from SAMPLE [25] based on CAD 3D modeling and electromagnetic calculation simulation; (c) A sample picture based on background transfer.
Figure 1. Three data generation methods around MSTAR. (a) Some sample pictures based on GANs; (b) Some sample pictures from SAMPLE [25] based on CAD 3D modeling and electromagnetic calculation simulation; (c) A sample picture based on background transfer.
Remotesensing 15 04558 g001
Figure 2. The pipeline for constructing the synthetic dataset.
Figure 2. The pipeline for constructing the synthetic dataset.
Remotesensing 15 04558 g002
Figure 3. Vehicle segmentation label, containing a mask of the vehicle and its shadow and a rotated bounding box of its visually salient part. (a) The label of the vehicle when the boundary is relatively clear; (b) the label of the vehicle when the boundary is blurred.
Figure 3. Vehicle segmentation label, containing a mask of the vehicle and its shadow and a rotated bounding box of its visually salient part. (a) The label of the vehicle when the boundary is relatively clear; (b) the label of the vehicle when the boundary is blurred.
Remotesensing 15 04558 g003
Figure 4. (a) The pipeline for extracting grass and calculating the cosine similarity; (b) The histogram of the grass in Chips and Clutters.
Figure 4. (a) The pipeline for extracting grass and calculating the cosine similarity; (b) The histogram of the grass in Chips and Clutters.
Remotesensing 15 04558 g004
Figure 5. (a) A picture from the test set with 10346 × 1784 pixels: (b) Densely ranked vehicles; (c) sparsely ranked vehicles; (d) town scene; (e) field scene.
Figure 5. (a) A picture from the test set with 10346 × 1784 pixels: (b) Densely ranked vehicles; (c) sparsely ranked vehicles; (d) town scene; (e) field scene.
Remotesensing 15 04558 g005
Figure 6. Data statistics for Mix MSTAR (a) The area distribution of different categories of vehicles; (b) histogram of the number of annotated instances per image; (c) The number of vehicles in different azimuths; (d) The length–width distribution and aspect ratio distribution of vehicles.
Figure 6. Data statistics for Mix MSTAR (a) The area distribution of different categories of vehicles; (b) histogram of the number of annotated instances per image; (c) The number of vehicles in different azimuths; (d) The length–width distribution and aspect ratio distribution of vehicles.
Remotesensing 15 04558 g006
Figure 7. The architecture of the Rotated Retinanet.
Figure 7. The architecture of the Rotated Retinanet.
Remotesensing 15 04558 g007
Figure 8. The architecture of S2A-Net.
Figure 8. The architecture of S2A-Net.
Remotesensing 15 04558 g008
Figure 9. The architecture of R3Det.
Figure 9. The architecture of R3Det.
Remotesensing 15 04558 g009
Figure 10. The architecture of the ROI Transformer.
Figure 10. The architecture of the ROI Transformer.
Remotesensing 15 04558 g010
Figure 11. The architecture of Oriented RCNN.
Figure 11. The architecture of Oriented RCNN.
Remotesensing 15 04558 g011
Figure 12. The architecture of the Gliding Vertex.
Figure 12. The architecture of the Gliding Vertex.
Remotesensing 15 04558 g012
Figure 13. The architecture of ReDet.
Figure 13. The architecture of ReDet.
Remotesensing 15 04558 g013
Figure 14. The architecture of Rotated FCOS.
Figure 14. The architecture of Rotated FCOS.
Remotesensing 15 04558 g014
Figure 15. The architecture of Oriented RepPoints.
Figure 15. The architecture of Oriented RepPoints.
Remotesensing 15 04558 g015
Figure 16. (a) Confusion matrix of Oriented RepPoints on Mix MSTAR; (b) P-R curves of models on Mix MSTAR.
Figure 16. (a) Confusion matrix of Oriented RepPoints on Mix MSTAR; (b) P-R curves of models on Mix MSTAR.
Remotesensing 15 04558 g016
Figure 17. Some detection results of different models on Mix MSTAR. (a) Ground truth; (b) result of S2A-Net; (c) result of ROI Transformer; (d) result of Oriented RepPoints.
Figure 17. Some detection results of different models on Mix MSTAR. (a) Ground truth; (b) result of S2A-Net; (c) result of ROI Transformer; (d) result of Oriented RepPoints.
Remotesensing 15 04558 g017
Figure 18. (a) The result of the ROI Transformer on concatenated Chips; (b) class activation map of concatenated Chips.
Figure 18. (a) The result of the ROI Transformer on concatenated Chips; (b) class activation map of concatenated Chips.
Remotesensing 15 04558 g018
Figure 19. (a,c) T72 A05 Chips; (e,g) T72 A07 Chips; (b,d) class activation map of T72 A05 Chips; (f,h) class activation map of T72 A07 Chips.
Figure 19. (a,c) T72 A05 Chips; (e,g) T72 A07 Chips; (b,d) class activation map of T72 A05 Chips; (f,h) class activation map of T72 A07 Chips.
Remotesensing 15 04558 g019
Figure 20. The loss of pretrained/unpretrained models during training on Mini SAR. (a) Rotated Retinanet; (b) S2A-Net; (c) R3Det; (d) ROI Transformer; (e) Oriented RCNN; (f) ReDet; (g) Gliding Vertex; (h) Rotated FCOS; (i) Oriented RepPoints.
Figure 20. The loss of pretrained/unpretrained models during training on Mini SAR. (a) Rotated Retinanet; (b) S2A-Net; (c) R3Det; (d) ROI Transformer; (e) Oriented RCNN; (f) ReDet; (g) Gliding Vertex; (h) Rotated FCOS; (i) Oriented RepPoints.
Remotesensing 15 04558 g020
Figure 21. The mAP of pretrained/unpretrained models during training on Mini SAR. (a) Rotated Retinanet; (b) S2A-Net; (c) R3Det; (d) ROI Transformer; (e) Oriented RCNN; (f) ReDet; (g) Gliding Vertex; (h) Rotated FCOS; (i) Oriented RepPoints.
Figure 21. The mAP of pretrained/unpretrained models during training on Mini SAR. (a) Rotated Retinanet; (b) S2A-Net; (c) R3Det; (d) ROI Transformer; (e) Oriented RCNN; (f) ReDet; (g) Gliding Vertex; (h) Rotated FCOS; (i) Oriented RepPoints.
Remotesensing 15 04558 g021aRemotesensing 15 04558 g021b
Figure 22. Some detection results of Rotated Retinanet on Mini SAR. (a) Ground truth; (b) Rotated Retinanet trained on Mini SAR only; (c) Rotated Retinanet pretrained on Mix MSTAR; (d) Rotated Retinanet trained on Mini SAR and Mix MSTAR.
Figure 22. Some detection results of Rotated Retinanet on Mini SAR. (a) Ground truth; (b) Rotated Retinanet trained on Mini SAR only; (c) Rotated Retinanet pretrained on Mix MSTAR; (d) Rotated Retinanet trained on Mini SAR and Mix MSTAR.
Remotesensing 15 04558 g022aRemotesensing 15 04558 g022b
Figure 23. The style transfer of optical and SAR by using CycleGAN. (a) An optical car image with a label from the DOTA domain; (b) a transferred image on the Mix MSTAR domain.
Figure 23. The style transfer of optical and SAR by using CycleGAN. (a) An optical car image with a label from the DOTA domain; (b) a transferred image on the Mix MSTAR domain.
Remotesensing 15 04558 g023
Figure 24. Detection result of Redet on FARAD KA BAND. (a) Ground truth; (b) result.
Figure 24. Detection result of Redet on FARAD KA BAND. (a) Ground truth; (b) result.
Remotesensing 15 04558 g024
Table 1. Detailed information on existing public SAR vehicle datasets with large scenes.
Table 1. Detailed information on existing public SAR vehicle datasets with large scenes.
DatasetsResolution (m)Image Size (pixel)Images (n)Vehicle QuantityNoise Interference
FARAD X BAND [13]0.1016 × 0.10161682 × 3334–
5736 × 4028
30Large
FARAD KA BAND [14]0.1016 × 0.1016436 × 1288–
1624 × 4080
175Large
Spotlight SAR [15]0.1000 × 0.10003000 × 175464Small×
Mini SAR [16]0.1016 × 0.10162510 × 1638–
2510 × 3274
20Large×
MSTAR [10]Clutters0.3047 × 0.30471472 × 1784–
1478 × 1784
1000×
Chips0.3047 × 0.3047128 × 128–
192 × 193
17658Large×
Table 2. Basic radar parameters of Chips and Clutters in MSTAR.
Table 2. Basic radar parameters of Chips and Clutters in MSTAR.
Collection ParametersChipsClutters
Center Frequency9.60 GHz9.60 GHz
Bandwidth0.591 GHz0.591 GHz
PolarizationHHHH
Depression15°15°
Resolution (m)0.3047 × 0.30470.3047 × 0.3047
Pixel Spacing (m)0.202148 × 0.2031250.202148 × 0.203125
Platformairborneairborne
Radar Modespot lightstrip map
Data typefloat32uint16
Table 3. Analysis of grassland data from Chips and Clutters in the same period.
Table 3. Analysis of grassland data from Chips and Clutters in the same period.
GrasslandCollection DateMeanStdCV (Std/Mean)CSIM
BMP2 SN95631 September 1995

2 September 1995
0.0493059620.0302801590.6141277400.99826
BMP2 SN95660.0469899630.0283604450.6035426120.99966
BMP2 SN C210.0465607290.028306990.6079584790.99958
BTR70 SN C710.0468565230.0281432570.6006262350.99970
T72 SN1320.0459605050.0280471730.6102451010.99935
T72 SN8120.045461040.0275590570.6062126380.99911
T72 SNS70.0417912450.0253192190.6058498380.99260
Clutters5 September 199563.288125537.8502630.5980626331 1
1 The CSIM of Clutters is equal to 1, which means to compare itself.
Table 4. Original imaging algorithm and improved imaging algorithm.
Table 4. Original imaging algorithm and improved imaging algorithm.
Original Imaging AlgorithmImproved Imaging Algorithm
Input:
Amplitude in MSTAR Data a > 0,
Whether to enhance enhance = T or F
Input:
Amplitude in MSTAR Data a > 0,
Threshold thresh
Output: uint8 image imgOutput: uint8 image img
1: fmin←min(a), fmax←max(a)1: for pixel∈a do
2: frange←fmax-fmin, fscale←255/frange2:        if pixel>thresh then
3: a←(a-fmin)/fscale3:              pixel←thresh
4: img←uint8(a)4: scale←255/thresh
5: if enhance then5: img←uint8(a/scale)
6:        hist8←hist(img)6: Return img
7:        maxPixelCountBin←index[max(hist8)]
8:        minPixelCountBin←index[min(hist8)]
9:        if minPixelCountBin>maxPixelCountBin then
10:             thresh←minPixelCountBin-maxPixelCountBin
11:             scaleAdj←255/thresh
12:             img←img * scaleAdj
13:        else
14:             img←img * 3
15:        img←uint8(img)
16: Return img
Table 5. The division of Mix MSTAR.
Table 5. The division of Mix MSTAR.
ClassTrainTestTotal
2S119282274
BMP2195392587
BRDM219282274
BTR6013659195
BTR7013759196
D719282274
T6219182273
T72 A0419282274
T72 A0519282274
T72 A0719282274
T72 A1019081271
T72 A3219282274
T72 A6219282274
T72 A6319282274
T72 A6419282274
T72 SN13213759196
T72 SN81213659195
T72 SNS713457191
ZIL13119282274
ZSU23419282274
Total356018325392
Table 6. Performance evaluation of models on Mix MSTAR.
Table 6. Performance evaluation of models on Mix MSTAR.
CategoryModelParams(M)FLOPs (G)FPSmAPPrecisionRecallF1-Score
One-stageRotated Retinanet 36.74218.1829.261.03 ± 0.7530.9889.3646.01
Refine-stageS2A-Net38.85198.1226.372.41 ± 0.1031.5795.7447.48
R3Det37.52235.1926.170.87 ± 0.3122.2897.11 136.24
Two-stageROI Transformer55.39225.3225.375.17 ± 0.2446.9093.2762.42
Oriented RCNN41.37211.4426.573.72 ± 0.4538.2493.5654.29
ReDet31.754.4818.470.27 ± 0.7545.8389.9960.73
Gliding Vertex41.37211.3128.571.81 ± 0.1944.1791.7859.64
Anchor-freeRotated FCOS32.16207.1629.772.27 ± 1.2727.5296.4742.82
Oriented RepPoints36.83194.3526.875.37 ± 0.8034.7395.2550.90
1 The bold format represents the best indicator, and the following tables are the same.
Table 7. AP50 of each category on Mix MSTAR.
Table 7. AP50 of each category on Mix MSTAR.
ClassRotated Retinanet S2A-NetR3DetROI
Transformer
Oriented RCNNReDetGliding VertexRotated FCOSOriented RepPointsMean
2S187.9598.0295.1699.4897.5295.4895.3897.2298.1696.0
BMP288.1590.6990.6290.8290.7390.6790.6590.6690.8090.4
BRDM290.8699.6598.8399.6299.0398.1498.3999.7299.2298.2
BTR6071.8688.5288.0788.0285.6788.8486.1886.5586.5485.6
BTR7089.0398.0695.3697.5797.6892.5397.0296.6795.0695.4
D789.7690.7593.3898.0295.7093.4295.5195.5296.2194.3
T6278.4688.6691.2090.2992.3986.5389.7089.9990.2088.6
T72 A0437.4356.7150.2355.9755.4250.4450.0150.4653.4051.1
T72 A0531.0940.7143.1046.1748.2744.5645.9342.0950.0443.5
T72 A0729.5040.2840.1337.1338.2233.4038.4933.6144.3737.2
T72 A1027.9939.8236.0040.7136.8134.0434.4440.6747.5737.6
T72 A3269.2479.9683.0582.5780.7777.4877.0274.5678.6578.1
T72 A6241.0649.4950.0554.3247.3141.9745.7153.7754.0048.6
T72 A6338.1051.0746.4553.6350.0643.7949.2749.4453.0548.3
T72 A6435.5158.2857.5467.3766.6557.9563.3858.4766.1459.0
T72 SN13234.1854.9545.7159.8558.1656.8052.2354.3565.3853.5
T72 SN81249.2772.0161.8677.4274.2365.1371.3473.3372.1468.5
T72 SNS743.6159.3756.7766.4365.7057.4364.5664.4367.6260.7
ZIL13196.0696.2497.7699.0097.8898.7896.1695.2499.5897.4
ZSU23491.5595.0396.1799.0996.1598.0494.9098.6599.2696.5
mAP61.0372.4170.8775.1773.7270.2771.8172.2775.3771.4
Table 8. Best mAP of pretrained/unpretrained models on the Mini SAR validation set.
Table 8. Best mAP of pretrained/unpretrained models on the Mini SAR validation set.
ModelUnpretrainedPretrained
Rotated Retinanet 38.00 ± 15.5271.40 ± 0.75
S2A-Net65.63 ± 1.9469.81 ± 0.89
R3Det66.30 ± 2.6670.35 ± 0.18
ROI Transformer79.42 ± 0.6180.12 ± 0.01
Oriented RCNN70.49 ± 0.4780.07 ± 0.24
ReDet79.47 ± 0.5879.64 ± 0.31
Gliding Vertex70.71 ± 0.2077.64 ± 0.49
Rotated FCOS10.82 ± 3.9474.93 ± 2.60
Oriented RepPoints72.72 ± 2.0479.02 ± 0.39
Table 9. mAP of pretrained/unpretrained/mixed trained models on Mini SAR.
Table 9. mAP of pretrained/unpretrained/mixed trained models on Mini SAR.
ModelTrained on Mini SAR OnlyPretrained on Mix MSTARAdd Mix MSTAR
Rotated
Retinanet
38.00 ± 15.5271.40 ± 0.7578.62 ± 0.42
Rotated FCOS10.82 ± 3.9474.93 ± 2.6077.70 ± 0.10
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liu, Z.; Luo, S.; Wang, Y. Mix MSTAR: A Synthetic Benchmark Dataset for Multi-Class Rotation Vehicle Detection in Large-Scale SAR Images. Remote Sens. 2023, 15, 4558. https://doi.org/10.3390/rs15184558

AMA Style

Liu Z, Luo S, Wang Y. Mix MSTAR: A Synthetic Benchmark Dataset for Multi-Class Rotation Vehicle Detection in Large-Scale SAR Images. Remote Sensing. 2023; 15(18):4558. https://doi.org/10.3390/rs15184558

Chicago/Turabian Style

Liu, Zhigang, Shengjie Luo, and Yiting Wang. 2023. "Mix MSTAR: A Synthetic Benchmark Dataset for Multi-Class Rotation Vehicle Detection in Large-Scale SAR Images" Remote Sensing 15, no. 18: 4558. https://doi.org/10.3390/rs15184558

APA Style

Liu, Z., Luo, S., & Wang, Y. (2023). Mix MSTAR: A Synthetic Benchmark Dataset for Multi-Class Rotation Vehicle Detection in Large-Scale SAR Images. Remote Sensing, 15(18), 4558. https://doi.org/10.3390/rs15184558

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop