Next Article in Journal
Detecting Ocean Eddies with a Lightweight and Efficient Convolutional Network
Previous Article in Journal
Combination of Remote Sensing and Artificial Intelligence in Fruit Growing: Progress, Challenges, and Potential Applications
Previous Article in Special Issue
Underutilized Feature Extraction Methods for Burn Severity Mapping: A Comprehensive Evaluation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Technical Note

Region-Focusing Data Augmentation via Salient Region Activation and Bitplane Recombination for Target Detection

1
Department of Electronic Engineering, Tsinghua University, Beijing 100084, China
2
School of Mechatronical Engineering, Beijing Institute of Technology, Beijing 100081, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2024, 16(24), 4806; https://doi.org/10.3390/rs16244806
Submission received: 26 October 2024 / Revised: 17 December 2024 / Accepted: 20 December 2024 / Published: 23 December 2024

Abstract

:
As the performance of a convolutional neural network is logarithmically proportional to the amount of training data, data augmentation has attracted increasing attention in recent years. Although the current data augmentation methods are efficient because they force the network to learn multiple parts of a given training image through occlusion or re-editing, most of them can damage the internal structures of targets and ultimately affect the results of subsequent application tasks. To this end, region-focusing data augmentation via salient region activation and bitplane recombination for the target detection of optical satellite images is proposed in this paper to solve the problem of internal structure loss in data augmentation. More specifically, to boost the utilization of the positive regions and typical negative regions, a new surroundedness-based strategy for salient region activation is proposed, through which new samples with meaningful focusing regions can be generated. And to generate new samples of the focusing regions, a region-based strategy for bitplane recombination is also proposed, through which internal structures of the focusing regions can be reserved. Thus, a multiplied effect of data augmentation by the two strategies can be achieved. In addition, this is the first time that data augmentation has been examined from the perspective of meaningful focusing regions, rather than the whole sample image. Experiments on target detection with public datasets have demonstrated the effectiveness of this proposed method, especially for small targets.

1. Introduction

With the rise of Chat-GPT and its variants, data-driven large foundation models have attracted increasing attention. A recent study has shown that the performance of a convolutional neural network (CNN) is logarithmically proportional to the amount of training data [1]. But in most cases, the training data are difficult to collect or expensive to annotate, especially in the field of satellite remote sensing, such as the landslide data [2], disaster emergency data [3] and hyperspectral image [4,5,6]. It is generally acknowledged that, without sufficient annotated training data, over-parameterized CNNs would be over-fitted.
To solve the over-fitting problem, data augmentation has attracted increasing attention in recent years. According to the actual effect on training, data augmentation methods can be roughly categorized into two groups: positive data augmentation and negative data augmentation. The former is dedicated to increasing new augmented data through the finite or conscious distribution shift of the original data, and the latter is dedicated to increasing the robustness of the model through increasing data complexity. More specifically, in the group of positive data augmentation, geometric data augmentation methods, including cropping, flipping [7] and rotation [8], multiply the geometric variation in the sample images. Color jitter [9] changes the contrast, saturation and brightness of the sample images. Additionally, the bitplane recombination method [10] gives an idea that there are certain differences between different bitplanes, and the recombination of them can provide different structural information at different levels. In addition, metalantis [11] enhances underwater images through metamergence, metalief and metaebb phases. Inconsistent knowledge distillation (IKD) [12] utilized sample-specific data augmentation to capture distinct frequency components and adversarial feature augmentation to extract the non-robust features. The above methods can really boost the utilization of the original data with finite distribution shift, but attention has not been paid to the influence of sub-regions in the sample images, especially those that may contain positive or typical negative regions. Apart from these, mixup [13] forces CNNs to learn convex combinations between different sample images to generate new augmented samples, maximizing the margin by pushing the decision boundary away from the original samples. Manifold mixup [14] mixes the features from the intermediate layers of CNN to generate new augmented samples with feature diversity. These methods have also boosted the utilization of the original data through conscious distribution shift, but the newly generated samples are usually unrealistic, unreliable and inconsistent with human perception.
On the other hand, in the category of negative data augmentation, GANs have been successfully introduced to generate new samples, including BAGAN [15], DiffAugment [16] and the APA method [17]. However, most of them may be unfeasible for remotely sensed images or medical images, as the newly generated samples are unrealistic and unreliable. In addition, Cutout [18] randomly replaces a certain part of a sample image with a square patch of constant value, to prevent CNN from over-fitting. Random erasing [19] pushes CNN to learn occlusion by randomly determining whether to mask a region or not. Hide-and-Seek randomly hides image patches to force CNN to learn multiple parts of the original image [20]. GridMask [21] masks multiple regions in evenly spaced grids to maintain a good balance between the deletion and retention of critical information. CutMix [7] cuts a certain region in the sample image and replaces it with others, to differentiate two classes within a single image. RICAP [22] crops and patches four sample images to generate new samples. Even though these methods concentrate on forcing CNN to learn multiple parts of the original image by occlusion or re-editing operations, they actually damage the internal structures of a given sample image, and ultimately affect the results of subsequent application tasks. Taking the target detection in remotely sensed images as an example, a target may only consist of a few pixels, while there are thousands of small targets in one image. Most of these methods are inapplicable for this situation, as they may completely occlude or crop off such small targets. Furthermore, most of these methods concentrated only on the whole sample image, rather than the meaningful sub-regions inside the sample image.
To this end, this paper proposes Region-fOcusing data augmentation via salient region activation and BITplane recombination for target detection, termed ROBIT, to solve the problem of internal structure loss in the current data augmentation methods. More specifically, to boost the utilization of the positive regions and typical negative regions, a new surroundedness-based strategy for the salient region activation is proposed. And to generate new samples of the focusing regions, a region-based strategy for the bitplane recombination is proposed. Experiments on target detection tasks demonstrated the effectiveness of this method. The novelties and contributions of the proposed method can be summarized as follows:
  • To the best of our knowledge, this is the first time that data augmentation has been examined from the perspective of meaningful focusing regions, rather than the whole sample image, and both positive regions and typical negative regions have been considered at the same time.
  • A region-based strategy for bitplane recombination is proposed, which can maintain the internal structures of the focusing regions. And through a combination with the region-focusing strategy, a multiplied rate of data augmentation can be achieved.

2. Related Works

2.1. Saliency Detection

In recent years, convolutional neural networks (CNNs) have succeeded in saliency detection; they are generally utilized to construct encoders and decoders for feature extraction and generate visual saliency maps [23]. Subsequently, Long-Short Term Memory (LSTM) networks were also applied to capture both local and long-range visual features, to boost the accuracy of saliency detection [24]. Recently, by learning spatial long-range dependencies, transformer-based networks have achieved significant improvements in saliency detection [25]. However, most of these methods are designed for singular visual contexts, which cannot be further adapted to different tasks [26].

2.2. Bitplane Techniques

On the other hand, to exploit the internal structural information of the sample image, we should pay attention to the bitplane techniques. Khan et al. combined fuzzy logic with bitplane to locate and segment the regions of interest (RoI) in Computed Tomography (CT) images [27]. Dubey et al. adopted the local bitplane decoded pattern (LBDP) to retrieve CT images [28]. Tuan et al. utilized the adaptive fast marching and bitplane to segment the brain for the Magnetic Resonance Images (MRI) [29]. To date, bitplane techniques have attracted attention mostly from medical image processing, rather than from deep-learning-related tasks, especially data augmentation or target detection tasks.

3. Methods

This proposed region-focusing data augmentation method mainly consists of the following three parts: empirical risk of data augmentation, region-focusing based on salient region activation and region-based data augmentation.

3.1. Empirical Risk of Data Augmentation

Let X and Y denote the data space and label space, respectively, and each sample x with label y can be denoted by ( x , y ) , and a joint distribution of these data can be denoted by D ( x , y ) . The learning task aims to find a function f : x y , which minimizes the pre-defined loss l ( f x , y ) over the distribution D , and this is known as the expected risk R ( f | D ) defined by
R f D = l ( f x , y ) d   D ( x , y )
However, the joint distribution of training data is usually unknown in practice. Thus, a general solution is empirical risk minimization (ERM) [30], which optimizes the empirical risk of the training dataset { ( x i , y i ) } i = 1 N to mimic the data distribution:
R ^ f D = 1 N i = 1 N l ( f x i , y i )
Here, from Equation (2), we can see that the expected risk R ^ f D increases with the amount of training data. When conducting data augmentation, new samples { ( x i ~ , y i ~ ) } i = 1 N can be generated by transformation, occlusion or re-editing on { ( x i , y i ) } . Since most data augmentation methods will not change the annotated labels, newly generated samples { ( x i ~ , y i ~ ) } will share the same labels with { ( x i , y i ) } , i.e., y i ~ = y i . Data augmentation converts the distribution D to a new distribution D , and the learning task is changed from minimizing the expected risk R ^ f D to minimizing R ^ f D :
R ^ f D = 1 N i = 1 N l ( f x i ~ , y i ~ )
where N denotes the total scale of new augmented dataset, N N . According to the Vapnik–Chervonenkis Dimension theory (VC theory) [30], given a binary classifier f F with the finite VC-Dimension | F | V C , the upper bound of the expected risk can be formulated as follows with the probability 1 δ :
R f D R ^ f D + O ( ( | F | V C l o g δ N ) α )
where 1 2 α 1 , α = 1 2 is the non-separable case and α = 1 is the separable case [31]. Similarly, the model trained on the new augmented dataset can be expressed as
R f D R ^ f D + O ( ( | F | V C l o g δ N ) α )
Although the generalization error O in Equation (5) is smaller than that in Equation (4), there is discrepancy between the risk terms due to the distribution. Assuming D D , it can be expressed as
R f D R f D = ϵ 1 0
Thus, Equation (5) can be reformulated as
R f D R ^ f D + ϵ 1 + O ( ( | F | V C l o g δ N ) α )
From Equation (7), it can be seen that the benefits of data augmentation can be attributed to (1) the distribution shift, i.e., ϵ 1 , being fractional; (2) the scale of new augmented dataset being large, i.e., N N .
Actually, there is a certain trade-off between the amount of new augmented data and the distribution shift. And the distribution shift between the original dataset and new augmented dataset can be expressed as
R ^ f D R ^ f D = ϵ 2 0
Combined with Equations (2) and (3), we have
ϵ 2 = 1 N i = 1 N l ( f x i , y i ) 1 N i = 1 N l ( f x i ~ , y i ~ )
On the other hand, there are a series of data augmentation methods that generate new augmented images by occlusion or re-editing of the original images, such as Cutout [18], RE [19], HaS [20], etc. In these augmentation methods, the relationship between the augmented image and the original training image is one-to-one, i.e., each augmented image could find a corresponding original training image. Therefore, in these augmentation methods, N could be replaced by N :
ϵ 2 = 1 N i = 1 N l f x i , y i l ( f x i ~ , y i )
Then, the objective function can be applied with Taylor expansion:
l f x i ~ , y i l f x i , y i + l f x i ~ , y i T ( x i ~ x i )
where denotes differential operation and T denotes transpose operation. Combined with Equation (10), we have
ϵ 2 = 1 N i = 1 N l f x i ~ , y i T ( x i ~ x i )
where the term ( x i ~ x i ) corresponds to the internal structure loss in the data augmentation methods, leading to the distribution shift. Meanwhile for the latest data augmentation methods, like BIRD [10], as the transformation is conducted on certain bitplanes, the continuity of the internal structure of the training images can be maintained without changing the distribution of the original dataset.
Inspired by these methods, here, a region-focusing data augmentation method via salient region activation and bitplane recombination is proposed. More specifically, to boost the utilization of the positive regions and typical negative regions, a new surroundedness-based strategy for salient region activation is proposed. And to generate new samples of the focusing regions, a region-based strategy for bitplane recombination is proposed, without changing the distribution of the original dataset.

3.2. Region-Focusing Based on Salient Region Activation

Most of the current data augmentation methods are based on occlusion or re-editing operations; even if they concentrate on occlusion robust or forcing the CNN to learn multiple parts of the original sample image, these methods still damage the continuity of internal structures of a given sample image. As a result, most of the current data augmentation methods may be inapplicable if we want to take the detection of small targets shown as in Figure 1a as our final application task. Intuitively, the occlusion or re-editing operations in those methods should be more conservative. Based on the above considerations, here, a new surroundedness-based strategy is proposed to focus the data augmentation operations only on some meaningful sub-regions rather than the whole sample image. In more detail, the meaningful sub-regions are extracted by the subsequent salient region activation, and then bitplane recombination is conducted for the extracted sub-regions.
It should be mentioned that, as the meaningful sub-regions may or may not contain the annotated target, those sub-regions containing the target can be taken as positive regions, while those sub-regions without the target can be taken as typical negative regions, as shown in Figure 2. It can be seen that, if only positive regions are selected to be augmented, the network would come across the long-tail problem. In contrast, our proposed surroundedness-based strategy with both positive and negative regions could alleviate this problem effectively, because both positive regions and typical negative regions can contribute to the network training, through the following data augmentation operation. A comparison between the use of all meaningful focusing regions and only the positive regions will be given and discussed in the Experiment section.
To begin with, a single channel of a C -channels image can be taken as an 8-bit grayscale image, P c R H × W × 1 . Our proposed salient region activation method is to select a meaningful focusing region P e . To select meaningful focusing regions, a morphological-based approach, i.e., surroundedness, is introduced, which has demonstrated its effectiveness on sensing figure-ground assignments in neuroscience findings [32]. The surroundedness refers to the enclosure topological relationship between diverse visual components, which is invariant in different scenarios, independent of the shape or scale of visual components. Generally, to analyze surroundedness, a given image P o r i can be characterized by a set of binary images { B r } , which are sampled from a pre-given distribution function. Then, to depict the effects of each binary image B r , which can be termed as salient map Ψ ( B r ) , surroundedness analysis and normalization are conducted on each binary image. Finally, to acquire meaningful focusing regions, the salient maps { Ψ ( B r ) } are averaged and thresholded, as shown in Figure 1.
Concretely, a set of binary images { B r } can be generated by randomly thresholding the feature maps of the given sample image P o r i , through the pre-defined threshold θ c of each channel of the feature maps:
p i , j = Γ ( P c , θ c ) ,   θ c ~ d θ
where each individual channel image P c is selected as a feature map; p i , j represents the value of pixel ( i , j ) on the binary image, i.e., pixel value p i , j = 1 if P c , i , j > θ c and p i , j = 0 otherwise; d θ represents the prior distribution of threshold sampling on the feature map; and θ c follows the distribution of d θ .
To further simplify the sampling of a multi-channel image P o r i , the color space can be whitened before the binary image sampling. Then, the threshold θ c can be sampled at a fixed step δ , which is set to 8 in the experiments. In the limit, this fixed-step sampling is equivalent to uniform sampling [32].
To acquire the salient map Ψ ( B r ) , a binary image B r can be found firstly by analyzing the surrounding areas to acquire the activation map A ( B r ) , and then by normalizing the activation map to emphasize regions of small targets. Specifically, the pixels on a binary image B r can be divided into the figure set B r , f : { p i , j = 1 } and the ground set B r , g : { p i , j = 0 } . Intuitively, a figure pixel is surrounded if it is enclosed by the ground set. And the surroundedness can be defined by the connectivity of a pixel to the image boundary pixels. On a binary image B r , a pixel p i , j is surrounded if there is no path in B r , f or B r , g that connects p i , j and any pixel of the image borders, where a path denotes a sequence of pixels in which any pair of consecutive pixels are adjacent. Therefore, by using the pixels on the image boundary as seeds, non-surrounded pixels can be filtered out by the Flood Fill algorithm [32] efficiently. Thus, on the activation map A ( B r ) , the surrounded pixels are set to 1 and the others are set to 0. Then, to normalize the activation map A ( B r ) , each activation map is divided into two sub-activation maps:
A + B r = A ( B r ) B r , A B r = A ( B r ) ¬ B r
where is a pixel-wise Boolean conjunction operation and ¬ is a negation sign. A + B r and A B r denote the selected surrounded regions on B r and ¬ B r , respectively. The intuitive explanation of these sub-activation maps is that A + B r activates the surrounded regions greater than the threshold, and A B r activates the surrounded regions less than the threshold. Therefore, normalization can be taken to emphasize the rare topographic features. To further penalize sub-activation maps for small targets, the sub-activation maps are dilated before normalization:
A + B r = A + B r K , A B r = A B r K
where is dilate operation and K is a square dilation kernel of width w . After that, to emphasize small active regions, l 2 -normalization is applied to the sub-activation maps:
Ψ B r = A + B r | | A + B r | | 2 + A B r | | A B r | | 2
Finally, to acquire an averaged salient map Ψ B r ¯ , all salient maps are averaged and small targets are emphasized, as shown in Figure 2.
The overall surroundedness-based strategy for salient region activation is simple, efficient and training-free. As a result, the center point v e = ( i e , j e ) and its size ( h e , w e ) of each meaningful focusing region P e can be acquired as in Figure 1d, for the next part: the region-based data augmentation.

3.3. Region-Based Data Augmentation

After acquiring the meaningful focusing regions, to generate new samples of the focusing regions, a region-based strategy of bitplane recombination is conducted by which the internal structures of focusing regions can be reserved.
Here, firstly, to utilize the symbolic system, the basic Algorithm 1 of bitplane recombination will be briefly introduced. Given a typical 8-bit C -channel image P o r i R H × W × C , it will be firstly split into C -separated channels P c :
P o r i = P 1 P 2 P 3 P c
where P c R H × W × 1 , each channel can be taken as an 8-bit grayscale image, and each grayscale image can be sliced into 8 bitplanes hierarchically, from the 0 t h bitplane to the 7 t h bitplane as follows:
P c = m = 0 7 i = 1 H j = 1 W b i , j m 2 m { P c m } m = [ 0,7 ]
where b i , j m represents the bits located at pixel ( i , j ) on the m t h bitplane (marked as m t h BP). Each bitplane P c m R H × W × 1 reserves its own bits, the m t h BP, taking no account of the bits on other bitplanes. For an 8-bit grayscale image, bitplanes from the 0 t h BP to the 7 t h BP provide internal structure information from the details to the entirety.
After bitplane slicing, to take advantage of the bitplane properties in region-based data augmentation, certain numbers of bitplanes will be extracted to maintain the internal structures of meaningful focusing regions. In detail, the extracted bitplanes can be formulated as
P s = c = 1 C s = l 1 l k i = 1 H j = 1 W b c , i , j s 2 s ,   s = { l 1 , , l k }
where l 1 , , l k denotes the order of the extracted bitplanes, i.e., l k t h BP.
The appropriate number of extracted bitplanes depends on the difference in extracted bitplanes over the recombined new image. The recombined new image P n e w l 1 , , l k generated from the extracted bitplanes is utilized to estimate the original image by using the maximum a posteriori (MAP) theory [33], and the error probability between the original image and the estimated image, i.e., P o r i and P ~ o r i , is used to evaluate the difference in the number of extracted bitplanes as follows:
P ~ o r i a r g m a x log p ( P o r i , P n e w l 1 , , l k )
For a given 8-bit image, the averaged error probability is large when 1 or 2 bitplanes are extracted. However, when 3, 4, 5, 6 and 7 bitplanes are extracted, the averaged error probability is acceptable. Thus, the suitable number of extracted bitplanes is in the range of 3 BPs~7 BPs.
Since the recombined images P n e w l 1 , , l k can be regarded as the reconstructed images of the original image P o r i , Mean Squared Error (MSE), Peak Signal-to-Noise Ratio (PSNR) and Structural Similarity Index Measure (SSIM) indicators are selected to measure the recombination quality. As a result, the MSE of P n e w 0,1 , 2,3 , 4,5 , 6 and P n e w 0,1 , 2,3 , 4,5 , 7 is quite large. In addition, the PSNR of P n e w 0,1 , 2,3 , 4,5 , 6 and P n e w 0,1 , 2,3 , 4,5 , 7 is so fractional that it is better to reserve both the 7 t h and the 6 t h bitplane. Furthermore, the SSIM indicates that reserving both the 7 t h and the 6 t h bitplane contributes to maintaining the major internal structures.
Therefore, from the view of recombination quality, bitplane extraction (BE) operations preserving both the 7 t h and the 6 t h bitplane can be adopted to maintain the internal structural information. In addition, adjacent bitplane extraction [10] is selected as the basic strategy for the BE operation.
For a focused region on a single channel image P c , the bits within the selected focusing region P e on the extracted bitplanes are reserved according to Equation (9):
P e = s = l 1 l k i = i e i e + h e j = j e j e + w e b i , j s 2 s
For the C -channels image P o r i R H × W × C , there would be 4 representative sub-policies { Ω i } i = [ 1,4 ] for the region-based data augmentation, taking C = 3 as an example:
  • The meaningful focusing regions are all selected, and the extracted bitplanes of different regions are the same order bitplanes, as shown in Figure 3c:
    Ω 1 : P e 1 = P e 2 = P e 3 ,   s 1 = s 2 = { s 3 }
  • The meaningful focusing regions are partially selected, but the extracted bitplanes of different regions are the same order bitplanes, as shown in Figure 3d:
    Ω 2 : P e 1 P e 2 P e 3 ,   s 1 = s 2 = { s 3 }
It should be mentioned that, for a given training image, the partially selected focusing regions change by a pre-given percentage in each training epoch.
3.
The meaningful focusing regions are all selected, but the extracted bitplanes of different regions are different order bitplanes, as shown in Figure 3e:
Ω 3 : P e 1 = P e 2 = P e 3 ,   s 1 s 2 { s 3 }
Here, it should also be mentioned that the different order bitplanes is not the different individual bitplanes contained in a bitplane extraction operation, but the different bitplane extraction operations. Taking the 6-times 5-BPs as an example, whose adjacent bitplane extraction operations include P n e w 3,4 , 5,6 , 7 , P n e w 0,4 , 5,6 , 7 , P n e w 2,3 , 5,6 , 7 , P n e w 1,2 , 5,6 , 7 , P n e w 0,1 , 5,6 , 7 and P n e w 0,1 , 2,6 , 7 , here, different order bitplanes refer to s = { 3,4 , 5,6 , 7 } , s = { 0,4 , 5,6 , 7 } or s = { 2,3 , 5,6 , 7 } , etc.
4.
The meaningful focusing regions are partially selected, and the extracted bitplanes of different regions are different order bitplanes, as shown in Figure 3f:
Ω 4 : P e 1 P e 2 P e 3 ,   s 1 s 2 { s 3 }
By applying these region-focused bitplane recombination operations, the positive regions and typical negative regions within the meaningful focusing regions of the given sample images can be exploited effectively. In these region-focused bitplane recombination operations, sub-polices Ω 2 emphasize the spatial variation, while sub-polices Ω 3 emphasize the bitplane variation. And sub-polices Ω 4 can combine the variation in them both. The performances of these sub-policies are shown in Figure 3.

3.4. Total Framework of the Proposed Method

After salient region activation and region-based bitplane extraction, the extracted bitplanes within meaningful focusing regions will be recombined to generate a new sample of training image P n e w l 1 , , l k R H × W × C , as follows:
P n e w l 1 , , l k = ¬ { P e } P e 1 P e 2 P e n
It can be seen from Equation (26) that, in the meaningful focusing regions of the given sample image, the internal structures can be completely maintained, and the variety among different bitplanes can also be full expressed, through the recombination of extracted bitplanes.
We can see that, given a selected region P e in the sample image P o r i , compared with the occlusion- or re-editing-based data augmentation methods, our proposed method can reduce the amount of internal structure loss by pixels from P e down to P e × p × b r 2 r 255 , where r { l 1 , , l k } .
The overall framework of our proposed ROBIT method can be summarized as follows, as shown in Figure 4.
Algorithm 1. ROBIT data augmentation
Input: Training dataset P o r i = { P 1 , P 2 , , P N } .
Parameter setting: N b , number of image blocks of image P i ;
          p , the probability threshold of the data augmentation.
Output: Augmented set P n e w = { P n e w , 1 , , P n e w , N } .
For i = 1 to N do
  Crop P i into image blocks { P i , j } j = [ 1 , N b ] .
  For j = 1 to N b do
    For each epoch do
      While averaged focusing map is generated
        Extract focusing regions { P e } ;
        In each focusing region do
          Randomly generate the probability of data augmentation τ from ( 0,1 ) .
          If τ p do
            Region focusing based on salient region activation;
            Region-focused Bitplane extraction as in Equation (21);
            Bitplane recombination to generate new images as in Equation (26).
          End
        End
      End
    End
  End
End

4. Results

To evaluate the effectiveness of our proposed method, experiments were conducted on the object detection task using two different public datasets.

4.1. Dataset

For object detection task, our proposed method was evaluated on the public remotely sensed image dataset, the DOTA dataset. The DOTA dataset [8] is the most popular aerial remote sensing object detection dataset, containing 2806 images, 15 categories and 188,282 instances of different scales, orientations and shapes. This dataset is also representative in small target detection.
In addition, experiments are also conducted on the HRSC2016 dataset [34], which is designed for the ship detection of remotely sensed images. HRSC2016 is collected from Google Earth, which includes 1061 images ranging from 300 × 300 to 1500 × 900. This dataset contains 436 training images, 181 validation images and 444 test images.

4.2. Experimental Settings of Our Proposed Method

In all the following experiments, the data augmentation rate of the bitplane recombination is set to six times, which is defined by the scale of the new generated dataset relative to that of the original dataset. In addition, the percentage of the region selection in region-focusing is set to 0.5 and the training epoch is set to 12 for our proposed method. Since the final data augmentation rate of our proposed ROBIT method is equal to the number of selected focusing regions times the rate of region-based data augmentation, the final data augmentation rate using sub-policy Ω 1 is 1   ×   6 = 6 times, the final rate using sub-policy Ω 2 is 0.5   ×   12   ×   6 = 36 times, the final rate using sub-policy Ω 3 is 1   ×   6 = 6 times and the final rate by using sub-policy Ω 4 is 0.5   ×   12   ×   6 = 36 times, respectively.

4.3. Comparison with Different Object Detection Methods

In this section, the target detection results of the RoI Transformer (RT) [35] method with and without our proposed method are compared with other popular target detection methods on the DOTA dataset, to validate the effectiveness of our proposed method. The RT method is the official baseline method of the DOTA dataset, and the compared methods include RRPN [36], DCN [37], ICN [38] and FFA-3 [39], which are influential works in the target detection of remotely sensed images. In this experiment, according the settings of RT [35], 25% training images are augmented, the learning rate is set to 0.0005 and the moment is set to 0.9. And the mean average precision (mAP) on the DOTA server is utilized to evaluate the detection accuracy.
By using all the original and augmented images, the mAP of the 15 categories is shown in Table 1. From left to right, the classes are plane, baseball diamond, bridge, ground track field, small vehicle, large vehicle, ship, tennis court, basketball court, storage tank, soccer ball field, roundabout, harbor, swimming pool and helicopter. The best results are shown in bold and the second-best results are shown with an underline. As can be seen in Table 1, the mAP can be improved by 2.9%, compared with the latest Pytorch version of the RT method (RT-Pytorch). And our proposed method can even enhance the mAP by 7.1%, compared with the Mxnet version RT method (RT-Mxnet). It is also obvious that, while there are subtle discrepancies in the accuracy of different categories, our proposed method can boost most categories, especially the categories with small objects, such as the helicopter category (HC).

4.4. Comparison with Different Data Augmentation Methods

In this section, data augmentation methods for target detection are compared on the DOTA dataset, including data augmentation using our proposed method and those using traditional “Crop and Rotation (C&R)” [35], “Crop and Scaling (C&S)” and the latest BIRD method [10]. In addition, we also implemented a random digging-based BIRD data augmentation method (noted as rBIRD) as a comparison. To make a fair comparison, all the input images were cropped to the same size. According to the detection accuracy, the official baseline method of the DOTA dataset, i.e., RT-Pytorch, was chosen as the benchmark target detection method. And the detection accuracy was evaluated by the mAP as well. From Table 2, it can be seen that, compared with the baseline RT-Pytorch method, the mAP can be enhanced by 2.9% by using our proposed method. Moreover, our proposed ROBIT method also surpasses most of the other comparative methods and the random digging-based BIRD data augmentation method across different categories, especially the categories with small targets, such as the helicopter (HC) category.
Given a C -channel original image P o r i R H × W × C with the height and width of a meaningful focusing region P e denoted as h e and w e , the computational complexity of single-stream ROBIT data augmentation is O ( n × h e × w e ) ; here, n denotes the number of meaning focusing regions. Therefore, the computational complexity of a given training image O ( n × h e × w e ) is approximate to O ( H × W ) . And the computational complexity of the entire training dataset is O ( N × H × W ) , which is similar to that of the official data augmentation of the DOTA dataset, i.e., C&R and C&S.

4.5. Effectiveness of the Sub-Policies in Region-Based Operations

In this section, to investigate the effectiveness of the sub-policies of region-based data augmentation operations for the C -channel sample images, target detection experiments are carried out on the DOTA dataset. RT-Pytorch is chosen as the benchmark detection method as well, and the mAP is applied to verify the effectiveness of different sub-policies. The bitplane extraction operation is selected as 7-BPs.
From Table 3, it can be seen that both the sub-policy Ω 2 that partial regions selected with same order bitplanes extracted and the sub-policy Ω 3 that all regions selected with different order bitplanes extracted boost the target detection accuracy remarkably. What is more, the sub-policy that partial regions selected with different order bitplanes extracted, i.e., the Ω 4 sub-policy, can combine the advantages of them both and achieves the best performance.

4.6. Ablation Study

In this section, to validate the effectiveness of each part of our proposed method, ablation studies are conducted on the DOTA dataset. The official baseline method of the DOTA dataset, i.e., RT-Pytorch, is also chosen as the benchmark target detection method, and the effectiveness is evaluated by the mAP as well. Experiments results are shown in Table 4.
It can be seen that each part of our proposed data augmentation method can really boost the performance of target detection. And the combination of them gives full play to the advantages of the two individual parts, so the final mAP can be remarkably enhanced by 2.9% compared with the baseline RT-Pytorch method.

4.7. Meaningful Focusing Regions vs. Only Positive Regions

In this section, a comparison between the use of all the meaningful focusing regions and only the positive regions for data augmentation will be given. Object detection for typical small ship targets in optical remotely sensed images is selected as the comparison task, as in [43]. And to accurately depict the positive regions, HeatNet [44] is selected as the baseline detection method.
The detection results are shown in Table 5. From Table 5, it can be clearly seen that the result of using all the meaningful focusing regions surpasses that of only using positive regions, which validates the effectiveness of using of all the meaningful focusing regions for data augmentation operations.

4.8. Comparison on the HRSC2016 Dataset

In this section, our proposed ROBIT method is also validated on the HRSC2016 dataset [34]. And the ReDet method [45] is selected as the baseline detection method to verify the effectiveness of our proposed method across different deep learning models. The experimental results are shown in Table 6. We can see that the detection accuracy can be improved by 1.3% using our proposed ROBIT method.

5. Discussion

As mentioned in Section 3.4, given a C -channel original image P o r i R H × W × C with the height and width of a meaningful focusing region P e denoted as h e and w e , the computational complexity of single-stream ROBIT data augmentation is O ( n × h e × w e ) ; here n denotes the number of meaning-focusing regions. Therefore, the computational complexity of a given training image O ( n × h e × w e ) is approximate to O ( H × W ) . And the computational complexity of the entire training dataset is O ( N × H × W ) . Compared with the training of the target detection network, the runtime of our proposed data augmentation is negligible. When there are too many targets in the remote sensing image, such as if thousands of targets are included in one typical image, the computational complexity and runtime will increase.

6. Conclusions

To solve the problem of internal structure loss in data augmentation, this paper proposes a region-focusing data augmentation via salient region activation and bitplane recombination. More specifically, to boost the efficiency of positive regions and typical negative regions, a new strategy of surroundedness-based salient region activation is proposed, by which new samples with meaningful focusing regions can be generated. And to generate new samples of the focusing regions, a region-based strategy of bitplane recombination is conducted, by which the internal structures can be reserved. Thus, a multiplied effect could be acquired by the advantages of these two strategies. Finally, experiments on target detection have demonstrated the effectiveness of this proposed ROBIT method, especially for small targets. As a matter of fact, although we only used our proposed method for the task of target detection in multi-spectral images, it can also be used for other related tasks with different image sources, such as classification in multi-spectral images, spectral target detection in hyperspectral images and even lesion detection in medical images; these would be some of the future works we need to consider.

Author Contributions

Methodology, H.Z.; Investigation, X.H.; Writing—original draft, H.Z.; Writing—review and editing, W.S.; Supervision, W.S. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded in part by the National Natural Science Foundation (41971294, 82471999), the Beijing Institute of Technology Research Fund Program for Young Scholars, and the Cross-Media Intelligent Technology Project of BNRist (BNR2019TD01022) of China.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Sun, C.; Shrivastava, A.; Singh, S.; Gupta, A. Revisiting Unreasonable Effectiveness of Data in Deep Learning Era. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 843–852. [Google Scholar]
  2. Xie, Y.; Zhan, N.; Zhu, J.; Xu, B.; Chen, H.; Mao, W.; Luo, X.; Hu, Y. Landslide Extraction from Aerial Imagery Considering Context Association Characteristics. Int. J. Appl. Earth Obs. Geoinf. 2024, 131, 103950. [Google Scholar] [CrossRef]
  3. Zhu, J.; Zhang, J.; Chen, H.; Xie, Y.; Gu, H.; Lian, H. A Cross-View Intelligent Person Search Method Based on Multi-Feature Constraints. Int. J. Digit. Earth 2024, 17, 2346259. [Google Scholar] [CrossRef]
  4. Xu, Y.; Hou, J.; Zhu, X.; Wang, C.; Shi, H.; Wang, J.; Li, Y.; Ren, P. Hyperspectral Image Super-Resolution with ConvLSTM Skip-Connections. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5519016. [Google Scholar] [CrossRef]
  5. Cao, S.; Feng, D.; Liu, S.; Xu, W.; Chen, H.; Xie, Y.; Zhang, H.; Pirasteh, S.; Zhu, J. BEMRF-Net: Boundary Enhancement and Multiscale Refinement Fusion for Building Extraction from Remote Sensing Imagery. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2024, 17, 16342–16358. [Google Scholar] [CrossRef]
  6. Zhang, H.; Han, X.; Deng, J.; Sun, W. How to Evaluate and Remove the Weakened Bands in Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2024. [Google Scholar]
  7. Yun, S.; Han, D.; Oh, S.J.; Chun, S.; Choe, J.; Yoo, Y. Cutmix: Regularization Strategy to Train Strong Classifiers with Localizable Features. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 6023–6032. [Google Scholar]
  8. Xia, G.S.; Bai, X.; Ding, J.; Zhu, Z.; Belongie, S.; Luo, J.; Datcu, M.; Pelillo, M.; Zhang, L. DOTA: A Large-scale Dataset for Object Detection in Aerial Images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3974–3983. [Google Scholar]
  9. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  10. Zhang, H.; Xu, Z.; Han, X.; Sun, W. Data Augmentation using Bitplane Information Recombination Model. IEEE Trans. Image Process. 2022, 31, 3713–3725. [Google Scholar] [CrossRef]
  11. Wang, H.; Zhang, W.; Bai, L.; Ren, P. Metalantis: A Comprehensive Underwater Image Enhancement Framework. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5618319. [Google Scholar] [CrossRef]
  12. Liang, J.; Liang, S.; Liu, A.; Ma, K.; Li, J.; Cao, X. Exploring Inconsistent Knowledge Distillation for Object Detection with Data Augmentation. In Proceedings of the 31st ACM International Conference on Multimedia, Ottawa, ON, Canada, 29 October–3 November 2023. [Google Scholar]
  13. Zhang, H.; Cisse, M.; Dauphin, Y.N.; Lopez-Paz, D. mixup: Beyond Empirical Risk Minimization. Int. Conf. Learn. Represent. 2018. [Google Scholar]
  14. Verma, V.; Lamb, A.; Beckham, C.; Najafi, A.; Mitliagkas, I.; Lopez-Paz, D.; Bengio, Y. Manifold mixup: Better Representations by Interpolating Hidden States. In Proceedings of the International Conference on Machine Learning, PMLR, Long Beach, CA, USA, 9–15 June 2019; pp. 6438–6447. [Google Scholar]
  15. Mariani, G.; Scheidegger, F.; Istrate, R.; Bekas, C.; Malossi, C. Bagan: Data Augmentation with Balancing GAN. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018. [Google Scholar]
  16. Zhao, S.; Liu, Z.; Lin, J.; Zhu, J.Y.; Han, S. Differentiable Augmentation for Data-efficient GAN Training. Adv. Neural Inf. Process. Syst. 2020, 33, 7559–7570. [Google Scholar]
  17. Jiang, L.; Dai, B.; Wu, W.; Loy, C.C. Deceive D: Adaptive Pseudo Augmentation for GAN Training with Limited Data. Adv. Neural Inf. Process. Syst. 2021, 34, 21655–21667. [Google Scholar]
  18. DeVries, T.; Taylor, G.W. Improved Regularization of Convolutional Neural Networks with Cutout. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  19. Zhong, Z.; Zheng, L.; Kang, G.; Li, S.; Yang, Y. Random Erasing Data Augmentation. AAAI Conf. Artif. Intell. 2020, 34, 13001–13008. [Google Scholar] [CrossRef]
  20. Kumar Singh, K.; Jae Lee, Y. Hide-and-seek: Forcing a Network to Be Meticulous for Weakly-supervised Object and Action Localization. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 3524–3533. [Google Scholar]
  21. Chen, P.; Liu, S.; Zhao, H.; Wang, X.; Jia, J. Gridmask Data Augmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020. [Google Scholar]
  22. Takahashi, R.; Matsubara, T.; Uehara, K. Ricap: Random Image Cropping and Patching Data Augmentation for Deep CNNs. In Proceedings of the Asian Conference on Machine Learning, PMLR, Beijing, China, 14–16 November 2018; pp. 786–798. [Google Scholar]
  23. Droste, R.; Jiao, J.; Noble, J. Unified image and video saliency modeling. In Proceedings of the European Computer Vision Conference, Glasgow, UK, 23–28 August 2020; Volume 16, pp. 419–435. [Google Scholar]
  24. Liu, N.; Han, J. A Deep Spatial Contextual Longterm Recurrent Convolutional Network for Saliency Detection. IEEE Trans. Image Process. 2018, 27, 3264–3274. [Google Scholar] [CrossRef] [PubMed]
  25. Djilali, Y.A.D.; McGuinness, K.; O’Connor, N. Learning Saliency from Fixations. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–8 January 2024; pp. 383–393. [Google Scholar]
  26. Hosseini, A.; Kazerouni, A.; Akhavan, S.; Brudno, M.; Taati, B. SUM: Saliency Unification through Mamba for Visual Attention Modeling. arXiv 2024, arXiv:2406.17815. [Google Scholar]
  27. Khan, Z.F.; Kannan, A. Intelligent Segmentation of Medical Images using Fuzzy Bitplane Thresholding. Meas. Sci. Rev. 2014, 14, 94–101. [Google Scholar] [CrossRef]
  28. Dubey, S.R.; Singh, S.K.; Singh, R.K. Local Bit-plane Decoded Pattern: A Novel Feature Descriptor for Biomedical Image Retrieval. IEEE J. Biomed. Health Inform. 2015, 20, 1139–1147. [Google Scholar] [CrossRef]
  29. Tuan, T.A.; Kim, J.Y.; Bao, P.T. 3D Brain Magnetic Resonance Imaging Segmentation by Using Bitplane and Adaptive Fast Marching. Int. J. Imaging Syst. Technol. 2018, 28, 223–230. [Google Scholar] [CrossRef]
  30. Vladimir, V. Statistical Learning Theory; Wiley: New York, NY, USA, 1998. [Google Scholar]
  31. He, Z.; Xie, L.; Chen, X.; Zhang, Y.; Wang, Y.; Tian, Q. Data Augmentation Revisited: Rethinking the Distribution Gap between Clean and Augmented Data. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  32. Zhang, J.; Sclaroff, S. Exploiting Surroundedness for Saliency Detection: A Boolean Map Approach. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 38, 889–902. [Google Scholar] [CrossRef] [PubMed]
  33. Koller, D.; Friedman, N. Probabilistic Graphical Models: Principles and Techniques; MIT Press: Cambridge, MA, USA, 2009. [Google Scholar]
  34. Liu, Z.; Yuan, L.; Weng, L.; Yang, Y. A High Resolution Optical Satellite Image Dataset for Ship Recognition and Some New Baselines. Int. Conf. Pattern Recognit. Appl. Methods 2017, 2, 324–331. [Google Scholar]
  35. Ding, J.; Xue, N.; Long, Y.; Xia, G.S.; Lu, Q. Learning RoI Transformer for Oriented Object Detection in Aerial Images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 2849–2858. [Google Scholar]
  36. Ma, J.; Shao, W.; Ye, H.; Wang, L.; Wang, H.; Zheng, Y.; Xue, X. Arbitrary-oriented Scene Text Detection via Rotation Proposals. IEEE Trans. Multimed. 2018, 20, 3111–3122. [Google Scholar] [CrossRef]
  37. Dai, J.; Qi, H.; Xiong, Y.; Li, Y.; Zhang, G.; Hu, H.; Wei, Y. Deformable Convolutional Networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 764–773. [Google Scholar]
  38. Azimi, S.M.; Vig, E.; Bahmanyar, R.; Körner, M.; Reinartz, P. Towards Multi-class Object Detection in Unconstrained Remote Sensing Imagery. In Asian Conference on Computer Vision; Springer International Publishing: Cham, Switzerland, 2018; pp. 150–165. [Google Scholar]
  39. Fu, K.; Chang, Z.; Zhang, Y.; Xu, G.; Zhang, K.; Sun, X. Rotation-aware and Multi-scale Convolutional Neural Network for Object Detection in Remote Sensing Images. ISPRS J. Photogramm. Remote Sens. 2020, 161, 294–308. [Google Scholar] [CrossRef]
  40. Yang, X.; Sun, H.; Sun, X.; Yan, M.; Guo, Z.; Fu, K. Position Detection and Direction Prediction for Arbitrary-oriented Ships via Multiscale Rotation Region Convolutional Neural Network. IEEE Access 2018, 6, 50839–50849. [Google Scholar] [CrossRef]
  41. Ding, J.; Xue, N.; Xia, G.-S.; Bai, X.; Yang, W.; Yang, M.Y.; Belongie, S.; Luo, J.; Datcu, M.; Pelillo, M.; et al. Object Detection in Aerial Images: A Large-scale Benchmark and Challenges. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 7778–7796. [Google Scholar] [CrossRef] [PubMed]
  42. Zhang, H.; Leng, W.; Han, X.; Sun, W. MOON: A Subspace-Based Multi-Branch Network for Object Detection in Remotely Sensed Images. Remote Sens. 2023, 15, 4201. [Google Scholar] [CrossRef]
  43. Yi, J.; Wu, P.; Liu, B.; Huang, Q.; Qu, H.; Metaxas, D. Oriented Object Detection in Aerial Images with Box Boundary-aware Vectors. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Virtual, 5–9 January 2021; pp. 2150–2159. [Google Scholar]
  44. Zhang, H.; Xu, Z.; Han, X.; Sun, W. Refining FFT-based Heatmap for the Detection of Cluster Distributed Targets in Satellite Images. In Proceedings of the British Machine Vision Conference, Online, 22–25 November 2021. [Google Scholar]
  45. Han, J.; Ding, J.; Xue, N.; Xia, G.S. Redet: A Rotation-equivariant Detector for Aerial Object Detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 19–25 June 2021; pp. 2786–2795. [Google Scholar]
  46. Yang, X.; Yan, J.; Feng, Z.; He, T. R3det: Refined Single-stage Detector with Feature Refinement for Rotating Object. Proc. AAAI Conf. Artif. Intell. 2021, 35, 3163–3171. [Google Scholar] [CrossRef]
Figure 1. The extraction of meaningful focusing regions on sample image which contains numerous small targets.
Figure 1. The extraction of meaningful focusing regions on sample image which contains numerous small targets.
Remotesensing 16 04806 g001
Figure 2. The diagram of emphasized small targets on the averaged salient map. (a) Original image which contains small targets, where only positive regions are marked with yellow boxes. (b) Averaged salient map which emphasizes all meaningful regions, where typical negative regions are marked with green boxes.
Figure 2. The diagram of emphasized small targets on the averaged salient map. (a) Original image which contains small targets, where only positive regions are marked with yellow boxes. (b) Averaged salient map which emphasizes all meaningful regions, where typical negative regions are marked with green boxes.
Remotesensing 16 04806 g002
Figure 3. Region-focused data augmentation with different sub-policies. (a) Original image. (b) Averaged salient map of the original image. (c) Region-focused data augmentation with the Ω 1 sub-policy. (d) Region-focused data augmentation with the Ω 2 sub-policy. (e) Region-focused data augmentation with the Ω 3 sub-policy. (f) Region-focused data augmentation with the Ω 4 sub-policy.
Figure 3. Region-focused data augmentation with different sub-policies. (a) Original image. (b) Averaged salient map of the original image. (c) Region-focused data augmentation with the Ω 1 sub-policy. (d) Region-focused data augmentation with the Ω 2 sub-policy. (e) Region-focused data augmentation with the Ω 3 sub-policy. (f) Region-focused data augmentation with the Ω 4 sub-policy.
Remotesensing 16 04806 g003
Figure 4. Total framework of our proposed method.
Figure 4. Total framework of our proposed method.
Remotesensing 16 04806 g004
Table 1. Comparison results of different object detection methods on the DOTA server.
Table 1. Comparison results of different object detection methods on the DOTA server.
PLBDBRGTFSVLVSHTCBCSTSBFRAHASPHCmAP
RRPN [36]80.965.835.367.459.950.955.890.766.972.455.152.255.153.448.261.0
Yang et al. [40]81.371.436.567.461.250.956.690.768.172.455.155.662.453.451.562.3
DCN [37]85.075.731.673.458.946.160.789.675.276.852.361.747.658.744.062.5
ICN [38]81.474.347.770.364.967.870.090.879.178.253.662.967.064.250.268.2
FFA-3 [39]88.874.448.857.963.675.979.690.880.382.954.360.066.966.842.568.9
RT-Mxnet [35]88.678.543.475.968.873.783.690.777.381.558.453.562.858.947.769.6
RT-Pytorch [41]88.076.152.672.178.077.887.390.484.782.753.962.775.468.256.473.8
RT-Pytorch + GHM [42]88.7 77.4 53.9 77.477.6 77.6 87.790.886.885.661.9 60.1 76.1 70.5 64.375.8
RT-Pytorch + ROBIT89.284.154.876.078.482.887.790.884.585.465.663.277.171.659.576.7
Table 2. Comparison results of different augmentation methods on the DOTA server.
Table 2. Comparison results of different augmentation methods on the DOTA server.
PLBDBRGTFSVLVSHTCBCSTSBFRAHASPHCmAP
RT-Pytorch88.076.152.672.178.077.887.390.484.782.753.962.775.468.256.473.8
RT-Pytorch + C&R89.478.254.372.772.476.587.790.880.585.760.161.976.072.458.774.5
RT-Pytorch + C&S88.976.454.376.873.476.787.490.886.978.764.763.075.871.155.274.7
RT-Pytorch + BIRD88.978.652.775.771.876.987.690.885.684.263.462.377.070.855.174.8
RT-Pytorch + rBIRD88.982.653.077.173.177.187.590.886.384.462.761.374.871.158.475.3
RT-Pytorch + IKD89.283.253.775.577.782.287.690.786.984.961.962.877.171.758.776.2
RT-Pytorch + ROBIT89.284.154.876.078.482.887.790.884.585.465.663.277.171.659.576.7
Table 3. Experiments on the sub-polices for region-based data augmentation.
Table 3. Experiments on the sub-polices for region-based data augmentation.
PLBDBRGTFSVLVSHTCBCSTSBFRAHASPHCmAP
RT-Pytorch88.076.152.672.178.077.887.390.484.782.753.962.775.468.256.473.8
RT-Pytorch + Ω 1 88.978.254.975.778.177.987.590.986.385.461.065.577.271.858.075.8
RT-Pytorch + Ω 2 89.183.354.674.378.377.787.890.887.085.767.262.277.170.963.676.6
RT-Pytorch + Ω 3 89.384.754.776.778.482.787.690.787.185.564.962.276.971.556.676.6
RT-Pytorch + Ω 4 89.284.154.876.078.482.887.790.884.585.465.663.277.171.659.576.7
Table 4. Ablation study on the DOTA server, where bitplane recombination is termed BR and region focusing is termed RF.
Table 4. Ablation study on the DOTA server, where bitplane recombination is termed BR and region focusing is termed RF.
PLBDBRGTFSVLVSHTCBCSTSBFRAHASPHCmAP
RT-Pytorch88.076.152.672.178.077.887.390.484.782.753.962.775.468.256.473.8
RT-Pytorch + BR88.978.652.775.771.876.987.690.885.684.263.462.377.070.855.174.8
RT-Pytorch + RF88.978.054.675.373.777.787.590.886.585.660.161.676.869.759.275.1
RT-Pytorch +ROBIT89.284.154.876.078.482.887.790.884.585.465.663.277.171.659.576.7
Table 5. Comparison results on all regions and only positive regions.
Table 5. Comparison results on all regions and only positive regions.
MethodmAP
HeatNet74.8
HeatNet + only positive samples75.2
HeatNet + ROBIT75.5
Table 6. Comparison results on the HRSC2016 dataset.
Table 6. Comparison results on the HRSC2016 dataset.
MethodmAP
RC2 [34]75.7
RRPN [36]79.1
RT [35]80.1
BBAvector [43]82.8
HeatNet [44]84.7
R3Det [46]89.3
ReDet [45]96.4
ReDet+ROBIT97.7
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, H.; Han, X.; Sun, W. Region-Focusing Data Augmentation via Salient Region Activation and Bitplane Recombination for Target Detection. Remote Sens. 2024, 16, 4806. https://doi.org/10.3390/rs16244806

AMA Style

Zhang H, Han X, Sun W. Region-Focusing Data Augmentation via Salient Region Activation and Bitplane Recombination for Target Detection. Remote Sensing. 2024; 16(24):4806. https://doi.org/10.3390/rs16244806

Chicago/Turabian Style

Zhang, Huan, Xiaolin Han, and Weidong Sun. 2024. "Region-Focusing Data Augmentation via Salient Region Activation and Bitplane Recombination for Target Detection" Remote Sensing 16, no. 24: 4806. https://doi.org/10.3390/rs16244806

APA Style

Zhang, H., Han, X., & Sun, W. (2024). Region-Focusing Data Augmentation via Salient Region Activation and Bitplane Recombination for Target Detection. Remote Sensing, 16(24), 4806. https://doi.org/10.3390/rs16244806

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop