Next Article in Journal
A Method for Underwater Acoustic Target Recognition Based on the Delay-Doppler Joint Feature
Next Article in Special Issue
A Novel Hybrid Deep-Learning Approach for Flood-Susceptibility Mapping
Previous Article in Journal
FFEDet: Fine-Grained Feature Enhancement for Small Object Detection
Previous Article in Special Issue
Spatiotemporal Mechanism-Based Spacetimeformer Network for InSAR Deformation Prediction and Identification of Retrogressive Thaw Slumps in the Chumar River Basin
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

DiffuPrompter: Pixel-Level Automatic Annotation for High-Resolution Remote Sensing Images with Foundation Models

1
College of Information Science and Engineering, Northeastern University, Shenyang 110819, China
2
Peng Cheng Laboratory, Shenzhen 518055, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2024, 16(11), 2004; https://doi.org/10.3390/rs16112004
Submission received: 23 April 2024 / Revised: 24 May 2024 / Accepted: 29 May 2024 / Published: 2 June 2024

Abstract

:
Instance segmentation is pivotal in remote sensing image (RSI) analysis, aiding in many downstream tasks. However, annotating images with pixel-wise annotations is time-consuming and laborious. Despite some progress in automatic annotation, the performance of existing methods still needs improvement due to the high precision requirements for pixel-level annotation and the complexity of RSIs. With the support of large-scale data, some foundational models have made significant progress in semantic understanding and generalization capabilities. In this paper, we delve deep into the potential of the foundational models in automatic annotation and propose a training-free automatic annotation method called DiffuPrompter, achieving pixel-level automatic annotation of RSIs. Extensive experimental results indicate that the proposed method can provide reliable pseudo-labels, significantly reducing the annotation costs of the segmentation task. Additionally, the cross-domain validation experiments confirm the powerful effectiveness of large-scale pseudo-data in improving model generalization performance.

1. Introduction

With the development of deep learning, the interpretation of RSIs has also made significant progress [1]. Instance segmentation is a crucial part of remote sensing interpretation. However, it is also a data-intensive task that requires a significant amount of pixel-level annotations, which are labor-intensive and expensive, limiting the development of the task. To reduce the annotation costs, some scholars have introduced the automatic image annotation task, AIA for short, which has become an integral component of computer vision [2]. Although the scarcity of pixel-level annotated datasets in the remote sensing domain may be more severe, previous AIA methods have mainly focused on natural images. The perspective effect in natural images often results in distinct foreground and background elements. Therefore, some methods use the attention maps of classification networks to locate the foreground region and consider the image category as the foreground class to generate a pseudo-mask for objects [3,4,5]. In contrast, the top-down perspective and complex image content in RSIs diminish the effectiveness of these AIA methods.
Recently, the potential of big data has been further explored, leading to the emergence of many fundamental models trained on large-scale datasets, such as the Stable Diffusion Model (SDM) [6] and the Segment Anything Model (SAM) [7], which show significant contributions to downstream tasks [8]. Although fundamental models have become plentiful, none are tailored for RSI, which limits their application in RSI tasks. The main goal of this paper is not to create a fundamental model tailored for RSIs but to investigate the applicability of existing fundamental models to pixel-level AIA in RSIs.
With the development of text-guided generation models such as SDM, a pixel-level AIA technical route based on synthetic images has emerged in natural images [9,10]. These methods utilize generative models to synthesize data and generate masks for objects in synthetic images via vision–text alignment knowledge during generation [11]. Generative-based methods account for two assumptions: (1) the synthetic images are realistic enough to avoid domain shift issues between training and testing sets, and (2) the vision–text alignment knowledge can guide the generation of sufficiently accurate object masks. While these assumptions hold in natural images, they do not necessarily apply in RSIs. Generative models trained on natural images cannot synthesize RSIs realistically enough, and the complex and diverse scenes in remote sensing images seriously interfere with vision–text alignment, making it difficult to segment objects accurately. Some scholars incorporate the SAM model into instance segmentation methods for more accurate results [12]. After training on over one billion masks, the SAM model has demonstrated an outstanding ability to segment anything. However, SAM is a class-agnostic segmentation method and requires prior positional cues, such as points and bounding boxes, to segment target objects. These limitations prevent SAM from being directly applied to the instance segmentation task.
As shown in Figure 1, we present new insight into automatically obtaining mask annotations for authentic images using pre-trained foundational models to reduce the annotation cost of the instance segmentation task. Based on the insight, we propose a training-free prompt generation method called DiffuPrompter, which transforms SAM from a category-agnostic into a category-aware segmentation method to label RSIs automatically. Specifically, the proposed DiffuPrompter model leverages the text-concept grounding capabilities of a pre-trained diffusion model to provide coarse localization results for target objects. These localization results are then used as visual segmentation prompts for the SAM model, enabling precise segmentation of the target objects. We tested various automatic annotation methods on remote sensing datasets, and the experimental results validated the superiority of DiffuPrompter, i.e., it achieved 27.3% and 15.4% AP on the NWPU and iSAID datasets, respectively. Furthermore, the cross-domain study demonstrates the positive impact of pseudo-labels on improving model generalization performance, providing a valuable reference for future research.
The main contributions of this paper can be summarized as follows:
  • We present the novel insight that it is possible to automatically obtain the mask annotation of authentic images using off-the-shelf foundational models.
  • We propose a training-free prompt generation method, DiffuPrompter, that transforms SAM from a class-agnostic segmenter to a class-aware segmenter to label RSIs automatically.
  • We tested several automatic annotation methods on remote sensing datasets, and the extensive results validated the superiority of the proposed DiffuPrompter while proving the positive impact of pseudo-labels on enhancing model generalization performance. The results may provide a reference for future work.

2. Theory and Methods

DiffuPrompter utilizes the pre-trained SDM to explore generating semantically explicit prompts for SAM, enabling it to generate masks for specified remote sensing objects automatically. Section 2.1 introduces the working principles of SDM and SAM. In Section 2.2.1, we introduce how to realize the grounding of textual concepts into input images, and we discuss noise suppression in Section 2.2.2. Section 2.2.3 introduces how to prompt SAM to segment specific objects.

2.1. Preliminary Knowledge

2.1.1. Overview of SDM

SDM [6] is derived from a perceptual compression model consisting of an autoencoder, a U-Net, and a decoder. Specifically, when we input an image x R H × W × 3 , the encoder E encodes x into a latent representation z = E ( x ) R h × w × c , and the decoder D reconstructs the image from z, i.e., x ˜ = D ( z ) = D ( E ( x ) ) . The encoder downsamples the image according to the sampling factor f = H / h = W / w , where f has different values in different layers of the encoder and U-Net, namely f = 2 m , where m N .
The training process of SDM consists of a forward diffusion and a backward denoising stage. In the forward diffusion stage, SDM adds noise to z for T steps until z is completely replaced by noise N ( 0 , 1 ) . In the backward denoising stage, SDM learns to gradually remove the noise by U-Net based on the textual condition to recover z. Finally, z is decoded into an image by D ( z ) . SDM achieves semantic mapping between visual and textual inputs through the cross-attention mechanism in U-Net [13]. To pre-process the condition text prompt P t e x t , SDM introduces a domain-specific encoder τ θ that projects P t e x t to an intermediate representation, which is then mapped to the intermediate layers of the U-Net via cross-attention layers as follows:
Attention ( Q , K , V ) = softmax Q K T d · V , Q = W Q ( i ) · φ i ( z t ) , K = W K ( i ) · τ θ ( P t e x t ) , V = W V ( i ) · τ θ ( P t e x t ) .
Here, φ i z t R N × d ϵ i refers the intermediate embedding in the U-Net implementation, and W V ( i ) R d × d E i and W Q ( i ) R d × d τ & W K ( i ) R d × d τ are learnable projection matrices [13,14].

2.1.2. Overview of SAM

SAM is an interactive segmentation approach predicated on provided prompts such as instance points and bounding boxes. The mask-generation process can be expressed as follows:
z img = E i - e n c ( x ) , z i n t e r = E p - e n c P i n t e r , z m a s k = E p - e n c P m a s k , z o u t = Cat T m c - f i l t e r , T I o U , z i n t e r , M = E m - d e c z i m g + z m a s k , z out ,
where z i m g R h × w × c represents the latent representation of the input image; P i n t e r denotes the interactive prompts, including points and bounding boxes; P m a s k R k × c signifies the mask prompt tokens, which are from the previous prediction iteration; E p - e n c encodes prompts into features z i n t e r R h × w × c and z m a s k R h × w × c ; T m c - f i l t e r R 4 × c and T I o U R 1 × c are the pre-inserted learnable tokens representing four different mask filters and their corresponding IoU predictions; and M denotes the predicted masks. The primary objective of DiffuPrompter is to provide P i n t e r (points and bounding boxes) for SAM to segment the target object.

2.2. Proposed Method

2.2.1. Textual Concept Grounding

Upon further exploration of the SDM training process, we discovered that, during the training of SDM, the U-Net restores the latent presentations of the input image step by step based on the text description. Equation (1) illustrates that textual concepts V = W V ( i ) · τ θ ( P t e x t ) are directed into latent presentations through cross-attention of the cross-modal spatial transformer module in U-Net.
Based on this observation, we constructed a text semantic grounding pipeline centered on the cross-attention map. This pipeline utilizes classification datasets as its data source and grounds image categories into the images. As illustrated in Figure 2, given an image from a classification dataset, the textual description is achieved by using the ‘Photo of a category’ template to process the corresponding class name. Then, the cross-attention layer grounds each text semantic in the template into the visual space by cross-attention maps as follows:
A = Softmax Q K T d ,
where A R H × W denotes the re-shaped attention map. For the j-th text token, e.g., airplane in Figure 3a, the corresponding cross-attention A j R H × W shows the visual location in φ ( z t ) of the j-th token.
We propose integrating grounding results at different resolutions to enhance the accuracy and robustness. The cross-attention pyramid A j s is obtained by applying Equation (3) to different layers in U-Net, where s denotes the attention map from the s-th layer of U-Net. We extract four resolutions in this paper, i.e., 8 × 8 , 16 × 16 , 32 × 32 , and 64 × 64 , as shown in Figure 3b. Then, we aggregate multi-scale grounding results in the cross-attention pyramid by calculating the average map as follows:
A ^ j = Norm ( 1 S s S A j s max ( A j s ) ) ,
where S represents the total number of layers (i.e., four for U-Net). Finally, the attention maps are transformed into probability maps through a normalization layer to facilitate subsequent binarization processing.

2.2.2. Denoise by Noise

Figure 3a indicates that the highlighted regions in the cross-attention map correlate with the regions where that input token is presented. However, the maps are significantly noisy. Figure 4 compares the cross-attention maps between natural and remote-sensing images. Figure 4a shows a precise attention map for the ‘horse’ region. On the contrary, the attention map in Figure 4b for an RSI only roughly indicates the ‘playground’ region, showing a very weak correlation with regions a human would pick out as meaningful. Therefore, it is necessary to denoise the cross-attention maps in SDM before using them to localize remote sensing objects. However, the noise points tend to be localized and demonstrate high randomness in their distribution. Hence, achieving precise removal of these noise points poses a significant challenge.
Inspired by [15], we propose a Loop-Sampling Averaging Denoising (LSAD) strategy to suppress noise interference. In LSAD, we model the observed cross-attention map as a combination of a noise-free attention map and additive noise, as follows:
A ( x , y ) = A ( x , y ) + N ( x , y ) ,
where A ( x , y ) represents the value of the captured cross-attention map at coordinates ( x , y ) , A ( x , y ) signifies the value of the noise-free map, and N ( x , y ) denotes the value of the noise. The denoising process is the procedure of approximating A ( x , y ) from the known A ( x , y ) . For multiple cross-attention maps of the same input image and token, the A ( x , y ) will remain constant, and N ( x , y ) is random. Thus, the mean of L maps of the same image can be represented as follows:
A ¯ ( x , y ) = 1 L i = 1 L A ( x , y ) + N i ( x , y ) = A ( x , y ) + 1 L i = 1 L N i ( x , y ) ,
where A ¯ ( x , y ) denotes the mean value of the maps at coordinates ( x , y ) , and L is the total number of maps considered. As the noise is random and unrelated, the expectation of the mean value approximates zero, i.e., i = 1 k N i ( x , y ) 0 . Therefore, the expected mean and variance of the cross-attention maps can be expressed as follows:
E A ¯ ( x , y ) = A ( x , y ) ,
σ A ¯ ( x , y ) = 1 L σ N ( x , y ) ,
where σ represents the standard deviation. Equation (7) shows that the expected mean value of multiple cross-attention maps is a map without noise. However, there will be some disturbances, and the standard deviation determines the noise’s intensity. The essence of denoising is reducing the standard deviation. Equation (8) indicates that, by increasing the value of L , i.e., increasing the number of averaged maps, the noise can be suppressed effectively.
The main challenge of applying Equation (8) to denoise cross-attention maps lies in introducing random noise onto A ( x , y ) . As mentioned in Section 3, the SDM is trained to construct a clear image from Gaussian noise by removing Gaussian noise step by step. Noise contributes to generation diversity [16], implying attention maps are variable. Therefore, injecting Gaussian noise into the latent embeddings of the input image will result in multiple noisy cross-attention maps. Fortunately, the forward diffusion process in SDM is the noise addition process. Therefore, given an image, we perform an iterative forward diffusion process, preserving the cross-attention pyramid maps during each loop. Subsequently, we apply LSAD to them as follows:
A j ¯ ( x , y ) = 1 L l L A ^ j l ( x , y ) ,
where l represents the l-th sampling iteration. Figure 5 illustrates the workflow of the LSAD algorithm.
Figure 6 visualizes the performance of LSAD on natural and remote sensing images. It can be observed that there is no noticeable noise in the cross-attention map of the natural image. LSAD does not show a significant enhancement effect on the cross-attention map. In contrast, there is much noise in the cross-attention map of the RSI, making it challenging to locate the target object accurately. After LSAD processing, the noise is effectively suppressed, making the highlighted areas more meaningful.

2.2.3. Prompt for SAM

Given a normalized average attention map A ¯ R H × W for the j-th text token to get the target object region (e.g., ‘airplane’). As shown in Figure 3c, the solution to the binarization process is using a threshold value γ and refining with DenseCRF [17] as follows:
B = DenseCRF γ ; A ^ j argmax .
As shown in Figure 7, we take the minimum bounding box and centroid of the areas in B with a value of 1 as the box and point prompts for SAM. Then, SAM will output a mask list based on the prompts. We select the mask with the highest IoU, with the attention mask in B as the final segmentation result. If the selected mask contains multiple closed intervals, it is considered to have multiple objects, such as the boat, airplane, tennis court, and storage tank in Figure 7. If the selected mask contains a single closed interval, we consider that there is a single object, such as the playground. At this point, we have constructed a training-free, pixel-level AIA pipeline using SDM and SAM.

3. Results

3.1. Datasets

iSAID: iSAID [18] is a large-scale dataset for remote sensing instance segmentation inherited from DOTA [19]. The spatial resolutions of images range between 800 and 13,000. We split them into 512 × 512 patches during training and testing. It contains 15 classes of 655,451 instances in 2806 images: ship, storage tank, baseball diamond, tennis court, basketball court, playground, bridge, large vehicle, small vehicle, helicopter, swimming pool, roundabout, soccer ball field, plane, and harbor.
NWPU VHR-10: NWPU VHR-10 [20] is another widely used dataset for object detection of RSIs. It has 800 high-resolution images, among which 650 are positive and 150 are negative, without any objects of interest. This dataset contains annotations of 10 object categories: airplane, ship, storage tank, baseball diamond, tennis court, basketball court, ground track field, harbor, bridge, and vehicle.
Classification Dataset: We selected target images corresponding to the segmentation dataset from 11 classification datasets: UC Merced Land Use Dataset [21], WHU-RS19 [22,23], RSSCN7 [24], RS_C11 [25], NWPU-RESISC45 [26], AID [26], RSD46-WHU [27,28], PatternNet [29], OPTIMAL-31 [30], CLRS [31], and DLR Munich Vehicle [32]. Ultimately, we collected 9300 RSIs across 12 categories of 0.3∼3 m resolution: airplane, ship, storage tank, baseball diamond, swimming pool, tennis court, basketball court, roundabout, ground track field, harbor, bridge, and vehicle. The corresponding classes are selected when testing on the iSAID and NWPU datasets. The spatial resolution is uniformly set to 512 × 512 .

3.2. Evaluation Metrics

We adopted the commonly used mean average precision (mAP) metric to evaluate the performance of the proposed method. When the mask of an instance exists an intersection-over-union (IoU) with a mask of ground truth above a threshold and its predicted category matches the label, the prediction is considered to be true positive. In this study, we employ AP, AP 50 , AP 75 , AP S , AP M , and AP L for evaluation. AP refers to metrics averaged across all 10 IoU thresholds (0.50:0.05:0.95) and all categories. A larger AP value denotes more accurate predicted instance masks and superior instance segmentation performance. AP 50 represents the calculation under the IoU threshold of 0.50, while AP 75 embodies a stricter metric corresponding to the calculation under the IoU threshold 0.75. Therefore, if we havethe same AP 75 and AP 50 values, qhere the AP 75 indicates more accurate instance masks. AP L is set for large targets (area > 96 2 ); AP M is set for medium targets ( 32 2 < area < 96 2 ); and AP S is set for small targets (area < 32 2 ).

3.3. Implementation Details

In this paper, we do not train any parameters in the SDM and SAM. Due to the lack of information about vehicle sizes in the remote sensing classification datasets, we merged ‘small-vehicle’ and ‘large-vehicle’ as one category, i.e., ‘vehicle’, during testing on iSAID. Additionally, since the “helicopter” and “soccer ball field” categories do not exist in the classification dataset, we did not generate pseudo-labels for them. Finally, we collected 9300 RSIs. The total pseudo-label categories and their abbreviations annotated in this paper are: AI-airplane, SH-ship, ST-storage tank, BD-baseball diamond, TC-tennis court, BC-basketball court, RA-roundabout, PL-playground (ground track field), SP-swimming pool, HA-harbor, BR-bridge, and VE-vehicle. Mask R-CNN [33], Cascade R-CNN [34], and Mask2Former [34] were used as the baseline to evaluate our method. Eight Tesla V100 GPUs were used to generate pseudo-labels, which took approximately 96 h.

3.4. Qualitative Experiments

Visualizations in Figure 7 depict the intermediate results of DiffuPrompter in generating pseudo-labels. It can be observed that the proposed method can accurately segment any number of target objects, significantly increasing the number of positive samples in the pseudo-labels. In the Stable Diffusion model, time steps control the noise intensity, affecting the results of DiffuPrompter. Figure 8 visualizes the cross-attention maps at different time steps and shows that the cross-attention map at time step 40 is the clearest, maintaining the structure of the target object. The reason may be that, at t = 40, the artificially introduced noise intensity is close to the inherent noise intensity in the cross-attention map. Therefore, LSAD can effectively suppress noise without losing the contours of the target objects due to excessive noise. Thus, we use cross-attention in step 40 to conduct experiments throughout the paper.

3.5. Ablation Study

We also performed an extensive ablation analysis to better understand the effectiveness of each proposed module in our DiffuPrompter.

3.5.1. Comparison with Attention Map under Different Thresholds

Figure 3c illustrates the impact of different thresholds on the binary image. It is evident that the λ value significantly impacts the prompt quality for SAM. Table 1 qualitatively compares the segmentation performance under different binary thresholds. The term “cross-attention” means using the binarized attention map as pseudo-labels to train the segmentation model. The experimental results indicate that setting the threshold to 0.4 provides the best guidance for the model to segment target objects in both methods. Therefore, the value of λ in subsequent experiments is set to 0.4. Additionally, the performance of DiffuPrompter far exceeds that of the attention map across all thresholds. This can be attributed to the superior segmentation ability of SAM, which provides accurate pseudo-masks for the segmentation model.

3.5.2. Sampling Times

Table 2 provides the related ablation study for sampling times in LSAD. The results reflect that, with the increase in sampling iterations, there is a significant improvement in segmentation performance, and the performance stabilized after reaching 50 sampling iterations. We ultimately adopted 50 iterations as the universal sampling count to balance performance and economy.

3.6. Segmentation Performance Comparison

NWPU: Table 3 presents instance segmentation results on the NWPU. The baseline segmentation methods trained on the data labeled by DiffuPrompter can reach approximately half of the performance achieved with pure real data, e.g., 27.3 % vs. 58.3 % for mask A P of Mask R-CNN. Additionally, further fine-tuning on 600 (25% off) real data can achieve performance comparable to training on pure real data, e.g., 55.6% mask A P after fine-tuning vs. 58.3% mask A P training on pure real data with Mask R-CNN; 59.1% mask A P after fine-tuning vs. 59.8% mask A P training on pure real data with Cascade R-CNN; and 60.9% mask A P after fine-tuning vs. 61.3% mask A P training on pure real data with Mask2Former.
iSAID: Table 4 presents the results on iSAID. iSAID is more challenging than NWPU, as it includes more objects and complex backgrounds. Even in the absence of pseudo-labels for helicopter and football field categories, DiffuPrompter and iSAID still can present a competitive result, i.e., 34.8 % vs. 35.1 % mask A P of Mask R-CNN; 35.1 % vs. 35.6 % mask A P of Cascade Mask R-CNN; 36.8 % vs. 37.1 % mask A P of Mask2Former, when trained on 9300 pseudo-data and 2200 real data (saved 21.4% in manner effort).

3.7. Domain Generalization

Table 5 delivers the results of cross-domain validation, which can evaluate the generalization performance. We tested the performance on overlapping categories under the two datasets. The results indicate that DiffuPrompter plays a prominent role in domain generalization, e.g., 22.5% AP with DiffuPrompter and NWPU vs. 17.9% AP with NWPU on the iSAID test set, and 50.3% AP with DiffuPrompter and iSAID vs. 47.2% AP with iSAID on the NWPU test set.

3.8. Comparison with the State of the Art

Figure 9 visualizes the output results of two advanced AIA methods and our DiffuPrompter. CAM originates from classification models and focuses more on discriminative regions while neglecting details. Consequently, the masks it generates often fail to cover the object entirely. DiffuMask, a recently proposed advanced algorithm, also leverages diffusion models to generate pseudo-masks for objects. It optimizes attention maps in U-Net through noise learning and uses them as pseudo-labels. However, DiffuMask is designed to generate pseudo-labels for synthetic images, which limits its annotation ability for authentic images. From Figure 9, it is evident that the optimized attention map is still easily disturbed by the complex background in RSIs. Moreover, the additional training process significantly increases its annotation cost compared to DiffuPrompter. In contrast, the proposed DiffuPrompter can accurately locate and label masks for target objects without training.
Table 6 quantitatively compares the performance of the proposed method with other state-of-the-art algorithms. The results in Table 6 show that the proposed method significantly outperforms the baselines in terms of accuracy. However, there is still room for improvement in annotation speed.
Table 7 compares DiffuPrompter with some advanced weakly supervised methods. Here, we denote the supervision type as: I (image-level label), B (box-level label). DiffuPrompter M and DiffuPrompter C represent Mask R-CNN and Cascade Mask R-CNN trained with pseudo-labels generated by DiffuPrompter, respectively. The results show that DiffuPrompter M and DiffuPrompter C significantly outperform methods based on image-level supervision, e.g., 29.9% AP with DiffuPrompter C vs. 13.3% AP with BESTIE on NWPU, and achieve performance comparable to methods based on object-level supervision, e.g., 29.9% AP with DiffuPrompter C vs. 29.8% AP with MGWI-Net on NWPU. In conclusion, DiffuPrompter can perform better with less manual labor than existing advanced methods.
Table 8 compares the class-wise performance of DiffuPrompter M against several advanced supervised methods trained by ground truth. The results indicate that most categories achieve nearly half the performance of supervised methods. However, the performance of “bridge” and “vehicle” is significantly lower than that of supervised methods. We hypothesize that this is due to the significant scale difference between the pseudo-labels and the segmentation dataset for these two categories. The pseudo-labels do not adequately guide the segmentation of these objects in the target dataset. Therefore, addressing the cross-dataset issues between pseudo-labels and the target dataset is worth further investigation.

4. Discussion

To obtain more accurate prompts, we fine-tuned SDM on a combination of some RSI caption datasets (RSICD [44], UCM-captions [45], and Sydney-captions [45]). However, the cross-attention maps of the fine-tuned SDM, as shown in Figure 10, cannot provide precise guidance for SAM. The reason may be that the image caption datasets for RSIs are too small in scale and insufficient to convey accurate text–visual correspondence information to the SDM model. Therefore, the SDM used in this paper has yet to be fine-tuned. It is worth exploring how to fine-tune the SDM model on remote sensing datasets to obtain a more familiar understanding of remote sensing objects.
Another factor affecting performance is domain discrepancy. The significant differences between the classification and object detection datasets result in the knowledge provided by the pseudo-labels not being well applied to the dataset. Therefore, addressing the cross-domain issue between pseudo-labels and target data is also a valuable research direction.
In this paper, we determined the hyperparameter values based on the overall performance of the segmentation algorithm. However, the optimal hyperparameters vary slightly across different categories. Table 9 presents the category-level results under different parameters, indicating that designing adaptive hyperparameters could further improve the quality of the pseudo-labels. Therefore, developing an adaptive pseudo-labeling algorithm tailored to each category is a promising direction for future research.

5. Conclusions

This paper introduces a novel insight that demonstrates the possibility of automatically annotating object masks for RSIs by leveraging off-the-shelf foundational models. Building on the SDM and SAM models, we propose an AIA method, namely DiffuPrompter, capable of leveraging the text semantic grounding knowledge in SDM to generate semantically precise SAM prompts, enabling it to acquire instance masks autonomously. We comprehensively test the effectiveness of the proposed method on two general datasets. We first evaluate the efficacy of each component in DiffuPrompter through ablation studies. Then, the cross-domain validation experiments confirm the significant effectiveness of large-scale pseudo-data in improving model generalization performance. Finally, we compare our method with other state-of-the-art algorithms, and the results demonstrate the superiority of our proposed method over existing ones.

Author Contributions

Conceptualization, H.L.; theory and methodology, H.L.; software (Python 3.9), H.L. and H.P.; visualization, H.L.; formal analysis, H.L.; writing—original draft preparation, H.L.; writing—review and editing, H.L.; supervision, Y.W.; project administration, Y.W. and W.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Open Project Program Foundation of the Key Laboratory of Opto-Electronics Information Processing, Chinese Academy of Sciences (OEIP-O-202002), the National Nature Science Foundation of China (grant No. 61871106 and No. 61370152), and the Key R & D projects of Liaoning Province, China (grant No. 2020JH2/10100029).

Data Availability Statement

The original contributions presented in the study are included in the article, and further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Chen, K.; Zou, Z.; Shi, Z. Building extraction from remote sensing images with sparse token transformers. Remote Sens. 2021, 13, 4441. [Google Scholar] [CrossRef]
  2. Cheng, Q.; Zhang, Q.; Fu, P.; Tu, C.; Li, S. A survey and analysis on automatic image annotation. Pattern Recognit. 2018, 79, 242–259. [Google Scholar] [CrossRef]
  3. Wu, T.; Huang, J.; Gao, G.; Wei, X.; Wei, X.; Luo, X.; Liu, C.H. Embedded discriminative attention mechanism for weakly supervised semantic segmentation. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 16765–16774. [Google Scholar] [CrossRef]
  4. Xu, L.; Ouyang, W.; Bennamoun, M.; Boussaid, F.; Sohel, F.; Xu, D. Leveraging auxiliary tasks with affinity learning for weakly supervised semantic segmentation. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 6984–6993. [Google Scholar] [CrossRef]
  5. Ru, L.; Zhan, Y.; Yu, B.; Du, B. Learning affinity from attention: End-to-end weakly-supervised semantic segmentation with transformers. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 16846–16855. [Google Scholar] [CrossRef]
  6. Rombach, R.; Blattmann, A.; Lorenz, D.; Esser, P.; Ommer, B. High-resolution image synthesis with latent diffusion models. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 10684–10695. [Google Scholar] [CrossRef]
  7. Kirillov, A.; Mintun, E.; Ravi, N.; Mao, H.; Rolland, C.; Gustafson, L.; Xiao, T.; Whitehead, S.; Berg, A.C.; Lo, W.Y.; et al. Segment anything. In Proceedings of the 2023 IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 1–6 October 2023; pp. 4015–4026. [Google Scholar] [CrossRef]
  8. Chen, J.; Chen, H.; Chen, K.; Zhang, Y.; Zou, Z.; Shi, Z. Diffusion models for imperceptible and transferable adversarial attack. arXiv 2023, arXiv:2305.08192. [Google Scholar]
  9. Zhang, Y.; Ling, H.; Gao, J.; Yin, K.; Lafleche, J.F.; Barriuso, A.; Torralba, A.; Fidler, S. Datasetgan: Efficient labeled data factory with minimal human effort. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 10145–10155. [Google Scholar]
  10. Li, D.; Ling, H.; Kim, S.W.; Kreis, K.; Fidler, S.; Torralba, A. BigDatasetGAN: Synthesizing ImageNet with Pixel-wise Annotations. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 21330–21340. [Google Scholar] [CrossRef]
  11. Wu, W.; Zhao, Y.; Shou, M.Z.; Zhou, H.; Shen, C. Diffumask: Synthesizing images with pixel-level annotations for semantic segmentation using diffusion models. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 1–6 October 2023; pp. 1206–1217. [Google Scholar] [CrossRef]
  12. Chen, K.; Liu, C.; Chen, H.; Zhang, H.; Li, W.; Zou, Z.; Shi, Z. RSPrompter: Learning to prompt for remote sensing instance segmentation based on visual foundation model. IEEE Trans. Geosci. Remote Sens. 2024, 62, 4701117. [Google Scholar] [CrossRef]
  13. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Processing Syst. 2017, 30. [Google Scholar] [CrossRef] [PubMed]
  14. Jaegle, A.; Gimeno, F.; Brock, A.; Vinyals, O.; Zisserman, A.; Carreira, J. Perceiver: General perception with iterative attention. In Proceedings of the International Conference on Machine Learning, PMLR, Virtual, 18–24 July 2021; pp. 4651–4664. [Google Scholar] [CrossRef]
  15. Smilkov, D.; Thorat, N.; Kim, B.; Viégas, F.; Wattenberg, M. Smoothgrad: Removing noise by adding noise. arXiv 2017, arXiv:1706.03825. [Google Scholar]
  16. Ho, J.; Jain, A.; Abbeel, P. Denoising diffusion probabilistic models. Adv. Neural Inf. Process. Syst. 2020, 33, 6840–6851. [Google Scholar] [CrossRef]
  17. Krähenbühl, P.; Koltun, V. Efficient inference in fully connected crfs with gaussian edge potentials. Adv. Neural Inf. Process. Syst. 2011, 24, 109–117. Available online: https://dl.acm.org/doi/10.5555/2986459.2986472 (accessed on 22 April 2024).
  18. Waqas Zamir, S.; Arora, A.; Gupta, A.; Khan, S.; Sun, G.; Shahbaz Khan, F.; Zhu, F.; Shao, L.; Xia, G.S.; Bai, X. isaid: A large-scale dataset for instance segmentation in aerial images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Long Beach, CA, USA, 16–17 June 2019; pp. 28–37. Available online: https://api.semanticscholar.org/CorpusID:170079084 (accessed on 22 April 2024).
  19. Xia, G.S.; Bai, X.; Ding, J.; Zhu, Z.; Belongie, S.; Luo, J.; Datcu, M.; Pelillo, M.; Zhang, L. DOTA: A large-scale dataset for object detection in aerial images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3974–3983. [Google Scholar] [CrossRef]
  20. Cheng, G.; Han, J.; Zhou, P.; Guo, L. Multi-class geospatial object detection and geographic image classification based on collection of part detectors. ISPRS J. Photogramm. Remote Sens. 2014, 98, 119–132. [Google Scholar] [CrossRef]
  21. Yang, Y.; Newsam, S. Bag-of-visual-words and spatial extensions for land-use classification. In Proceedings of the 18th SIGSPATIAL International Conference on Advances in Geographic Information Systems, San Jose, CA, USA, 2–5 November 2010; pp. 270–279. [Google Scholar] [CrossRef]
  22. Xia, G.S.; Yang, W.; Delon, J.; Gousseau, Y.; Sun, H.; Maître, H. Structural high-resolution satellite image indexing. In Proceedings of the ISPRS TC VII Symposium-100 Years ISPRS, Vienna, Austria, 5–7 July 2010; Volume 38, pp. 298–303. Available online: https://api.semanticscholar.org/CorpusID:18018842 (accessed on 22 April 2024).
  23. Dai, D.; Yang, W. Satellite Image Classification via Two-Layer Sparse Coding with Biased Image Representation. IEEE Geosci. Remote Sens. Lett. 2011, 8, 173–176. [Google Scholar] [CrossRef]
  24. Zou, Q.; Ni, L.; Zhang, T.; Wang, Q. Deep learning based feature selection for remote sensing scene classification. IEEE Geosci. Remote Sens. Lett. 2015, 12, 2321–2325. [Google Scholar] [CrossRef]
  25. Zhao, L.; Tang, P.; Huo, L. Feature significance-based multibag-of-visual-words model for remote sensing image scene classification. J. Appl. Remote Sens. 2016, 10, 035004. [Google Scholar] [CrossRef]
  26. Cheng, G.; Han, J.; Lu, X. Remote sensing image scene classification: Benchmark and state of the art. Proc. IEEE 2017, 105, 1865–1883. [Google Scholar] [CrossRef]
  27. Long, Y.; Gong, Y.; Xiao, Z.; Liu, Q. Accurate object localization in remote sensing images based on convolutional neural networks. IEEE Trans. Geosci. Remote Sens. 2017, 55, 2486–2498. [Google Scholar] [CrossRef]
  28. Xiao, Z.; Long, Y.; Li, D.; Wei, C.; Tang, G.; Liu, J. High-resolution remote sensing image retrieval based on CNNs from a dimensional perspective. Remote Sens. 2017, 9, 725. [Google Scholar] [CrossRef]
  29. Zhou, W.; Newsam, S.; Li, C.; Shao, Z. PatternNet: A benchmark dataset for performance evaluation of remote sensing image retrieval. ISPRS J. Photogramm. Remote 2018, 145, 197–209. [Google Scholar] [CrossRef]
  30. Wang, Q.; Liu, S.; Chanussot, J.; Li, X. Scene classification with recurrent attention of VHR remote sensing images. IEEE Trans. Geosci. Remote Sens. 2018, 57, 1155–1167. [Google Scholar] [CrossRef]
  31. Li, H.; Jiang, H.; Gu, X.; Peng, J.; Li, W.; Hong, L.; Tao, C. CLRS: Continual learning benchmark for remote sensing image scene classification. Sensors 2020, 20, 1226. [Google Scholar] [CrossRef]
  32. Liu, K.; Mattyus, G. Fast Multiclass Vehicle Detection on Aerial Images. IEEE Geosci. Remote Sens. Lett. 2015, 12, 1938–1942. [Google Scholar] [CrossRef]
  33. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask r-cnn. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2961–2969. [Google Scholar] [CrossRef]
  34. Cheng, B.; Misra, I.; Schwing, A.G.; Kirillov, A.; Girdhar, R. Masked-attention mask transformer for universal image segmentation. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 1290–1299. [Google Scholar] [CrossRef]
  35. Zhou, B.; Khosla, A.; Lapedriza, A.; Oliva, A.; Torralba, A. Learning deep features for discriminative localization. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 2921–2929. [Google Scholar] [CrossRef]
  36. Kolesnikov, A.; Lampert, C.H. Seed, expand and constrain: Three principles for weakly-supervised image segmentation. In Proceedings of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Proceedings, Part IV 14. Springer: Berlin/Heidelberg, Germany, 2016; pp. 695–711. [Google Scholar]
  37. Ahn, J.; Kwak, S. Learning pixel-level semantic affinity with image-level supervision for weakly supervised semantic segmentation. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4981–4990. [Google Scholar] [CrossRef]
  38. Kim, B.; Yoo, Y.; Rhee, C.E.; Kim, J. Beyond semantic to instance segmentation: Weakly-supervised instance segmentation via semantic knowledge transfer and self-refinement. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 4278–4287. [Google Scholar] [CrossRef]
  39. Dai, J.; He, K.; Sun, J. Boxsup: Exploiting bounding boxes to supervise convolutional networks for semantic segmentation. In Proceedings of the 015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 1635–1643. [Google Scholar] [CrossRef]
  40. Chen, M.; Zhang, Y.; Chen, E.; Hu, Y.; Xie, Y.; Pan, Z. Meta-Knowledge Guided Weakly Supervised Instance Segmentation for Optical and SAR Image Interpretation. Remote Sens. 2023, 15, 2357. [Google Scholar] [CrossRef]
  41. Kirillov, A.; Wu, Y.; He, K.; Girshick, R. Pointrend: Image segmentation as rendering. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 9799–9808. [Google Scholar]
  42. Chen, H.; Sun, K.; Tian, Z.; Shen, C.; Huang, Y.; Yan, Y. Blendmask: Top-down meets bottom-up for instance segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 8573–8581. [Google Scholar]
  43. Liu, Y.; Li, H.; Hu, C.; Luo, S.; Luo, Y.; Chen, C.W. Learning to aggregate multi-scale context for instance segmentation in remote sensing images. IEEE Trans. Neural Netw. Learn. Syst. 2024, 1–15, (Early Access). [Google Scholar] [CrossRef] [PubMed]
  44. Lu, X.; Wang, B.; Zheng, X.; Li, X. Exploring Models and Data for Remote Sensing Image Caption Generation. IEEE Trans. Geosci. Remote Sens. 2017, 56, 2183–2195. [Google Scholar] [CrossRef]
  45. Qu, B.; Li, X.; Tao, D.; Lu, X. Deep semantic understanding of high resolution remote sensing image. In Proceedings of the 2016 International Conference on Computer, Information and Telecommunication Systems (Cits), Kunming, China, 6–8 July 2016; pp. 1–5. [Google Scholar] [CrossRef]
Figure 1. DiffuPrompter classification images, with pixel-level annotations labeled by DiffuPrompter.
Figure 1. DiffuPrompter classification images, with pixel-level annotations labeled by DiffuPrompter.
Remotesensing 16 02004 g001
Figure 2. Pipeline for our method with the prompt ‘Photo of a stadium’. DiffuPrompter mainly includes three steps: (1) Organize the object name into the template and use it as a text prompt. (2) Object mask proposal generation. (3) The denoising strategy is applied to refine the proposals.
Figure 2. Pipeline for our method with the prompt ‘Photo of a stadium’. DiffuPrompter mainly includes three steps: (1) Organize the object name into the template and use it as a text prompt. (2) Object mask proposal generation. (3) The denoising strategy is applied to refine the proposals.
Remotesensing 16 02004 g002
Figure 3. Cross-attention maps of SDM. Text prompt: ‘Photo of an airplane’.
Figure 3. Cross-attention maps of SDM. Text prompt: ‘Photo of an airplane’.
Remotesensing 16 02004 g003
Figure 4. The cross-attention maps of natural and remote-sensing images.
Figure 4. The cross-attention maps of natural and remote-sensing images.
Remotesensing 16 02004 g004
Figure 5. Flow chart of LSAD.
Figure 5. Flow chart of LSAD.
Remotesensing 16 02004 g005
Figure 6. Visualization of denoising effects with different sampling times; t = 40 was applied to each sampling process. (a) visualizes the denoising effect on a natural image; (b) visualizes the denoising effect on a remote sensing image.
Figure 6. Visualization of denoising effects with different sampling times; t = 40 was applied to each sampling process. (a) visualizes the denoising effect on a natural image; (b) visualizes the denoising effect on a remote sensing image.
Remotesensing 16 02004 g006
Figure 7. Visualization of the DiffuPrompter mask generation process: (a) original image, (b) cross-attention map, (c) binarized map, (d) box and point prompts, (e) segmentation result.
Figure 7. Visualization of the DiffuPrompter mask generation process: (a) original image, (b) cross-attention map, (c) binarized map, (d) box and point prompts, (e) segmentation result.
Remotesensing 16 02004 g007
Figure 8. Visualization of cross-attention maps with different time steps. The L in LSAD is set to 50.
Figure 8. Visualization of cross-attention maps with different time steps. The L in LSAD is set to 50.
Remotesensing 16 02004 g008
Figure 9. Some examples of pseudo-labels generated by different methods.
Figure 9. Some examples of pseudo-labels generated by different methods.
Remotesensing 16 02004 g009
Figure 10. Cross-attention map with fine-tuned SDM.
Figure 10. Cross-attention map with fine-tuned SDM.
Remotesensing 16 02004 g010
Table 1. The performance of Mask R-CNN trained by pseudo-labels generated by DiffuPrompter vs. cross-attention with different λ thresholds.
Table 1. The performance of Mask R-CNN trained by pseudo-labels generated by DiffuPrompter vs. cross-attention with different λ thresholds.
Mask λ NWPUiSAID
AP AP 50 AP 75 AP AP 50 AP 75
Cross-Attention0.310.317.713.23.17.45.5
0.417.225.113.59.215.210.6
0.515.622.319.75.39.17.4
DiffuPrompter0.322.537.930.15.715.610.9
0.427.350.236.115.431.219.4
0.525.447.933.013.127.516.3
Table 2. The performance of Mask R-CNN trained on pure synthesis data under different L loop sampling times.
Table 2. The performance of Mask R-CNN trained on pure synthesis data under different L loop sampling times.
TimesNWPUiSAID
AP AP 50 AP 75 AP AP 50 AP 75
14.36.74.22.43.72.1
57.610.38.64.57.36.5
1010.312.510.55.49.67.8
2020.548.633.312.527.816.5
5027.350.236.115.431.219.4
10027.250.436.215.631.119.5
Table 3. The performance of Mask R-CNN and Cascade R-CNN on the NWPU. ‘P’ and ‘R’ refer to ‘Pseudo’ and ‘Real’.
Table 3. The performance of Mask R-CNN and Cascade R-CNN on the NWPU. ‘P’ and ‘R’ refer to ‘Pseudo’ and ‘Real’.
Training SetMethodSizeAP AP 50 AP 75 AP S AP M AP L
Training with Pure Real Label
NWPUMask R-CNNR: 0.8 k (all)58.390.260.740.956.661.1
Cascade R-CNNR: 0.8 k (all)59.891.966.645.360.067.3
Mask2FormerR: 0.8 k (all)61.392.568.646.362.769.5
Mask R-CNNR: 0.6 k50.283.155.431.248.357.5
Cascade R-CNNR: 0.6 k55.486.162.440.349.762.1
Mask2FormerR: 0.6 k56.287.164.142.553.864.2
Training with Pure Pseudo-Label
DiffuPrompterMask R-CNNP: 9.3 k27.350.236.117.225.135.4
Cascade R-CNNP: 9.3 k29.952.338.119.428.735.2
Mask2FormerR: 9.3 k30.353.440.121.330.537.2
Training with Pseudo and Real Label
DiffuPrompter & NWPUMask R-CNNP: 9.3 k R: 0.4 k50.685.354.232.453.358.6
Cascade R-CNNP: 9.3 k R: 0.4 k52.186.559.133.455.364.0
Mask2FormerP: 9.3 k R: 0.4 k54.688.764.240.953.662.1
Mask R-CNNP: 9.3 k R: 0.6 k55.689.360.237.456.361.6
Cascade R-CNNP: 9.3 k R: 0.6 k59.190.565.145.358.967.0
Mask2FormerP: 9.3 k R: 0.6 k60.992.666.946.061.868.2
Table 4. The performance of Mask R-CNN and Cascade R-CNN on the iSAID. Mask A P is for 15 classes. ‘P’ and ‘R’ refer to ‘Pseudo’ and ‘Real’.
Table 4. The performance of Mask R-CNN and Cascade R-CNN on the iSAID. Mask A P is for 15 classes. ‘P’ and ‘R’ refer to ‘Pseudo’ and ‘Real’.
Training SetMethodSizeAP AP 50 AP 75 AP S AP M AP L
Training with Pure Real Label
iSAIDMask R-CNNR: 2.8 k34.857.437.020.543.250.3
Cascade R-CNNR: 2.8 k35.657.838.020.844.352.7
Mask2FormerR: 2.8 k37.159.440.220.545.155.6
Mask R-CNNR: 2.2 k29.552.331.412.537.746.9
Cascade R-CNNR: 2.2 k31.251.431.613.539.747.2
Mask2FormerR: 2.2 k33.853.133.614.139.947.8
Training with Pure Pseudo-Label
DiffuPrompterMask R-CNNP: 9.3 k15.431.219.49.320.633.7
Cascade R-CNNP: 9.3 k16.132.521.511.122.736.2
Mask2FormerR: 9.3 k17.936.224.314.522.536.3
Training with Pseudo and Real Label
DiffuPrompter & iSAIDMask R-CNNP: 9.3 k R: 1.4 k31.654.332.415.341.748.4
Cascade R-CNNP: 9.3 k R: 1.4 k32.755.134.315.843.351.1
Mask2FormerP: 9.3 k R: 1.4 k33.456.435.316.345.754.5
Mask R-CNNP: 9.3 k R: 2.2 k35.157.336.417.342.750.4
Cascade R-CNNP: 9.3 k R: 2.2 k35.257.138.317.843.952.5
Mask2FormerP: 9.3 k R: 2.2 k36.858.738.917.945.255.3
Table 5. Performance of domain generalization between different datasets. Mask R-CNN with ResNet50 is used as the baseline.
Table 5. Performance of domain generalization between different datasets. Mask R-CNN with ResNet50 is used as the baseline.
Training SetData SizeTest SetAP AP 50 AP 75
NWPUR: 0.8 K (all)iSAID17.929.020.1
iSAIDR: 2.8 K (all)NWPU47.278.551.0
DiffuPrompter & NWPU9.3K + R: 0.6KiSAID22.540.323.0
DiffuPrompter & iSAID9.3K + R: 2.2KNWPU45.472.648.1
DiffuPrompter & NWPU9.3K + R: 0.8 K (all)iSAID25.245.126.0
DiffuPrompter & iSAID9.3K + R: 2.8 K (all)NWPU50.382.753.5
Table 6. Comparison of different AIA methods. The results are from segmentation methods trained on pseudo-labels constructed by these AIA methods. Seconds/im represents the time consumed by labeling each image with one V100 GPU.
Table 6. Comparison of different AIA methods. The results are from segmentation methods trained on pseudo-labels constructed by these AIA methods. Seconds/im represents the time consumed by labeling each image with one V100 GPU.
MethodSeconds/imNWPUiSAID
AP AP 50 AP 75 AP AP 50 AP 75
CAMMask R-CNN2.58.415.310.73.17.45.5
Mask2Former7.916.111.23.56.95.7
DiffuMaskMask R-CNN35.312.317.913.19.513.611.2
Mask2Former12.918.114.510.214.210.9
DiffuPrompterMask R-CNN297.327.350.236.115.431.219.4
Mask2Former30.353.440.117.936.224.3
Table 7. Performance comparison with some weakly supervised methods on the NWPU and iSAID datasets. ‘Sup’ refers to the supervision type.
Table 7. Performance comparison with some weakly supervised methods on the NWPU and iSAID datasets. ‘Sup’ refers to the supervision type.
MethodSupNWPUiSAID
AP AP 50 AP 75 AP AP 50 AP 75
CAM [35] I 8.415.310.73.17.45.5
SEC [36] I 21.430.023.14.212.710.3
AffinityNet [37] I 22.537.930.15.715.610.9
BESTIE [38] I 13.325.913.87.39.48.5
BoxSup [39] B 27.553.137.910.320.214.1
MGWI-Net [40] B 29.862.925.8---
DiffuPrompter M I 27.350.236.111.422.614.4
DiffuPrompter C I 29.952.338.113.125.516.3
Table 8. Comparison of class-wise results with advanced supervision methods on NWPU dataset.
Table 8. Comparison of class-wise results with advanced supervision methods on NWPU dataset.
MethodAPAISHSTBDTCBCPLHBBRVE
Mask R-CNN58.328.452.869.681.459.669.684.360.725.850.6
PANet [18]64.850.653.578.483.573.078.187.258.633.851.6
PointRend [41]65.454.553.275.784.372.474.490.158.835.954.7
BlendMask [42]65.748.151.179.884.072.476.791.558.939.654.6
CATNet [43]73.351.964.487.189.475.879.795.065.053.272.0
DiffuPrompter M 27.317.927.535.337.130.431.140.127.99.915.8
Table 9. Class-wise results of Mask R-CNN on the NWPU VHR-10 test set with different values of λ and t.
Table 9. Class-wise results of Mask R-CNN on the NWPU VHR-10 test set with different values of λ and t.
t λ APAISHSTBDTCBCPLHABRVE
300.312.38.911.58.622.511.321.615.99.69.14.1
0.420.516.719.715.718.422.628.526.427.811.917.3
0.518.516.020.719.719.425.012.919.223.910.617.5
400.322.527.730.021.130.019.728.022.121.810.514.1
0.427.317.927.535.337.130.431.140.127.99.915.8
0.525.435.035.020.135.018.918.935.030.812.712.6
500.319.523.714.225.021.124.021.111.323.025.06.4
0.423.730.619.126.332.028.829.433.016.78.712.4
0.520.132.430.520.924.415.921.024.616.88.56.2
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, H.; Wei, Y.; Peng, H.; Zhang, W. DiffuPrompter: Pixel-Level Automatic Annotation for High-Resolution Remote Sensing Images with Foundation Models. Remote Sens. 2024, 16, 2004. https://doi.org/10.3390/rs16112004

AMA Style

Li H, Wei Y, Peng H, Zhang W. DiffuPrompter: Pixel-Level Automatic Annotation for High-Resolution Remote Sensing Images with Foundation Models. Remote Sensing. 2024; 16(11):2004. https://doi.org/10.3390/rs16112004

Chicago/Turabian Style

Li, Huadong, Ying Wei, Han Peng, and Wei Zhang. 2024. "DiffuPrompter: Pixel-Level Automatic Annotation for High-Resolution Remote Sensing Images with Foundation Models" Remote Sensing 16, no. 11: 2004. https://doi.org/10.3390/rs16112004

APA Style

Li, H., Wei, Y., Peng, H., & Zhang, W. (2024). DiffuPrompter: Pixel-Level Automatic Annotation for High-Resolution Remote Sensing Images with Foundation Models. Remote Sensing, 16(11), 2004. https://doi.org/10.3390/rs16112004

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop