Next Article in Journal
Rapid Inspection of Large Concrete Floor Flatness Using Wheeled Robot with Aided-INS
Previous Article in Journal
Retrieving the Infected Area of Pine Wilt Disease-Disturbed Pine Forests from Medium-Resolution Satellite Images Using the Stochastic Radiative Transfer Theory
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Cycle and Self-Supervised Consistency Training for Adapting Semantic Segmentation of Aerial Images

1
Institute of Remote Sensing and Geographic Information System, Peking University, Beijing 100871, China
2
International Research Center of Big Data for Sustainable Development Goals, Beijing 100094, China
3
Key Laboratory of Digital Earth Science, Aerospace Information Research Institute, Chinese Academy of Sciences, Beijing 100094, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2022, 14(7), 1527; https://doi.org/10.3390/rs14071527
Submission received: 28 January 2022 / Revised: 3 March 2022 / Accepted: 18 March 2022 / Published: 22 March 2022

Abstract

:
Semantic segmentation is a critical problem for many remote sensing (RS) image applications. Benefiting from large-scale pixel-level labeled data and the continuous evolution of deep neural network architectures, the performance of semantic segmentation approaches has been constantly improved. However, deploying a well-trained model on unseen and diverse testing environments remains a major challenge: a large gap between data distributions in train and test domains results in severe performance loss, while manual dense labeling is costly and not scalable. To this end, we proposed an unsupervised domain adaptation framework for RS image semantic segmentation that is both practical and effective. The framework is supported by the consistency principle, including the cycle consistency in the input space and self-supervised consistency in the training stage. Specifically, we introduce cycle-consistent generative adversarial networks to reduce the discrepancy between source and target distributions by translating one into the other. The translated source data then drive a pipeline of supervised semantic segmentation model training. We enforce consistency of model predictions across target image transformations in order to provide self-supervision for the unlabeled target data. Experiments and extensive ablation studies demonstrate the effectiveness of the proposed approach on two challenging benchmarks, on which we achieve up to 9.95% and 7.53% improvements in accuracy over the state-of-the-art methods, respectively.

1. Introduction

Semantic segmentation of remote sensing (RS) images has attracted increasing attention and research interest. Many applications, such as environmental monitoring, crop production, and urban planning, need accurate and efficient segmentation mechanisms [1,2,3,4,5]. These demands coincide with the rise of deep learning methods in the RS field and application target-related RS image interpretation, including segmentation, object detection, and classification [6,7,8,9]. Semantic segmentation assigns a predefined ground truth label to each pixel in an image by clustering parts of an image that belong to the same object class [10], and usually applies end-to-end dense prediction networks to achieve pixelwise prediction. However, dense prediction architectures rely on pixel-level annotations of all categories to extract rich semantics and locate object boundaries accurately. Obtaining such dense annotations for semantic segmentation is notoriously laborious and expensive, and is the major limitation of semantic segmentation methods. Moreover, a trained model suffers from performance decreases in practical tasks due to the complexity of RS data, such as in the diversity of acquisition conditions, varied geospatial locations, and different ground sampling distances [11]. The requirement to develop generalizable models with unlabeled data has motivated work on deep domain adaptation (DA) approaches for semantic segmentation. DA is a special case of transfer learning that uses labeled data in one or more relevant source domains to implement new tasks in the target domain [12]. It aims to improve the generalization ability of the model across domains and realize the cross-domain transfer reuse of domain-invariant knowledge. DA methods can be classified into supervised, semisupervised, and unsupervised DA according to whether the target domain has labels. Obviously, unsupervised domain adaptation (UDA) is a promising approach for semantic segmentation and other tasks for which large-scale data annotation is costly and time-consuming.
The UDA methods for semantic segmentation have been extended to address issues in the RS community. According to existing works in the RS fields, UDA methods can roughly be categorized into three methods: generative-based method, adversarial learning, and self-training [13]. Specifically, generative-based methods translate source data to the visual appearances of a target domain by performing distribution alignment in the image space by using generative adversarial networks (GANs). Adversarial training aims to minimize the discrepancy between source and target feature distributions by introducing a discriminator network alongside the main segmentation network. Self-training-based methods minimize the loss of latent variables by alternatively generating pseudolabels on target data and retraining the model with these labels. Recent works on UDA for RS image semantic segmentation have primarily used generative-based methods [14,15,16,17,18,19,20,21,22]. To the best of our knowledge, Benjdira et al. [14] first addressed the problem of DA for RS image semantic segmentation by GANs. This study performed image-level translation from the source domain to the target domain using CycleGAN [23], and the results showed that the method is capable of substantially improving the accuracy of the segmentation model. Most recently, Li et al. [15] found that the DualGAN [24] is recommended to conduct unsupervised image translation between the source and target domains to carry out a weakly supervised transfer invariant constraint (WTIC). Tazar et al. [16] designed a novel GAN architecture named color mapping generative adversarial networks (ColorMapGAN), which can generate fake training images that are semantically the same as training images but with a spectral distribution similar to that of test images. Due to the architecturally simple (the generator has no convolution and pooling layers) but powerful design, ColorMapGAN performs well in both accuracy and computational complexity. Most studies mentioned above were implemented on a single space (e.g., input, feature, or output). In the study by [18], an end-to-end GAN-based full-space DA for classification was proposed. In this framework, the source and target images are fully aligned in the image, feature, and output spaces by different methods. Mateo-Garcia et al. [20] used CycleGAN to train a domain adaptation method between Proba-V and upscaled Landsat images. This was applied to train a cloud detection algorithm and eventually achieved higher cloud detection accuracy. Some researchers also used the CycleGAN approach for DA in deforestation detection in the Amazon Forest [21]. Kou et al. [22] proposed progressive domain adaptation (PDA) by embedding a convolutional long short-term memory (ConvLSTM) network into a conditional generative adversarial network (cGAN [25]) for change detection using season-varying remote sensing images.
In addition to the dominant generative-based methods, other methods have been widely explored in RS fields. Zhang et al. [26] introduced a layer alignment method and a feature covariance loss function. They also adopted a self-training method to fine-tune the model with pseudolabels on target data and further improve segmentation performance. Zhang et al. [27] proposed a curriculum-style local-to-global cross-domain adaptation framework for the segmentation of very-high-resolution (VHR) RS images. The result showed that the curriculum-style cross-domain adaptation (CCDA) strategy and local-to global feature alignment (LGFA) module achieved better performance on common benchmarks. Similar to this idea, Shen et al. [28] introduced a spatial frequency, detached the target domain into an easy and hard split, and then adopted intradomain adaptation by self-supervised training to improve the performance of the hard split. In addition, to be more efficient in addressing domain shifts on aerial datasets, Chen et al. [29] used entropy minimization and classwise distribution alignment to produce high-confidence prediction. In addition to extracting features from the image domain, Liu et al. [30] proposed a bispace alignment network for DA, named BSANet, which is able to extract features in the image and wavelet domains simultaneously. The dual-branch structure utilizing two discriminators minimizes the discrepancy in the fused feature space and output spaces.
While previous studies explored UDA for RS image semantic segmentation, challenges still exist: (i) the generative-based methods used above ignore the spatial scale difference of data across domains, and only one-sided translated images are used in the subsequent process, which results in a huge waste of computation; (ii) adversarial learning computationally involves adversarial objectives in different space that are highly sensitive to hyperparameters and are difficult to train; (iii) self-training methods do not sufficiently explore the intrinsic invariant characteristics of the target data caused by the data transformation. In this paper, we propose a cycle and self-supervised consistency (CSC) training framework. The three main contributions of our work can be summarized as follows:
  • We propose a novel UDA framework for RS image semantic segmentation that combines cycle-consistent image-to-image translation techniques and self-supervised transformation consistency training.
  • We investigate multiple transformation functions and enforce transformation consistency to provide supervision for self-supervised training of the model on unlabeled target data. These functions provide different perspectives on how to learn domain-invariant features for semantic segmentation.
  • Compared with previous methods, the framework we proposed achieves state-of-the-art performance on two challenging benchmarks of UDA for RS image semantic segmentation. Each consistency component independently outperformed some previous methods. This was verified by extensive ablation studies on our framework.
  • The proposed domain adaptation framework is easy to implement and is readily embedded in any existing semantic segmentation model to improve the prediction performance on unlabeled data.

2. Methods

In the following, we first give an overview of the proposed framework, and then introduce the ResiDualGAN to conduct unsupervised image translation between source and target domains to carry out cycle consistency training. Finally, we give the detailed descriptions of the proposed consistency training strategy.

2.1. Overview

In the setting of unsupervised domain adaptation for semantic segmentation, we consider a labeled source domain S and an unlabeled target domain T with the same set of semantic classes. In general, RS images from S and T are composed of different image attributes with different geographic locations, which leads to the cross-domain RS image semantic segmentation being more complex than other tasks in UDA fields. The goal of the purposed framework is to train a model on labeled domain S, and perform across-domain semantic segmentation tasks with minimal prediction error on the unlabeled target domain T. Intuitively, the overall workflow of the proposed consistency-based framework is shown in Figure 1, and is driven by consistency principles in different training stages. The framework consists of two consistency criteria: the cycle consistency across domains in the input space and the self-supervised transformation consistency in the training stage. A cycling generative adversarial network with scale change modules is trained in the input space to eliminate spectral- and spatial-scale discrepancies across domains by iteratively reconstructing and discriminating. A mapping of the data distribution between the source and target domains is established, and the target-stylized source domain data are produced to train a semantic segmentation model in a supervised way. In the subsequent training stage, a self-supervised training module is incorporated to enforce the consistency of target semantic maps produced by the model across a serious image transformation. Such consistency training is based on the idea that a strong model’s output should be consistent, with respect to the transformation of inputs, and enable the model to learn features from unlabeled target data in a self-supervised way.

2.2. Cycle Consistency

As mentioned in Section 1, GAN-based unpaired image translation algorithms have been widely used in cross-domain adaptation tasks for semantic segmentation. The original GANs can teacha generator to capture the distribution of real data by introducing an adversarial discriminator that evolves to discriminate between the real data and the fake. Therefore, it usually includes two neural networks, in which generative network G generates candidates while the discriminate network D distinguishes candidates produced by G from the true data distribution. With such a strategy, a bidirectional GAN architecture enables cross-domain image-to-image translation. Suppose there are two collections of images X S and X T from source domain S and target domain T, respectively. The generator G S T maps an image x S X S to an image x T X T , while the dual task is to train an inverse generator G T S . Discriminators D S and D T are trained to measure how well the generated candidates fit in the corresponding domains. This idea of implementing inverse mapping and having two generators and two discriminators is common among a series of GAN-family methods for image translation, such as DiscoGAN [31], CycleGAN, and DualGAN. For RS image translation, Li et al. [15] quantitatively verified that DualGAN performed better than other unsupervised methods (e.g., CycleGAN and DiscoGAN) in across-domain semantic segmentation tasks.
However, this line of work on RS image translation suffers from two main limitations in practical applications. First, the spatial scale discrepancy of images between two domains is not considered, while the scale factor is very important in RS image interpretation tasks. Previous works have proven that RS images with similar ground sampling distance (GSD) are easier to adapt across domains [19,32]. Second, cycle translation in appearance may locally change the semantic information of certain ground types; hence such semantic consistency should be ensured during translation. To solve the above problems, we introduce ResiDualGAN [32] to align the distribution in the image space. Based on DualGAN, ResiDualGAN redesigned the structure of generators and named it Resi-Generator, where a residual module and a resize module were incorporated (Figure 2). It is proposed specifically for optimizing the translation results of RS images with different spatial resolutions. For translation of X S X T , the Resi-Generator can be denoted as R e s i G S T , containing a regular generator G S T and a resize module R e s i z e S T . As Figure 2 shows, the design of R e s i G S T was inspired by the structure of skip connections (i.e., shortcuts) from the residual network [33]. The outputs of G S T are added as an identity mapping X S to enhance the original semantic information and then fed into R e s i z e S T to decrease the spatial resolution:
X S T = R e s i G S T ( X S ) = R e s i z e S T ( G S T ( X S ) + X S ) .
The mapping of X S X T :image x s X S is translated to domain X T using R e s i G S T . How well translation R e s i G S T ( x S ) fits in X T is evaluated by D T . R e s i G S T ( x S ) is then translated back to domain X S using R e s i G T S , which outputs R e s i G T S ( R e s i G S T ( x S ) ) as the reconstructed version of x S . Similarly, x T X T is translated to X S as R e s i G T S ( x T ) and then reconstructed as R e s i G S T ( R e s i G T S ( x T ) ) . Note that the reverse mapping X T X S and both D T and D S are not illustrated in Figure 1 for simplification.
The objective functions of ResiDualGAN consist of a feature-level adversarial loss and an image-level cycle loss. The adversarial loss function is integrated from loss functions used in D T and D S , which can be defined as
L a d v = ( D S ( R e s i G T S ( x T ) ) D T ( R e s i G S T ( x S ) ) ) .
Loss functions for generators R e s i G S T and R e s i G T S , however, are the same, as both share the same objective. The cycle loss function measured by the L 1 distance can be defined as
L c y c = ψ c y c ( | | x S R e s i G T S ( R e s i G S T ( x S ) ) | | + | | x T R e s i G S T ( R e s i G T S ( x T ) ) | | ) + ψ a d v ( D S ( R e s i G T S ( x T ) ) D T ( R e s i G S T ( x S ) ) ) ,
where ψ c y c and ψ a d v are two hyperparameters. By alternately minimizing L c y c and L a d v , cycle consistency enforces cross-domain consistent predictions in the context of image-to-image translation. The translated source and target data are used to train a semantic segmentation model in the next stage.

2.3. Self-Supervised Consistency

Our ultimate goal is to train a semantic segmentation model that is capable of high performance on unlabeled target domains. Cycle consistency reduces the distribution of data between the source domain and target domain. The semantic segmentation model can be trained on labeled, target-stylized source data and learn the features transferred from the target domain. However, the transferred features are insufficient for mapping real target data to ground truth. Inspired by approaches from self-training learning and semisupervised learning communities [34,35,36,37,38,39,40], a consistency training strategy is designed in this stage. This strategy is based on the assumption that the outputs from a well-trained model should be consistent across input transformations. For example, let f : X Y represent a pixelwise mapping from images x X to semantic output y Y . Denote t as a transformation function. We denote t p : X X as a photometric image transformation such as whitening or style transfer, and t q : X X as a spatial image transformation such as flipping or scaling. The consistency training is under the following hypothesis that for any image x X , f is invariant under t p : f ( t p ( x ) ) = f ( x ) and equivariant under t q : f ( t q ( x ) ) = t q ( f ( x ) ) . If x X T , then the f ( x ) can be seen as pseudo labeling and the prediction inconsistency cases provide self-supervision for target data. As shown in Figure 3, the training step is composed of a standard pipeline of a supervised segmentation network driven by labeled target-stylized source domain data and a self-supervised consistency training branch with pseudo labeling. In one training loop, a batch of target-stylized source images are fed in the model f and the predicted semantic maps are used to calculate supervised loss with labels. After that, a batch of unlabeled target images are fed into the model and pseudolabels are created by selecting confident pixels from the averaged map using thresholds. A transformation function t mentioned above is used to obtain a transformed version of the same pair of target images and pseudolabels. The consistency between these two version predictions compels the model to learn the features from the target domain in a self-supervised way. From another perspective, minimizing the unsupervised consistency loss progressively propagates semantic information from labeled data to unlabeled ones.
Concisely, consistency training methods simply regularize model predictions to be invariant to perturbations applied to either input samples or hidden states [40]. Under this framework, we are interested in what perturbations or transformations to apply are beneficial for cross-domain semantic segmentation of RS images. In this paper, we explore three types of transformation functions for such self-supervised consistency training:
  • Translation consistency: The GAN-based image translation outputs are bidirectional. However, in previous generative-based methods, source-stylized target images are useless in subsequent procedures, which results in a huge waste of computation. With the help of ResiDualGAN, the spatial scale and semantic information are well preserved during target-to-source stylization. Therefore, we regard such translation as transformation and enforce consistency between the model’s outputs on the original target images and the translated target image (Figure 4b).
  • Augmentation consistency: Data augmentation in model training is a technique to increase the amount of data by adding modified copies of already existing data. It acts as a regularizer and helps reduce overfitting when training a deep learning model. We randomly augment the target data by flipping, cropping, and brightness-changing. The corresponding augmentation is also performed on pseudolabels (Figure 4c).
  • CutMix consistency: CutMix is a regularization method designed for image classification and transfer learning [41]. We leverage this strategy to randomly cut and paste patches from source data to target data. The ground truth labels or pseudolabels are also mixed proportionally. The added patches on target data can be seen as a transformation for consistency training. Pasting patches from the source domain further increases confidence in pseudo labeling (Figure 4d).
In this paper, we adopt a combination objective function that combines a distribution-based cross-entropy loss and a region-based Dice loss. In this way, we can combine local with global information to improve the segmentation results. The loss function is defined as follows:
L C E _ D C ( y , y ^ ) = ( i c j N y i j log y ^ i j ) + ( 1 1 c i = 0 c j N 2 y i j y ^ i j + ϵ j N y i j + j N y ^ i j + ϵ ) ,
where y i j and y ^ i j denote the ith channel at the jth pixel location of the reference labels and neural network softmax output, respectively. We use c to denote the total channel count, N to denote the total pixel count in a mini-batch, and ϵ as a small constant plugged to avoid numerical problems. In addition to the supervised loss, we use transformation functions to obtain transformed pairs of target data and pseudolabels to calculate consistency loss by the same loss function. Therefore, the final loss L s e g involves the source supervised loss and the consistency training loss, which can be defined as follows:
L s e g = ( 1 λ ) L C E _ D C ( y S T , y ^ S T ) + λ L C E _ D C ( y T , y ^ T ) ,
where y S T and y ^ S T are labels and predictions of target-stylized source data, respectively. Meanwhile, y T and y ^ T are transformed pseudo labels and predictions of transformed target data, respectively. λ is a hyperparameter for weighting the two loss terms.

3. Experiments

In this section, we describe the datasets and experimental implementation details first and then we give the experimental results.

3.1. Datasets

We evaluate our framework on the two standard large-scale RS image semantic segmentation datasets Potsdam and Vaihingen [42]. These two common datasets are published by the International Society for Photogrammetry and Remote Sensing (ISPRS) as benchmarks of a 2D semantic labeling contest, providing airborne images with very high-resolution true orthophotos and labeled ground truth. The semantic categories from both datasets are the same and have been defined in the ranges of impervious surfaces, buildings, low vegetation, trees, cars, and clutter/background. Both areas cover urban scenes captured from different geographic locations (the city of Potsdam and Vaihingen) with different geospatial distributions. Potsdam shows a typical historic city with large building blocks, narrow streets, and dense settlement structures, while Vaihingen is small with many detached buildings and small multistory buildings. There are also differences in the imaging modes, including the channels and resolution. Potsdam datasets contain subsets with different spectral band combinations such as IR-R-G and R-G-B, while images from Vaihingen datasets have only IR-R-G channels. Moreover, the ground sampling distance or spatial resolution of the Potsdam datasets is 5 cm, while that of the Vaihingen datasets is 9 cm.
The similarities and differences mentioned above imply that these two datasets are quite suitable to evaluate the performance of UDA methods because, as two domains, they are the same in task objective and label space but different in data distribution and feature space. Li et al. [15] first developed them as new benchmarks of unsupervised domain adaptation for RS image semantic segmentation, evaluating the performance of the UDA model from the perspective of variation in geographic location and imaging modes. Since then, some work has been evaluated using this benchmark [15,27,32]. In this paper, we continue to adopt this established evaluation protocol. Images and labels from the Potsdam datasets serve as the source data, and the images from the Vaihingen datasets serve as the target data without available semantic labels. The two domain adaptation scenarios in the following are denoted as Potsdam IRRG → Vaihingen IRRG and Potsdam RGB → Vaihingen IRRG, respectively. The former represents the scenario in which domain adaptation tasks are across domains with different geographic regions. The latter crosses domains with different geographic regions plus different spectral combinations. Note that both adaptation scenarios involve the difference in ground sample distance.

3.2. Implementation Details

We implement the proposed framework in PyTorch [43] and adopt DeepLabv3+ [44] with ResNet-34 backbone as the segmentation architecture. Backbone initializes from the models pretrained on ImageNet. In the first stage, we implement cycle consistency using ResiDualGAN to translate the data from both domains. The loss parameters ψ c y c and ψ a d v are set to 10 and 5 for training. Based on the ratio of the spatial resolution of the data in the two domains, Potsdam IRRG and Potsdam RGB are cropped into the size of 896 × 896 while images from the target domain are cropped into the size of 512 × 512 . Finally, a total of 2368 Potsdam images and labels serve as source data for translation and training. The Vaihingen datasets contain 1697 images as target data for self-supervised training stage, 440 of which are used for validation. ResiDualGAN applies the color rendering and spatial resolution style of Vaihingen datasets (target domain) to higher resolution images of Potsdam datasets (the reverse translation occurred at the same time). We obtained the Vaihingen stylized Potsdam IRRG and Potsdam RGB images with downscale size of 512 × 512 . In the self-supervised transformation consistency training stage, we train the model with labeled target-stylized source data in a supervised way. Joint training proceeds in a self-supervised manner with unlabeled target data. The total gradient is calculated after alternating source-target forward passes to update the weights of the model. The batch size of 16 comprises eight target-stylized images and eight real target images at a size of 512 × 512 , which is a common practice. The optimizer is Adam and the initial learning rate is set to 1 × 10 4 and is scaled down by a factor of 0.5, according to the patience, which is a number of epochs with no improvement. To investigate the effects of each component and hyperparameter of our framework on performance, we present an extensive set of ablation studies. All experiments are conducted on a machine equipped with an Intel Core i7-7800X CPU, 16 GB of RAM, and one NVIDIA GeForce RTX 2080 Ti GPU with 11 GB of VARM to accelerate the training process. We measure the segmentation accuracy with per-class intersection-over-union (IoU) and F1-score, following recent works [15,27]. We denote the number of true positives, false positives, and false negatives by T P , F P , and F N respectively. The formulation of IoU and F1-score can be written as follows:
IoU = T P T P + F P + F N ,
F 1 - s c o r e = 2 × T P 2 T P + F P + F N .
The F1-score tends to measure something closer to average performance, while the IoU score measures something closer to the worst-case performance.

3.3. Performance of the CSC

We present our experimental results of the proposed CSC training framework on the Potsdam RGB → Vaihingen IRRG and Potsdam IRRG → Vaihingen IRRG benchmarks in Table 1 and Table 2. Depending on the transformation functions involved, our methods in Table 1 and Table 2 are denoted as CSC-Trans, CSC-Aug, and CSC-CutMix, respectively. The qualitative results of these approaches can be seen in Figure 4 and Figure 5. In both domain adaptation scenarios, augmentation-based consistency performs best, followed by CutMix-based and translation-based consistency. We adopt DeepLabv3+ as the baseline segmentation model to show the practical performance of the segmentation model in the presence of a domain gap. The baseline model is trained on labeled source data and tested on the unlabeled target datasets. Compared to the baseline, the CSC improves the accuracy of the segmentation result on target domain by up to 83% on Potsdam RGB → Vaihingen IRRG and 65% on Potsdam IRRG → Vaihingen IRRG. The solid performance of augmentation-based consistency suggests that the simple geometric or photometric transformations provide more significant equivariant constraints to the model than CutMix and image translation. CutMix-based consistency involves patches from source domain data, which provide high-confidence pseudo labeling while facilitating model-learning domain-invariant features. Note that in the scenario of the Potsdam IRRG → Vaihingen IRRG, the CutMix-based method has a significantly higher segmentation accuracy for the category of cars than the other methods. Translation-based consistency performs the worst relative to the other two transformation methods of our framework. As shown in column (d) of Figure 5 and Figure 6, the results of the translation-based method misclassified many regular objects into the class of clutter. Such poor performance is particularly evident in the average accuracy of the Potsdam RGB → Vaihingen IRRG. This implies that the reverse appearance translation as a transformation function for self-supervised consistency training is too intense for domains with vastly different imaging modes. By comparing the results of the baselines, methods on the domains with the same spectral band combination can bring an 18% performance improvement. This is another verification that aligning the domains directly in image space is effective for adapting domains with different imaging modes.

3.4. Comparison to State-of-the-Art Methods

Since we adopt an established evaluation protocol, including the same test datasets and metrics, we compare our framework to previous work from the benchmark list. The methods involved in the comparison include an adversarial learning method (AdaptSegNet [45]), an image-to-image translation-based method (GAN-RSDA [14]), a curriculum learning method (CCDA [27]), and two hybrid methods (MUCSS [15] and RDG-OSA [32]). Among them, GAN-RSDA, MUCSS, CCDA, and RDG-OSA are specifically designed for RS image semantic segmentation across domains, and the AdaptSegNet is designed for natural images. AdaptSegNet aligns the distribution of features in the output space using a discriminator network, which is a representative adversarial learning in the UDA field. GAN-RSDA is the pioneer in the use of image-to-image translation techniques in UDA for RS images and introduces the CycleGAN for appearance adaptation in image space. MUCSS goes a step further from the GAN-RSDA by using the more effective DualGAN and involving a self-training strategy and consistency regularization. CUSS designs a curriculum-style local-to-global cross-domain adaptation method to learn domain-invariance features. RDG-OSA combines the ResiDualGAN with an output space adaptation method and claims the best published results. For fairness, all the methods we reproduce in experiments employ the same segmentation model and training setup as our proposed framework.
As shown in Table 1 and Table 2, our proposed framework substantially outperforms the other methods and sets a new standard in terms of IoU and F1-score. Compared with the state-of-the-art methods, our framework increases the IoU and F1-score of segmentation results by 9.95% and 6.61% on Potsdam RGB → Vaihingen IRRG, respectively, while by 7.53% and 5.02% on Potsdam IRRG → Vaihingen IRRG. From the perspective of the adaptation scenario, the average performance improvement on the Potsdam RGB → Vaihingen IRRG is superior to that on the Potsdam IRRG → Vaihingen IRRG. For example, the mIoU is improved by 37 % compared to MUCSS on Potsdam RGB → Vaihingen IRRG and shows an up to 28% improvement on Potsdam IRRG → Vaihingen IRRG. This phenomenon does not include the CCDA, which ignores the class of clutter on the Potsdam IRRG → Vaihingen IRRG. Specific to the accuracy comparison of each class, CCDA achieves the best accuracy of clutter on the Potsdam RGB → Vaihingen IRRG while MUCSS achieves the best accuracy on the Potsdam IRRG → Vaihingen IRRG. In addition, augmentation-consistency-based CSC achieves the best accuracy in other categories in both scenarios. The proposed framework shows strong robustness and great generalization capability. Figure 7 and Figure 8 present a few qualitative examples, comparing our framework to the baseline and previous methods. Particularly prominent are the refinements of the small-scale elements such as cars and clutters, which may benefit from spatial resolution alignment in the cycle consistency. Moreover, the boundaries of the segments in our results are more precise by enforcing self-supervised consistency with transformation function, which makes our framework less prone to the contextual bias, leading to coarse boundaries.

4. Discussion

In this section, we first discuss the effects of components and hyperparameters sensitivity of the proposed CSC training framework. Then, we discuss the limitations of the proposed method and possible further improvements.

4.1. Ablation Study

In this subsection, we examine an extensive set of ablation studies to determine what makes our framework effective. These studies concern each component of the proposed framework and hyperparameters, such as the loss function ratio and pseudolabel threshold. Note that the self-supervised module of the full framework (cycle + self-supervised consistency) in this subsection is augmentation-consistency-based. First, we analyze the contribution of two components: cycle consistency and self-supervised consistency. We independently switch off each component and report the results in Table 3. On the Potsdam RGB → Vaihingen IRRG, compared to the full framework, disabling the self-supervised consistency module (transformation functions) leads to a 12% IoU decrease, while abolishing cycle consistency (without ResiDualGAN) leads to a drop of 34% IoU. For similar cases in the Potsdam IRRG → Vaihingen IRRG scenario, the mIoU decreases by 7% and 17%, respectively.
These statistics suggest that both cycle consistency and self-supervised consistency play a critical role in the entire framework. The contribution of cycle consistency is larger, especially in domain adaptation scenarios where there are vast differences in appearance or imaging mode. It is worth mentioning that the cycle-consistency-only cases in our methods significantly outperform CycleGAN-based GAN-RSDA and DualGAN-based MUCSS. This demonstrates the improvement in accuracy brought by the resize-residual module in the ResiDualGAN. To demonstrate the effectiveness of cycle consistency in reducing differences in data distribution, we randomly selected 500 images in each of two adaptation scenarios and extracted the high-dimensional features in the latent space of the backbone for semantic segmentation. These high-dimensional features are reduced to the two-dimensional space for visualization. As shown in Figure 9, the ResiDualGAN well matches the data distribution of the source domain data with the target domain data. The feature distribution of most translated images is similar to the target domain data feature distribution. In addition, it can be seen from Figure 9 that the matching difficulty of the data distribution between Potsdam IRRG and Vaihingen IRRG is less than the matching difficulty of Potsdam RGB and Vaihingen IRRG, which again illustrates the impact of imaging mode differences on domain adaptation.
Figure 10 illustrates the contribution of each component to the accuracy of individual classes on the Potsdam IRRG → Vaihingen IRRG. For most categories, the contribution of the different components to their accuracy is similar to that of the overall accuracy. However, for the impervious surface and clutter, there is little difference between the cycle-consistency-only-based method and the self-supervised-only-based method for accuracy improvement. In addition, the combination of cycle consistency and self-supervision gives the car an almost negligible improvement in accuracy.
Next, we analyze the sensitivity of hyperparameters λ and τ (Figure 11). Our proposed framework involves tuning the single hyperparameter λ , which trades off between the strength of the source supervised loss and self-supervised consistency loss. We present results for λ from the range of 0.1–0.9 and find that the performance of the model decreases gradually as λ grows. Similarly, we present results for the pseudo label threshold τ from the range of 0.1–0.9. We find that the accuracy starts to decrease significantly when the threshold τ > 0.5 . It should be noted that due to the datasets consist of 6 classes, τ = 0.1 means there is no thresholding in the training process. We think that the decrease in accuracy with increasing pseudo label thresholds τ is due to the loss being dominated by easy classes when a high threshold is set.

4.2. Computational Complex Analysis

In Section 3, related domain adaptation methods involved in the comparison include an adversarial learning method (AdaptSegNet), a generative-based method (GAN-RSDA), a curriculum learning method (CCDA), and two hybrid methods (MUCSS and RDG-OSA).The complete framework of CSC can also be viewed as a hybrid of generative methods and self-supervised training. Since the objective of domain adaptation methods is to improve the performance of a semantic segmentation model on unlabeled target domain, the computational complexity of all methods involved in the comparison is consistent in the model inference. In this paper, the adapted semantic segmentation model in the experiments is DeepLabv3+ with with ResNet-34 backbone. However, the computational complexity of various domain adaptation frameworks has significant differences in the training step. Generative-based methods and hybrid methods that include generative methods are much more computationally complex than other methods due to the existence of cycle generative adversarial networks. For example, the network structure of ResiDualGAN contains two generators and two discriminators. The numbers of parameters for each generator and discriminator are 41.82 M and 6.96 M, respectively, and the multiply-accumulate operations (MACs) for each generator and discriminator are 144.44 G and 7.74 G, respectively. Similarly, domain adaptation methods based on adversarial learning and curriculum learning introduce additional neural network structures, resulting in higher computational complexity than self-supervised methods with shared parameters. In summary, the complete CSC training framework is similar to other generative methods in terms of the number of parameters and the computational complexity during the training step. When the self-supervised module of CSC is used alone, since there is no additional network structure, the computational complexity and the amount of parameters are smaller than other methods.

4.3. Limitations

We have verified that the proposed cycle and self-supervised consistency training framework with different transformation functions performs well on UDA of RS image semantic segmentation, but there are many issues that remain unexplored and undeveloped in our proposed framework. Although we reduced the loss of semantic information of images during reconstruction by ResiDualGAN, we still cannot completely avoid the error of local semantic information. Such errors are exacerbated when the source and target domains have different data imaging modes. The overall adaptation effect on the Potsdam IRRG → Vaihingen IRRG is better than that on the Potsdam RGB → Vaihingen IRRG. This phenomenon arises because the latter needs to adapt not only the domain gap posed by geographical location but also the gap posed by differences in imaging modes. For example, in the Potsdam RGB to Vaihingen IRRG translation, parts of the building roofs in the Potsdam datasets that appear close to red and have complex textures are generated as trees or low vegetation. This semantic error affects the supervised training of the source domain data in the subsequent process. In addition, the differences in the label space of domains can lead to additional problems in the unsupervised domain adaptation process. The limitations of the segmentation accuracy for all methods in this paper occur in the class of clutter/background. The official documents define this class as water, clutters, and others. However, there exists a perception difference of “clutter or background” in the manual labeling of the two datasets. Many previous studies of UDA of semantic segmentation for RS images excluded this class from the training and validation procedure [19]. Finally, we do not analyze the individual contribution of the transformation function in augmentation-based consistency or whether there exists an optimal combination of target data transformation functions when faced with certain domain adaptation scenarios. The problem of class imbalance still exists in that our loss function does not include parameters to weight classes, nor does the self-supervised consistency training step set classwise thresholds. These are issues that need to be addressed and will be explored in our subsequent work.

5. Conclusions

We addressed the task of unsupervised domain adaptation for RS image semantic segmentation by proposing a simple but effective training framework (CSC) that unifies cycle consistency with self-supervised consistency. The CSC leverages cycle-consistent generative adversarial networks to reduce distribution discrepancies between the source and target domains. To further improve the performance on the target domain, self-supervision is embedded into a supervised learning framework by consistency training, which forces the model predictions from various transformed target images to be consistent. Compared to other UDA methods for RS image semantic segmentation, our framework achieved state-of-the-art performance on the two representative benchmarks. Based on an extensive set of ablation studies, we believe that each consistency component can work independently for UDA, and the ensemble of the two consistency components further improves the performance. Moreover, this method can be embedded in various types of semantic segmentation domain adaptation methods to solve the problem of performance degradation of semantic segmentation models between datasets with large differences in spectrum, resolution, and geographic distribution. Future work will focus on different combinations of transformation functions in the self-supervised step and exploring other transformation functions in certain UDA scenarios.

Author Contributions

Conceptualization, H.G.; methodology, H.G., Y.Z.; software, H.G.; validation, H.G. and Y.Z.; formal analysis, H.G. and Y.T.; investigation, Z.S.; resources, H.G.; data curation, P.G.; writing—original draft preparation, H.G. and P.G.; writing—review and editing, Z.S. and Y.T.; visualization, H.G.; supervision, X.C.; project administration, X.C.; funding acquisition, Y.T. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Key R&D Program of China, grant number 2021YFB3900503; the National Natural Science Foundation of China, grant number 41972308, 42071312, 42171291.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The Potsdam and Vaihingen datasets are published by the International Society for Photogrammetry and Remote Sensing (ISPRS) and can be accessed at https://www.isprs.org/education/benchmarks/UrbanSemLab/Default.aspx, accessed on 27 January 2022.

Acknowledgments

Our sincere gratitude goes to the anonymous reviewers for the constructive comments and suggestions that have helped improve this paper substantially.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Audebert, N. Beyond RGB: Very high resolution urban remote sensing with multimodal deep networks. ISPRS J. Photogramm. Remote Sens. 2018, 140, 20–32. [Google Scholar] [CrossRef] [Green Version]
  2. Zhou, K.; Chen, Y.; Smal, I.; Lindenbergh, R. Building segmentation from airborne VHR images using Mask R-CNN. ISPRS Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2019, 42, 155–161. [Google Scholar] [CrossRef] [Green Version]
  3. Chen, K.; Fu, K.; Yan, M.; Gao, X.; Sun, X.; Wei, X. Semantic Segmentation of Aerial Images With Shuffling Convolutional Neural Networks. IEEE Geosci. Remote Sens. Lett. 2018, 15, 5. [Google Scholar] [CrossRef]
  4. Huang, J.; Zhang, X.; Xin, Q.; Sun, Y.; Zhang, P. Automatic building extraction from high-resolution aerial images and LiDAR data using gated residual refinement network. ISPRS J. Photogramm. Remote Sens. 2019, 151, 91–105. [Google Scholar] [CrossRef]
  5. Kemker, R.; Salvaggio, C.; Kanan, C. Algorithms for semantic segmentation of multispectral remote sensing imagery using deep learning. ISPRS J. Photogramm. Remote Sens. 2018, 145, 60–77. [Google Scholar] [CrossRef] [Green Version]
  6. Zhang, L.; Zhang, L.; Du, B. Advances in Machine Learning for Remote Sensing and Geosciences. IEEE Geosci. Remote Sens. Mag. 2016, 19, 22–40. [Google Scholar] [CrossRef]
  7. Zhu, X.X.; Tuia, D.; Mou, L.; Xia, G.S.; Zhang, L.; Xu, F.; Fraundorfer, F. Deep Learning in Remote Sensing: A Comprehensive Review and List of Resources. IEEE Geosci. Remote Sens. Mag. 2017, 5, 8–36. [Google Scholar] [CrossRef] [Green Version]
  8. Audebert, N.; Le Saux, B.; Lefèvre, S. Segment-before-Detect: Vehicle Detection and Classification through Semantic Segmentation of Aerial Images. Remote Sens. 2017, 9, 368. [Google Scholar] [CrossRef] [Green Version]
  9. Zhao, W.; Du, S.; Wang, Q.; Emery, W.J. Contextually guided very-high-resolution imagery classification with semantic segments. ISPRS J. Photogramm. Remote Sens. 2017, 132, 48–60. [Google Scholar] [CrossRef]
  10. Garcia-Garcia, A.; Orts-Escolano, S.; Oprea, S.; Villena-Martinez, V.; Garcia-Rodriguez, J. A Review on Deep Learning Techniques Applied to Semantic Segmentation. arXiv 2017, arXiv:1704.06857. [Google Scholar]
  11. Tuia, D.; Persello, C.; Bruzzone, L. Domain Adaptation for the Classification of Remote Sensing Data: An Overview of Recent Advances. IEEE Geosci. Remote Sens. Mag. 2016, 4, 41–57. [Google Scholar] [CrossRef]
  12. Wang, M.; Deng, W. Deep Visual Domain Adaptation: A Survey. arXiv 2018, arXiv:1802.03601. [Google Scholar] [CrossRef] [Green Version]
  13. Toldo, M.; Maracani, A.; Michieli, U.; Zanuttigh, P. Unsupervised Domain Adaptation in Semantic Segmentation: A Review. Technologies 2020, 8, 35. [Google Scholar] [CrossRef]
  14. Benjdira, B.; Bazi, Y.; Koubaa, A.; Ouni, K. Unsupervised Domain Adaptation Using Generative Adversarial Networks for Semantic Segmentation of Aerial Images. Remote Sens. 2019, 11, 1369. [Google Scholar] [CrossRef] [Green Version]
  15. Li, Y.; Shi, T.; Zhang, Y.; Chen, W.; Wang, Z.; Li, H. Learning deep semantic segmentation network under multiple weakly-supervised constraints for cross-domain remote sensing image semantic segmentation. ISPRS J. Photogramm. Remote Sens. 2021, 175, 20–33. [Google Scholar] [CrossRef]
  16. Tasar, O.; Happy, S.L.; Tarabalka, Y.; Alliez, P. ColorMapGAN: Unsupervised Domain Adaptation for Semantic Segmentation Using Color Mapping Generative Adversarial Networks. IEEE Trans. Geosci. Remote Sens. 2020, 58, 7178–7193. [Google Scholar] [CrossRef] [Green Version]
  17. Tasar, O.; Giros, A.; Tarabalka, Y.; Alliez, P.; Clerc, S. DAugNet: Unsupervised, Multisource, Multitarget, and Life-Long Domain Adaptation for Semantic Segmentation of Satellite Images. IEEE Trans. Geosci. Remote Sens. 2021, 59, 1067–1081. [Google Scholar] [CrossRef]
  18. Ji, S.; Wang, D.; Luo, M. Generative Adversarial Network-Based Full-Space Domain Adaptation for Land Cover Classification From Multiple-Source Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2021, 59, 3816–3828. [Google Scholar] [CrossRef]
  19. Wittich, D.; Rottensteiner, F. Appearance based deep domain adaptation for the classification of aerial images. ISPRS J. Photogramm. Remote Sens. 2021, 180, 82–102. [Google Scholar] [CrossRef]
  20. Mateo-García, G.; Laparra, V.; López-Puigdollers, D.; Gómez-Chova, L. Cross-Sensor Adversarial Domain Adaptation of Landsat-8 and Proba-V Images for Cloud Detection. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 747–761. [Google Scholar] [CrossRef]
  21. Soto, P.J.; Costa, G.A.O.P.; Feitosa, R.Q.; Happ, P.N.; Ortega, M.X.; Noa, J.; Almeida, C.A.; Heipke, C. Domain adaptation with cyclegan for change detection in the Amazon forest. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2020, 43, 1635–1643. [Google Scholar] [CrossRef]
  22. Kou, R.; Fang, B.; Chen, G.; Wang, L. Progressive Domain Adaptation for Change Detection Using Season-Varying Remote Sensing Images. Remote Sens. 2020, 12, 3815. [Google Scholar] [CrossRef]
  23. Zhu, J.Y.; Park, T.; Isola, P.; Efros, A.A. Unpaired Image-to-Image Translation using Cycle-Consistent Adversarial Networks. arXiv 2020, arXiv:1703.10593. [Google Scholar]
  24. Yi, Z.; Zhang, H.; Tan, P.; Gong, M. DualGAN: Unsupervised Dual Learning for Image-to-Image Translation. arXiv 2018, arXiv:1704.02510. [Google Scholar]
  25. Mirza, M.; Osindero, S. Conditional Generative Adversarial Nets. arXiv 2014, arXiv:1411.1784. [Google Scholar]
  26. Zhang, Z.; Doi, K.; Iwasaki, A.; Xu, G. Unsupervised Domain Adaptation of High-Resolution Aerial Images via Correlation Alignment and Self Training. IEEE Geosci. Remote Sens. Lett. 2021, 18, 746–750. [Google Scholar] [CrossRef]
  27. Zhang, B.; Chen, T.; Wang, B. Curriculum-Style Local-to-Global Adaptation for Cross-Domain Remote Sensing Image Segmentation. IEEE Trans. Geosci. Remote Sens. 2021. [Google Scholar] [CrossRef]
  28. Shen, W.; Wang, Q.; Jiang, H.; Li, S.; Yin, J. Unsupervised domain adaptation for semantic segmentation via self-supervision. In Proceedings of the 2021 IEEE International Geoscience and Remote Sensing Symposium IGARSS, Brussels, Belgium, 11–16 July 2021; pp. 2747–2750. [Google Scholar] [CrossRef]
  29. Chen, Y.; Ouyang, X.; Zhu, K.; Agam, G. Domain Adaptation on Semantic Segmentation for Aerial Images. arXiv 2020, arXiv:2012.02264. [Google Scholar]
  30. Liu, W.; Su, F.; Jin, X.; Li, H.; Qin, R. Bispace Domain Adaptation Network for Remotely Sensed Semantic Segmentation. IEEE Trans. Geosci. Remote Sens. 2020, 60, 1–11. [Google Scholar] [CrossRef]
  31. Kim, T.; Cha, M.; Kim, H.; Lee, J.K.; Kim, J. Learning to Discover Cross-Domain Relations with Generative Adversarial Networks. arXiv 2017, arXiv:1703.05192. [Google Scholar]
  32. Zhao, Y.; Gao, H.; Guo, P.; Sun, Z. ResiDualGAN: Resize-Residual DualGAN for Cross-Domain Remote Sensing Images Semantic Segmentation. arXiv 2022, arXiv:2201.11523. [Google Scholar]
  33. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. arXiv 2015, arXiv:1512.03385. [Google Scholar]
  34. Melas-Kyriazi, L.; Manrai, A.K. PixMatch: Unsupervised Domain Adaptation via Pixelwise Consistency Training. arXiv 2021, arXiv:2105.08128. [Google Scholar]
  35. Araslanov, N.; Roth, S. Self-supervised Augmentation Consistency for Adapting Semantic Segmentation. arXiv 2021, arXiv:2105.00097. [Google Scholar]
  36. Chen, T.; Kornblith, S.; Norouzi, M.; Hinton, G. A Simple Framework for Contrastive Learning of Visual Representations. In Proceedings of the 37th International Conference on Machine Learning, PMLR, Virtual Event, 12–18 July 2020; pp. 1597–1607. [Google Scholar]
  37. Berthelot, D.; Carlini, N.; Goodfellow, I.; Papernot, N.; Oliver, A.; Raffel, C. MixMatch: A Holistic Approach to Semi-Supervised Learning. arXiv 2019, arXiv:1905.02249. [Google Scholar]
  38. Berthelot, D.; Carlini, N.; Cubuk, E.D.; Kurakin, A.; Sohn, K.; Zhang, H.; Raffel, C. ReMixMatch: Semi-Supervised Learning with Distribution Alignment and Augmentation Anchoring. arXiv 2020, arXiv:1911.09785. [Google Scholar]
  39. Sohn, K.; Berthelot, D.; Li, C.L.; Zhang, Z.; Carlini, N.; Cubuk, E.D.; Kurakin, A.; Zhang, H.; Raffel, C. FixMatch: Simplifying Semi-Supervised Learning with Consistency and Confidence. Adv. Neural Inf. Process. Syst. 2020, 33, 596–608. [Google Scholar]
  40. Xie, Q.; Dai, Z.; Hovy, E.; Luong, T.; Le, Q. Unsupervised data augmentation for consistency training. Adv. Neural Inf. Process. Syst. 2020, 33, 6256–6268. [Google Scholar]
  41. Yun, S.; Han, D.; Oh, S.J.; Chun, S.; Choe, J.; Yoo, Y. Cutmix: Regularization strategy to train strong classifiers with localizable features. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27–28 October 2019; pp. 6023–6032. [Google Scholar]
  42. Gerke, M. Use of the Stair Vision Library within the ISPRS 2D Semantic Labeling Benchmark (Vaihingen); University of Twente: Enschede, The Netherlands, 2015. [Google Scholar] [CrossRef]
  43. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. PyTorch: An Imperative Style, High-Performance Deep Learning Library. Adv. Neural Inf. Process. Syst. 2019, 32. Available online: https://proceedings.neurips.cc/paper/2019/hash/bdbca288fee7f92f2bfa9f7012727740-Abstract.html (accessed on 2 January 2022).
  44. Chen, L.C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-Decoder with Atrous Separable Convolution for Semantic Image Segmentation. arXiv 2018, arXiv:1802.02611. [Google Scholar]
  45. Tsai, Y.H.; Hung, W.C.; Schulter, S.; Sohn, K.; Yang, M.H.; Chandraker, M. Learning to adapt structured output space for semantic segmentation. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 7472–7481. [Google Scholar] [CrossRef] [Green Version]
  46. Van der Maaten, L.; Hinton, G. Visualizing data using t-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
Figure 1. Overview of proposed framework. R e s i G S T and R e s i G T S are generators of the ResiDualGAN between two domains. L c y c and L a d v are cycle loss function and adversarial loss function, respectively. L s r c and L c o n are segmentation loss function for source data and target data, respectively.
Figure 1. Overview of proposed framework. R e s i G S T and R e s i G T S are generators of the ResiDualGAN between two domains. L c y c and L a d v are cycle loss function and adversarial loss function, respectively. L s r c and L c o n are segmentation loss function for source data and target data, respectively.
Remotesensing 14 01527 g001
Figure 2. Resi-Generator of ResiDualGAN.
Figure 2. Resi-Generator of ResiDualGAN.
Remotesensing 14 01527 g002
Figure 3. Flowchart of self-supervised consistency training: x S T and y S T are target-stylized source image and label; x T and y T are target image and pseudolabel. f denotes the pixelwise mapping from images to semantic output; t is the transformation function; x T and y T are the transformed target image and pseudolabel; y ^ S T and y ^ T are the semantic outputs of the f.
Figure 3. Flowchart of self-supervised consistency training: x S T and y S T are target-stylized source image and label; x T and y T are target image and pseudolabel. f denotes the pixelwise mapping from images to semantic output; t is the transformation function; x T and y T are the transformed target image and pseudolabel; y ^ S T and y ^ T are the semantic outputs of the f.
Remotesensing 14 01527 g003
Figure 4. Transformation of images from target domain for consistency training. (a) Target domain images without transformation. (b) Source-stylized target images. (c) Augmented target images. (d) Target images mixed by source image patches.
Figure 4. Transformation of images from target domain for consistency training. (a) Target domain images without transformation. (b) Source-stylized target images. (c) Augmented target images. (d) Target images mixed by source image patches.
Remotesensing 14 01527 g004
Figure 5. Qualitative results of cross-domain semantic segmentation on Potsdam RGB → Vaihingen IRRG. (a) Target images. (b) Ground truth. (c) Baseline. (d) CSC-Trans. (e) CSC-Aug and (f) CSC-CutMix.
Figure 5. Qualitative results of cross-domain semantic segmentation on Potsdam RGB → Vaihingen IRRG. (a) Target images. (b) Ground truth. (c) Baseline. (d) CSC-Trans. (e) CSC-Aug and (f) CSC-CutMix.
Remotesensing 14 01527 g005
Figure 6. Qualitative results of the cross-domain semantic segmentation on Potsdam IRRG → Vaihingen IRRG. (a) Target images. (b) Ground truth. (c) Baseline. (d) CSC-Trans. (e) CSC-Aug and (f) CSC-CutMix.
Figure 6. Qualitative results of the cross-domain semantic segmentation on Potsdam IRRG → Vaihingen IRRG. (a) Target images. (b) Ground truth. (c) Baseline. (d) CSC-Trans. (e) CSC-Aug and (f) CSC-CutMix.
Remotesensing 14 01527 g006
Figure 7. Qualitative results of the cross-domain semantic segmentation on Potsdam RGB → Vaihingen IRRG. (a) Target images. (b) Ground truth. (c) Baseline. (d) GAN-RSDA. (e) AdaptSegNet. (f) MUCSS. (g) RDG-OSA and (h) CSC-Aug.
Figure 7. Qualitative results of the cross-domain semantic segmentation on Potsdam RGB → Vaihingen IRRG. (a) Target images. (b) Ground truth. (c) Baseline. (d) GAN-RSDA. (e) AdaptSegNet. (f) MUCSS. (g) RDG-OSA and (h) CSC-Aug.
Remotesensing 14 01527 g007
Figure 8. Qualitative results of the cross-domain semantic segmentation on Potsdam IRRG → Vaihingen IRRG. (a) Target images. (b) Ground truth. (c) Baseline. (d) GAN-RSDA. (e) AdaptSegNet. (f) MUCSS. (g) RDG-OSA and (h) CSC-Aug.
Figure 8. Qualitative results of the cross-domain semantic segmentation on Potsdam IRRG → Vaihingen IRRG. (a) Target images. (b) Ground truth. (c) Baseline. (d) GAN-RSDA. (e) AdaptSegNet. (f) MUCSS. (g) RDG-OSA and (h) CSC-Aug.
Remotesensing 14 01527 g008
Figure 9. The visualization of distribution matching in latent space, where we map features to 2D space with t-SNE [46]: (a) Potsdam IRRG → Vaihingen IRRG; (b) Potsdam RGB → Vaihingen IRRG.
Figure 9. The visualization of distribution matching in latent space, where we map features to 2D space with t-SNE [46]: (a) Potsdam IRRG → Vaihingen IRRG; (b) Potsdam RGB → Vaihingen IRRG.
Remotesensing 14 01527 g009
Figure 10. Ablation study of contribution of each component to accuracy of individual classes on Potsdam IRRG → Vaihingen IRRG.
Figure 10. Ablation study of contribution of each component to accuracy of individual classes on Potsdam IRRG → Vaihingen IRRG.
Remotesensing 14 01527 g010
Figure 11. Ablation study of hyperparameter sensitivity.
Figure 11. Ablation study of hyperparameter sensitivity.
Remotesensing 14 01527 g011
Table 1. Quantitative results (%) of cross-domain semantic segmentation on Potsdam RGB → Vaihingen IRRG benchmark.
Table 1. Quantitative results (%) of cross-domain semantic segmentation on Potsdam RGB → Vaihingen IRRG benchmark.
MethodMetricsClutter/
Background
Impervious
Surface
CarTreeLow
Vegetation
BuildingOverall
Baseline (DeepLabv3+ [44])IoU2.6740.2418.3553.1412.8852.6329.98
F1-score4.6556.9330.4069.1922.6868.7442.10
GAN-RSDA [14]IoU2.2948.2725.7342.1623.3464.3334.35
F1-score3.5064.7940.2059.0337.5578.1347.20
AdaptSegNet [45]IoU6.2655.9134.0947.5623.1865.9738.83
F1-score9.5571.4450.3464.1737.2279.3652.01
MUCSS [15]IoU5.8754.2127.9543.7326.9468.7637.91
F1-score8.7770.0442.8960.5342.0981.2650.93
CCDA [27]IoU12.3864.4743.4352.8338.3776.8748.06
F1-score21.5577.7660.0569.6255.9486.9561.98
RDG-OSA [32]IoU9.8462.5954.2256.3137.8679.3350.02
F1-score14.5576.8170.0071.9254.5588.4162.71
CSC-TransIoU5.7957.3252.9351.7830.6174.3145.46
F1-score9.1672.6969.0268.1146.4085.1358.42
CSC-AugIoU8.1268.9157.4165.4748.3381.7855.00
F1-score11.2381.4872.7679.0464.7889.9466.54
CSC-CutMixIoU10.2110.2153.8956.4337.2978.3249.94
F1-score14.8114.8169.7472.0054.0287.7462.64
Table 2. Quantitative results (%) of cross-domain semantic segmentation on Potsdam IRRG → Vaihingen IRRG benchmark.
Table 2. Quantitative results (%) of cross-domain semantic segmentation on Potsdam IRRG → Vaihingen IRRG benchmark.
MethodMetricsClutter/
Background
Impervious
Surface
CarTreeLow
Vegetation
BuildingOverall
Baseline (DeepLabv3+ [44])IoU2.9947.8820.8258.7419.5761.3735.23
F1-score5.1864.4033.9373.8832.4775.8347.61
GAN-RSDA [14]IoU7.2657.3220.0444.2735.4765.3538.28
F1-score10.3272.6032.5361.0451.9978.8451.22
AdaptSegNet [45]IoU6.3262.5029.3155.7440.3070.4144.10
F1-score9.6776.6644.8171.3657.0182.5057.00
MUCSS [15]IoU11.1665.9426.3050.4939.8569.0743.80
F1-score14.7079.1540.7766.7656.5581.5356.58
CCDA [27]IoU\58.6428.1753.2830.3960.6046.22
F1-score\75.1345.8169.5247.6276.8962.99
RDG-OSA [32]IoU10.7070.3154.0459.2249.0381.2054.08
F1-score15.4882.4369.8574.2265.5289.5766.18
CSC-TransIoU8.4265.6754.7561.7242.6975.8851.52
F1-score12.7779.1570.5676.2359.4686.2064.06
CSC-AugIoU13.8375.5656.5865.5552.9284.1758.10
F1-score19.5986.0172.0179.0968.9691.3869.50
CSC-CutMixIoU10.4673.3157.9163.5851.5881.2556.35
F1-score14.9184.5073.0577.5767.8689.5667.91
Table 3. Ablation study of effects of components of our framework.
Table 3. Ablation study of effects of components of our framework.
MethodPotsdam RGB → Vaihingen IRRGPotsdam IRRG → Vaihingen IRRG
mIoU Δ mIoU Δ
Cycle + Self-supervised55.000.0058.100.00
Cycle consistency only48.50 12 % 54.09 7 %
Self-supervised consistency only36.56 34 % 48.10 17 %
Source only29.98 45 % 35.23 39 %
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Gao, H.; Zhao, Y.; Guo, P.; Sun, Z.; Chen, X.; Tang, Y. Cycle and Self-Supervised Consistency Training for Adapting Semantic Segmentation of Aerial Images. Remote Sens. 2022, 14, 1527. https://doi.org/10.3390/rs14071527

AMA Style

Gao H, Zhao Y, Guo P, Sun Z, Chen X, Tang Y. Cycle and Self-Supervised Consistency Training for Adapting Semantic Segmentation of Aerial Images. Remote Sensing. 2022; 14(7):1527. https://doi.org/10.3390/rs14071527

Chicago/Turabian Style

Gao, Han, Yang Zhao, Peng Guo, Zihao Sun, Xiuwan Chen, and Yunwei Tang. 2022. "Cycle and Self-Supervised Consistency Training for Adapting Semantic Segmentation of Aerial Images" Remote Sensing 14, no. 7: 1527. https://doi.org/10.3390/rs14071527

APA Style

Gao, H., Zhao, Y., Guo, P., Sun, Z., Chen, X., & Tang, Y. (2022). Cycle and Self-Supervised Consistency Training for Adapting Semantic Segmentation of Aerial Images. Remote Sensing, 14(7), 1527. https://doi.org/10.3390/rs14071527

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop