Next Article in Journal
ResSKNet-SSDP: Effective and Light End-To-End Architecture for Speaker Recognition
Previous Article in Journal
Rain Discrimination with Machine Learning Classifiers for Opportunistic Rain Detection System Using Satellite Micro-Wave Links
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Cross-Domain Object Detection by Dual Adaptive Branch

1
The School of Computer Science and Engineering, Tianjin University of Technology, No. 391 Bin Shui Xi Dao Road, Tianjin 300384, China
2
Tianjin Key Laboratory for Control Theory and Applications in Complicated System, Tianjin University of Technology, No. 391 Bin Shui Xi Dao Road, Tianjin 300384, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(3), 1199; https://doi.org/10.3390/s23031199
Submission received: 26 December 2022 / Revised: 17 January 2023 / Accepted: 18 January 2023 / Published: 20 January 2023
(This article belongs to the Section Environmental Sensing)

Abstract

:
The object detection task usually assumes that the training and test samples obey the same distribution, and this assumption is not valid in reality, therefore the study of cross-domain object detection is proposed. Compared with image classification, the cross-domain object detection task presents the greater challenge, which requires both accurate classification and localization of samples in the target domain. The teacher–student framework (the student model is supervised by pseudo-labels from the teacher model) has produced a large accuracy improvement in cross-domain object detection. Feature-level adversarial training is used in the student model, which allows features in the source and target domains to share a similar distribution. However, the direction and gradient of the weights can be divided into domain-specific and domain-invariant features, and the purpose of domain adaptive is to focus on the domain-invariant features while eliminating interference from the domain-specific features. Inspired by this, we propose a teacher–student framework named dual adaptive branch (DAB), which uses domain adversarial learning to address the domain distribution. Specifically, we ensure that the student model aligns domain-invariant features and suppresses domain-specific features in this process. We further validate our method based on multiple domains. The experimental results demonstrate that our proposed method significantly improves the performance of cross-domain object detection and achieves the competitive experimental results on common benchmarks.

1. Introduction

With the development of deep neural networks, many computer vision tasks have achieved great success. The convolutional neural network-based object detection has achieved excellent performance on various benchmark datasets. However, these successes also depend on a large amount of labeled data [1]. The collection of this data is very costly and time-consuming. Moreover, the performance of object detection models trained based on annotated data might be substantially degraded in a real scenario, and the main reason is a change in the captured view, appearance, background, lighting, or image quality. For these changing visual conditions, some work has begun to investigate the adaptive method [2]. Domain adaptive methods [3] can transfer knowledge from labeled source domains to unlabeled target domains, and it is a more cost-effective and functional option than annotating enough object samples. Therefore, the domain adaptive methods have a widely application in many fields.
The common domain adaptive methods include global feature adaptation [4], instance feature adaptation [5], and local feature adaptation [6]. The global feature adaptation easily confuses the object features in different categories, since each image contains multiple objects. Instance feature adaptation easily confuses the foreground and background features, since the detector output on the target domain is unstable, and most of the predicted foreground is actually background. Local feature adaptation is typically used to address the distribution shifts on the lower semantic levels. Domain adaptive in object classification only requires alignment to features in the same category of objects. However, which part of the features to align has also become a critical study problem in the object detection task because there is the uncertainty about the location of object.
In cross-domain object detection task, it generally contains a domain-invariant feature and a domain-specific feature. The domain-invariant feature is expressed as the composition of the object, and the domain-specific feature is expressed as whether the object has clear boundaries. As shown in Figure 1, we have analyzed the meaningful error types in cross-domain object detection. These error types are related to overall performance to minimize any confounding variables. This allows one to improve the interpretability of design decisions and to describe more clearly the strengths and weaknesses of the model. The pie chart shows the relative contribution of each error, while the bar plots show their absolute contribution. We can observe that the prediction results usually contain a large number of errors and false positives. This method cannot focus on domain-invariant feature alignment. If features are aligned as a whole instead of distinguishing between domain-invariant and domain-specific features, it will be detrimental to solve the errors and false positives problem in cross-domain object detection.
A desirable domain adaptive method should achieve the alignment of the domain-invariant features. Otherwise, it might have a “negative migration” effect on the target domain. The distribution difference metric function can minimize the difference between source and target domains. We can assume that the domains have the same or similar feature distributions when the difference is sufficiently minimal. However, considering the feature alignment method in this way, we find that it does not distinguish the domain-invariant and domain-specific features of the downstream task. Inspired by the above, we propose to suppress domain-specific features in the model. There is more focus on learning domain-invariant features on higher-level semantic space.
In this work, we propose a novel cross-domain object detection network with DAB structure. The purpose is to train a well performing model on a target domain with unknown domain annotations. We use adversarial learning and mutual learning to improve the detection performance of the target domain. Our model consists of two modules: a target-domain teacher model and a cross-domain student model. In the student model, we propose a domain-invariant feature alignment branch and a domain-specific feature suppression branch. A distribution difference measure function that focuses on domain invariant feature alignment while suppressing domain specific features is proposed. We map the features to a high-dimensional space and use discriminators with gradients for adversarial learning which adjusts the distribution differences between the source and target domains in the student model.
The main contributions of this paper are as follows:
(1)
A novel cross-domain object detection framework with dual adaptive branch is designed, which we called DAB. This framework utilizes two branches, domain-invariant feature alignment branch and domain-specific feature suppression branch, to overcome errors and false positives in cross-domain object detection.
(2)
The feature alignment branch and feature suppression branch are designed, respectively. With the purpose of domain-invariant feature alignment, we propose to map the features into a high-dimensional space and restrict the gradient using a distribution difference measure function, which minimizes the difference of domain-invariant features between two domains. With the purpose of domain-specific feature suppression, we propose to impose constraints on domain-specific features, which eliminate the influence of domain-specific features on cross-domain object detection.
(3)
The extensive experiments are conducted on various cross-domain benchmarks, and the experimental results demonstrate that our method achieves a significant performance improvement.

2. Related Works

Object Detection. Early object detection methods were based on the sliding-window methods, which apply the hand-crafted features and classifiers on dense image grids to find objects [7,8]. However, the traditional hand-crafted feature extraction method for object detection has some limitations, such as poor robustness to changing the objects, high time complexity, and redundant detection window. With the arrival of the deep convolutional neural network, one can solve the problems of traditional hand-crafted feature extraction methods and improve the detection speed and accuracy of object detection. The object detection task is quickly dominated by convolutional neural network (CNN), which can be divided into the two-stage object detection [9,10,11] and one-stage object detection [12,13,14,15,16].
Cross domain Object Detection. The purpose of cross-domain object detection is to detect objects in different domains. Xu et al. [16] have proposed to alleviate the domain shift problem of deformed component-based models (DPMs) by introducing adaptive support vector machine (SVM). Raj et al. [17] have proposed the subspace alignment methods to align the features extracted by using R-CNN models. The works mentioned above are either not trained in the end-to-end manner or focused on a specific case. Cross-domain object detection methods can be generally divided into two categories, namely, adversarial feature alignment [4,5,6,18,19,20,21,22,23,24,25,26] and self-training [27,28,29,30,31]. Besides the standard cross-domain object detection, the tasks of passive and multi-source have been studied in [21,32,33], respectively. In addition, refs. [34,35] have explored the problem of domain generalizability for object detection. This direction of research was first carried out by Chen et al. [5], who proposed a domain-adaptive faster R-CNN, which reduces the difference between image-level and instance-level distributions by embedding adversarial feature adaptation in a two-stage detection pipeline. Saito et al. [6] have proposed to align shallow local perceptual fields with deeper image-level features, i.e., strong local alignment and weak global alignment. Additionally, He et al. [35] have proposed a hierarchical domain feature alignment module and a weighted GRL to reweight the training samples. Kim et al. [36] have randomly expanded the source and target domains into multiple domains and solved the adaptation problem from the perspective of domain diversification. It addresses the issue of adaptability from the perspective of domain diversity. Ref. [37] has proposed a novel approach to domain adaption for object detection to mine the discriminative regions and focus on aligning them across both domains. Ref. [38] have adopted multi-level domain feature alignment. Ref. [39] has utilized classification consistency of image-level and instance-level predictions with the assistance of a multi-label classification model. Ref. [20] has proposed a center-aware feature alignment method that enables the discriminator focus on features from object regions. Refs. [4,25] have emphasized the different strategies for dealing with foreground and background features. Another popular methodology [27,28,39,40,41] is dedicated to address the problem of inaccurate labels in the target domain.
In summary, these methods have not properly addressed the potential contradiction between shift-ability and discrimin-ability in cross-domain object detectors. Therefore, we propose to design a novel dual adaptive branch, which addresses the domain-invariant feature alignment and domain-specific feature suppression in the cross-domain object detection task.

3. Proposed Method

3.1. Framework Overview

As shown in Figure 2, our model consists of two parts: a target-domain teacher model and a cross-domain student model. The teacher-student model is trained by mutual learning and adversarial learning. The target domain images are feeding into the teacher model to generate pseudo-labels, which are used to train the student model. The student model is updated with the teacher model by exponential moving average method (EMA). In the student model, we design two branches of feature alignment and feature suppression, respectively. The features are mapped to a high-dimensional space, and the differences between domain-invariant feature in two domain are minimized using a distribution difference measure function. Additionally, the domain-specific feature are constrained to eliminate the influence of domain-specific feature on cross-domain object detection.
The source domain image x i s and the target domain image x i t are as inputs, and the features F S ( x i s ) for the source domain and F T ( x i t ) for the target domain are obtained by the feature encoder. We design a novel DAB structure. This design allow us to efficiently align domain-invariant feature and suppress domain-specific features from different domains. Basically, our network consists of a target-domain teacher model and a cross-domain student model.
The feature encoder is to ensure that we can construct a feature space of all images, which is used to obtain the feature of the images. The DAB is designed to focuses on domain-invariant feature alignment while suppressing domain-specific feature. Additionally, the detector is used to output the predicted result. In particular, the purpose of the feature alignment branch is to minimize the differences between domain-invariant features using a measure function. The purpose of the feature suppression branch is to eliminate the influence of domain-specific feature.

3.2. Optimization Problem in Cross-Domain Student

In cross-domain object detection, we have N s source domains of labeled samples, defined as D S = { x i s , y i s } i = 1 N s , and N t target domains of unlabeled samples, defined as D T = { x i t } i = 1 N t . The x i s and x i t are denoted as the input samples in the source and target domains, respectively. The y i s = ( c i s , b i s ) denotes the labels of the corresponding input samples in the source domain, c i s is the category label, and b i s is the bounding box label.
We set the feature of the source domain sample as F S ( x i s ) = i = 1 N S f i s , and the feature of the target domain sample as F T ( x i t ) = i = 1 N T f i t . The regression representations in the source and target domains are denoted as P S ( f i s ) = i = 1 N S p i s and P T ( f i t ) = i = 1 N T p i t , respectively. Our objective is to achieve the same or similar distribution of features in the source and target domain. Therefore, the problem is converted into minimizing the distance of the feature matrix.
Firstly, we map the features in the source and target domains based on the Gaussian distribution. The mean values with different distributions are calculated by finding a continuous function in the sample space, which is used to evaluate the difference in distribution between the source and target domains, as shown in Equations (1) and (2).
F ^ S ( x i s ) = 1 N S i = 1 N S 2 γ N S ! ( f i s ) N S e γ ( f i s ) 2
F ^ T ( x i t ) = 1 N T i = 1 N T 2 γ N T ! ( f i t ) N T e γ ( f i t ) 2
where F ^ S and F ^ T denote mean of the feature mapping in source and target domains, respectively. γ = 1 δ 2 denotes the variance of the function.
We use the Frobenius norm to represent the mean of the samples with different distributions, as shown in Equation (3).
F ^ S F ^ T F = 1 N S i = 1 N S 2 γ N S ! ( f i s ) N S e γ ( f i s ) 2 1 N T i = 1 N T 2 γ N T ! ( f i t ) N T e γ ( f i t ) 2 F
Therefore, the objective function for minimizing the difference in feature distribution can be denoted as Equation (4).
min F ^ S F ^ T F
As shown in Figure 3a, Equation (4) is used to minimize the domain-invariant feature distribution, which only considers the means of the feature distributions in the source and target domains, but ignores the effect of the variance δ on the feature distributions, as shown in Figure 3b. Therefore, we design a domain adaptive kernel function for aligning the domain-invariant feature distribution and suppressing the domain-specific feature distribution in the source and target domains in parallel.
Therefore, we design f C ( F ^ S , F ^ T ) regular terms to suppress the variances of the domain-specific feature distribution, and obtain the optimization problem, as shown in Equation (5).
min F ^ S F ^ T F + f C ( F ^ S , F ^ T )
We define the domain adaptive kernel function K ( f i s , f i t ) as Equation (6).
K ( f i s , f i t ) = e ( f i s + f i t ) 2 δ 2 = e γ ( f i s + f i t ) 2 = e γ [ ( f i s ) 2 + ( f i t ) 2 ] e γ ( f i s f i t )
Then, we convert Equation (6) to Equation (7) according to Taylor’s formula.
K ( f i s , f i t ) = e γ [ ( f i s ) 2 + ( f i t ) 2 ] [ i = 1 n 2 γ ( f i s f i t ) n n ! + o ( f i s f i t ) n ]
When n , we can ignore the residual term to obtain Equation (8).
K ( f i s , f i t ) = i = 1 n 2 γ ( f i s f i t ) n n ! e γ [ ( f i s ) 2 + ( f i t ) 2 ] = i = 1 n 2 γ ( f i s f i t ) n n ! e γ ( f i s ) 2 e γ ( f i t ) 2 = i = 1 n 2 γ n ! ( f i s ) n e γ ( f i s ) 2 2 γ n ! ( f i t ) n e γ ( f i t ) 2
We set,
ϕ ( P ^ S ) = 2 γ n ! ( f 1 s ) n e γ ( f 1 s ) 2 2 γ n ! ( f 2 s ) n e γ ( f 2 s ) 2 2 γ n ! ( f n s ) n e γ ( f n s ) 2 , ϕ ( P ^ T ) = 2 γ n ! ( f 1 t ) n e γ ( f 1 t ) 2 2 γ n ! ( f 2 t ) n e γ ( f 2 t ) 2 2 γ n ! ( f n t ) n e γ ( f n t ) 2
Therefore, we are able to express the domain adaptive kernel function as Equation (10).
K ( f i s , f i t ) = ϕ ( P ^ S ) T ϕ ( P ^ T )
Then, the optimization problem Equation (5) can be expressed as Equation (11).
min 1 N S ϕ ( P ^ S ) 1 N T ϕ ( P ^ T ) F + 1 N S ϕ ( P ^ S ) T ϕ ( P ^ T ) 1 N T ϕ ( P ^ T ) T ϕ ( P ^ S ) 2
Theorem 1. 
In the kernel mapping operator, we define F ^ S F ^ T F and f C ( F ^ S , F ^ T ) . If F ^ S F ^ T F 0 , then we have f C ( F ^ S , F ^ T ) 0 .
Proof of Theorem 1. 
Since F ^ S F ^ T F 0 , we obtain ϕ ( P ^ S ) and ϕ ( P ^ T ) , which are similar, and, then, we have E ( ϕ ( P ^ S ) ) E ( ϕ ( P ^ T ) ) . By the theorem of expectation, we know that E ( A x ) = A E ( x ) , and then E ( ϕ ( P ^ S ) T ϕ ( P ^ T ) ) E ( ϕ ( P ^ T ) T ϕ ( P ^ S ) , thus f C ( F ^ S , F ^ T ) 0 . □

3.3. Feature Alignment Branch in Cross-Domain Student

In the domain adaptive task, it is easy to confuse the object features of different categories, since each image contains multiple objects. Therefore, we propose a feature alignment branch, which aims to minimize the difference of domain-invariant feature between source and target domain, as shown in Figure 4.
The feature encoder obtains the features of the source and target domain, respectively. At first, the quality focal loss is used to calculate the classification loss of the source domain for the source domain samples, which is densely supervised for the whole image.
The objective function is obtained as shown in Equation (12).
L Q F L ( P S ( f i s ) , y i g t ) = y i g t P S ( f i s ) β ( 1 y i g t ) log ( 1 P S ( f i s ) ) + y i g t log P S ( f i s )
where, y i g t denotes the pseudo label from teacher model, the parameter β controls the down-weighting rate smoothly, β = 2 .
Secondly, we train the feature encoder for obtaining the features F S ( x i s ) for the input x i s in the source domain and F T ( x i t ) for the input x i t in the target domain, respectively, and we map the feature to F ^ S and F ^ T , respectively. We use a gradient reversal layer (GRL) in the feature alignment branch, which allows the sign of the gradient to be reversed when the gradient passes through the GRL layer.
We design the feature alignment branch to align the domain-invariant feature in source and target domain. We can derive two distributions, the source anchor distribution D S a n c and the target anchor distribution D T a n c . Therefore, the objective function is obtained as shown in Equation (13).
L c l s a d v = E ( x i s , y i g t ) D S a n c F ^ S E ( x i t ) D T a n c F ^ T F + f C ( E ( x i s , y i g t ) D S a n c F ^ S E ( x i t ) D T a n c F ^ T )
where we denote each source-domain anchor x i s D S a n c with a ground truth label y i g t .
The Equation (12) is used to guide the feature alignment branch for correct classification in the source domain. Equation (13) is used to encourage the alignment of the mapping F ^ T on the target domain with the mapping F ^ S on the source domain, which reduces the difference in the domain-invariant feature distribution. In summary, the objective function of the feature alignment branch is obtained as shown in Equation (14).
E ( x i s , y i g t ) D S a n c L Q F L ( P S ( f i s ) , y i g t ) + λ L c l s a d v
where λ is the trade-off parameter.

3.4. Feature Suppression Branch in Cross-Domain Student

In the domain adaptive task, we aim to align the domain-invariant feature of two domains while suppressing the domain-specific feature, which gradually converges the feature distributions to the corresponding local optimal values. Therefore, we propose a feature suppression branch for constraining domain-specific feature, as shown in Figure 5. This branch can significantly improve the performance of cross-domain object detection. Especially, the model is well pre-trained on the source domain. The reason is the constraint inhibits the learning of the final task, but the pre-training process performed on the source domain will compensate this limitation. It removes a key obstacle to the cross-domain object detection task.
We have the ground truth category and bounding box labels in each object of the source domain. At first, we calculate the regression loss of the source domain. The objective function is obtained as shown in Equation (15).
L r e g ( P S ( f i s ) , y i g t ) = 1 P S ( f i s ) y i g t P S ( f i s ) y i g t + ( ρ 2 ( P S ( f i s ) , y i g t ) ) c 2 + α v
where v = 4 π 2 ( a r c t a n w i g t h i g t a r c t a n w i p r e h i p r e ) 2 and α = v ( 1 I o U ) + v . w i g t and h i g t denote the width and height of the ground truth bounding box, respectively. w i p r e and h i p r e denoted the width and height of the prediction bounding box obtained from P T ( f i t ) , respectively.
The ground truth labels proposed for the target domain are unknown, and we use P T ( f i t ) to calculate them. However, it is difficult to obtain a satisfactory detector at the target domain since there is the domain offset. Therefore, we propose a feature suppression branch for domain-specific feature.
In the target domain, we train the feature encoder for obtaining the feature F T ( x i t ) of the input x i t , and the P T ( f i t ) is obtained through the mapping representation F ^ T of feature F T ( x i t ) .
With the purpose of feature suppression branch, we measure the differences across domains and suppress the difference of domain-specific feature in source and target domain. The objective function is obtained, as shown in Equation (16).
L r e g a d v = P S ( f i s ) P T ( f i t ) F + f C ( P S ( f i s ) , P T ( f i t ) )
It is noted that the regression loss in the source domain is defined only for the bounding box corresponding to y i g t , while L r e g a d v in the target domain is defined only for the bounding box associated with the predicted category in P T ( f i t ) . Equation (15) guides the correct prediction in the source domain. Additionally, Equation (16) reduces the difference in the regression representation, which prevents the alignment of the regression representation P T ( f i t ) and P S ( f i s ) in the target domain. In summary, the objective function of the feature suppression branch is obtained, as shown in Equation (17).
L r e g ( P S ( f i s ) , y i g t ) + μ L r e g a d v
where μ is the trade-off parameter.

4. Experiment and Analysis

4.1. Datasets and Scenarios

4.1.1. Datasets

We perform experiments with popular benchmarks in cross-domain object detection. The details of the datasets are shown in Table 1.
Pascal VOC. It is a dataset collected from the real world and can be used for detection and segmentation. For the detection task, it mainly consists of 20 categories containing 2501 training images, 2510 validation images, and 4952 test images.
Clipart. It is an artistic image created by manual production. Clipart contains 1000 images in a total of 20 categories.
Watercolor. It contains watercolor style images, which consist of images from six categories.
DT Clipart. It uses CycleGAN for style migration, which converts the Pascal VOC dataset to the style of the Clipart dataset. Therefore, the annotation information is identical to the original Pascal VOC dataset.
Cityscapes. It is a semantic segmentation dataset consisting of 2975 training images, 500 validation images, and 1525 test images, each with a size of 1024 × 2048. Each image is annotated at pixel level and can be used for target detection tasks after conversion. The datasets are all urban scenes of different cities under normal weather, and the target objects are mainly pedestrians, vehicles, etc.
Foggy Cityscapes. It is created by adding synthetic fog into the Cityscapes dataset. Therefore, the annotation information is exactly the same as the original Cityscapes dataset.

4.1.2. Scenario

We evaluate our method in two adaptation scenarios.
Dissimilar domains. The purpose is to perform adaptation under dissimilar domains. Firstly, we use the Pascal VOC and Clipart as the source and target domains, respectively. The results are presented on the Clipart val set (Pascal VOC → Clipart). Secondly, we use the Pascal VOC and DT Clipart as the source and target domains, respectively. The results are presented on the DT Clipart test set (Pascal VOC → DT Clipart).
Adverse weather. The purpose is to perform adaptation under different weather conditions. We use the Cityscapes and Foggy Cityscapes as the source and target domains, respectively. The results are presented on the Foggy Cityscapes val set (Cityscapes → Foggy Cityscapes).

4.2. Implementation Details

The source code and models were trained and evaluated on the Pytorch toolbox, which is based on the Python 3.6 platform. All experiments were implemented on a NVIDIA RTX 3090Ti GPU. We train the network with a batch size of 16. We use an initial learning rate of 0.125 and a decay rate of 0.1 every 400 K steps. The different scales of detection correspond to different perception fields, and there are a total of 10,647 proposal boxes. We transform the training set in the model by data augmentation, which enriches the training set and enhances the generalization ability. The four images are randomly cropped and scaled, and then randomly arranged and stitched to form a single image. While enriching the dataset, the data of four images are calculated at once during the normalization operation. Therefore, the memory requirement of the model is reduced.

4.3. Experimental Results

4.3.1. Adaptation between Dissimilar Domains

Firstly, we have reported the adaptive experiments on dissimilar domains. We use the Pascal VOC dataset as the source domain and the Clipart dataset as the target domain. Source only indicates a model trained with only source domain data, and oracle indicates a model trained with labeled data from the source and target domains. The FA branch only denotes the method with feature alignment branch only, FS branch only denotes the method with feature suppression branch only, and Proposed Method denotes the method with DAB.
As shown in Table 2, we can observe that our method achieves 44.1% mAP with only feature alignment branch (FA branch only), which equals the advanced algorithm UMT [45]. This demonstrates that the feature alignment branch is effective in aligning domain-invariant feature. Additionally, our method achieves 45.9% mAP with only feature suppression branch (FS branch only), which exceeds the advanced algorithm UMT [45] by +1.8%. This illustrates the feature suppression branch can improve detection performance effectively by constraining domain-specific feature. Our method achieves 48.4% mAP with dual adaptive branch (Proposed Method), which outperforms all the other methods, and the detection performance has achieved an improvement of an order of magnitude.
Although in the experiments mentioned above, our method does not perform best in some categories, such as (‘chair’, ‘cow’, and ‘mbike’). However, we can observe that their differences of average precision is actually not large, and the detection performance of Proposed Method outperforms the method using only feature alignment branch (FA branch only) and only feature suppression branch (FS branch only) in these categories. Additionally, with the confusion matrix in Figure 6, it can be observed that ‘aero’ may be recognized as the ‘bird’ category, and ‘cat’ may be recognized as the ‘dog’ category. We analyze the reasons for the above situation, and it is influenced by the variation of image styles in cross-domain datasets, and it is easy to learn the approximate features in some small sample categories so that the accuracy of detection is reduced.
Figure 7 shows the heat maps for three exampled images from Clipart, where the main objects related to categories such as “sheep”, “person”, “chair”, “cow”, and “bottle” is localized. Figure 7a–c are heat maps of attention for the FA branch only, FS branch only, and Proposed Method, respectively. It can be observed that the proposed method enables a more accurate alignment of the critical regions and instances. Therefore, it can help the model to activate the main objects of interest more accurately and achieve improved detection performance.
In addition, we use the Pascal VOC dataset as the source domain and the DT Clipart dataset as the target domain. As shown in Table 3, we can observe that our method achieves 52.1% mAP with only the feature alignment branch (FA branch only), our method achieves 53.1% mAP with only the feature suppression branch (FS branch only), and our method achieves 54.7% mAP with dual adaptive branch (Proposed Method), which outperforms all the other methods and the detection performance has achieved the order of magnitude improvement.
We will further evaluate the generalization ability of our model on unknown domains. We trained the model on the source domain dataset with labels (PASCAL VOC) and another artistic dataset without labels (Watercolor). Then, we inferred the model on a target dataset (Clipart), which is unknown during training. We only trained the overlapping classes (six classes) between Clipart1k and Watercolor, and the results are given in Table 4.
Compared with AT [51] and MT [50], our model achieved the best performance in several categories. Despite this, in some categories (such as bicycle, dog, and person), our method is not achieving the optimal performance, but we can observe that their differences of average precision is actually not large. It demonstrates that our model can promote to unknown domains.

4.3.2. Adaptation between Adverse Weather

We have reported the experimental results of domain adaptive object detection under adverse weather condition. Source only indicates a model trained with only source domain data, and oracle indicates a model trained with labeled data from the source and target domains.The FA branch only denotes the method with feature alignment branch only, FS branch only denotes the method with feature suppression branch only, and Proposed Method denotes the method with DAB.
Table 5 shows the experimental results on Cityscapes → Foggy Cityscapes transfer. We can observe that our method achieves 43.6% mAP with feature alignment branch only (FA branch only), our method achieves 43.5% mAP with feature suppression branch only (FS branch only), and our method achieves 47.4% mAP with DAB (Proposed Method). It is worth noting that the mAP is approximated in the FA branch only method and the FS branch only method. However, in the categories of pedestrians and riders, the detection performance of the FS branch only method is better. The state-of-the-art method TDD [52] achieves 43.1% mAP, while our method achieves +4.3% gains. It shows that our method has the stable ability to solve the domain adaptive problem in adverse weather.
The confusion matrix for the source only method and our method are shown in Figure 8. We can clearly observe the improvement in detection quality. The results show that the proposed method significantly improves the performance, especially for the accuracy of localization.

4.4. Ablation Study

In this study, we investigate the performance of various strategies for aligning feature representations. We have used Pascal VOC → Clipart to conduct the study. For a fair comparison, all experiments have been performed under the same settings.
Trade-off parameter. Firstly, we investigate the effect of different trade-off parameters for the performance in domain adaptive object detection. We present the results of the ablation experiments for the trade-off parameter in Table 6.
We can observe that the trade-off parameter of EXP.2 achieves optimal performance. Both A P 0.5 and A P 0.5 : 0.95 outperformed the results of EXP.1 and EXP.3, with gains of +4.7% and +0.9%, and +2.7%, and +0.2%, respectively.
Scales. There is a potential scale shift between the source and target domain datasets. To investigate the effect of image scale on our method, we have changed the size of the image in the target domain, and the scale in the source domain is fixed at 640 pixels. We have plotted the detection performance at different image scales by changing the scales of the target domain images. The FA branch only denotes the method with feature alignment branch only, FS branch only denotes the method with feature suppression branch only, w/o DAB denotes the method without the DAB, and Proposed Method denotes the method with DAB.
As shown in Figure 9a,b, we observe that changing the scales under the same experimental conditions, the model with the DAB achieves better results at most scales. In Figure 9c, we can observe the detection performance of our model with the DAB at each scale. EXP.1-EXP.5 indicate the experiments were performed at various model depth conditions. In Figure 9d, we can observe the inference speed at different scales.
Comparing the two branch, we observe that the feature alignment branch is more robust to scale variation than the feature suppression branch. The reason is that scale variation is a global shift that affects all objects and backgrounds. While, in our method, the global domain shift is mainly solved by feature alignment branch to domain-invariant feature alignment, and the feature suppression branch is used to constraint domain-specific feature. When a significant global domain shift is present, the localization error is increased, and therefore the accuracy of the feature suppression branch is affected by the error in the domain-specific feature. Nevertheless, DAB consistently provide the optimal results at all scales.
Conv kernel. The depth of the model might affect the performance of feature extraction. To investigate the effect of different model depths for our method, we conduct experiments with different convolution kernel conditions. The reported results are set with Pascal VOC → Clipart. We conducted experiments at the same image scale as shown in Table 7.
We can observe that the EXP.5 obtained the optimal performance, the A P 0.5 and A P 0.5 : 0.95 achieved 46.2% and 25.9% mAP, respectively. We also report the results of different convolution kernels in Figure 10, which evaluates the accuracy and speed of different convolution kernel models. It can be observed that, for trade-off between accuracy and speed, EXP.5 also achieves optimal performance.
Branch structure. To verify the effectiveness of our dual adaptive branch structure, we have performed a set of ablation studies. The reported results are set with Pascal VOC → Clipart. The results of the different experiments are shown in Table 8.
We use Proposed Method as the baseline. It can be observed that Proposed Method outperforms FA branch only in A P 0.5 and A P 0.5 : 0.95 , achieving the gains of +2.5% and +0.9%, respectively. The Proposed Method outperforms the FS branch only in A P 0.5 and A P 0.5 : 0.95 , achieving gains of +4.3% and +1.1%, respectively. The Proposed Method outperforms all the above experiments in A P 0.5 and A P 0.5 : 0.95 , achieving 48.4% and 26.5% mAP. It demonstrates that the model performance is gradually improved with DAB involved in the training, which illustrates the utility of each branch. The experiment of DAB is superior to all single branch methods, which indicates that our method preserves useful source domain knowledge effectively and explores target domain information in parallel.

4.5. Error Analysis

To create a meaningful distribution of errors and identify the components of the mAP, we separated all false positives and false negatives of model with four types. We will represent the overlap between the maximum IoU of a false positive and the ground truth of a given category as IoUmax. The foreground IoU threshold is denoted as t f , the background threshold is denoted as t b , and the above thresholds are set at 0.5 and 0.1, respectively [57]. I o U m a x t f is denoted as classification error (Cls), which indicates localized correctly but classified incorrectly. t b I o U m a x t f is denoted as localization error (Loc), which indicates classified correctly but localized incorrectly. I o U m a x t b is denoted as background error (Bkg), which indicates detected background as foreground. Additionally, missed GT error (Miss) indicates undetected ground truth.
The error ratios for each model on Pascal VOC → Clipart are shown in Figure 11. It can be observed that the main errors in the target domain appear from: Miss (undetected ground truth), Cls (incorrect classification), and Loc (incorrect localization). As shown in column 2 of Figure 11, it is observed that the error ratio of Cls is effectively reduced after feature alignment branch, but the error ratio of Loc is increased. This also illustrates the necessity for the feature suppression branch is performed. As shown in column 3 of Figure 11, it can be observed that the error ratio of Loc effectively decreases through the constraint domain-specific feature. In summary, it is illustrated that DAB is reasonable.

4.6. Visualisation Results

Figure 12 shows the qualitative results in Pascal VOC → Clipart cross-domain detection. From top to bottom, the visualization results of ground truth, source only, DA-faster [5], and proposed method are shown, respectively. We can observe that there are many missing and incorrect results in the cource only. Compared with DA-faster [5], our method has more significant improvement in localization accuracy, which indicates that the problem of errors and false positives has been improved. We can clearly observe the improvement of the detection quality. The results demonstrate a significant improvement of performance with our method.
Figure 13 shows the qualitative results in Pascal VOC → DT Clipart cross-domain detection. Figure 14 shows the qualitative results in Cityscapes → Foggy Cityscapes cross-domain detection. From left to right, the visualization results of ground truth, source only, DA-faster and proposed method are shown, respectively. We can observe that the detection results of source only have incorrect detection and mis-location. In the DA-Faster [5] detection results, there are several omissions. Additionally, our method significantly improves the appearance of the above situations. Notably, our method shows a competitive performance with the oracle model. This demonstrates that our model can perceive the knowledge of the target domain while retaining the useful information of the source domain.

5. Conclusions

In this work, we address the domain shift in high-level semantic features by proposing a novel DAB structure. The purpose of domain adaptation is to focus on domain-invariant features and to eliminate the interference of domain-specific features. Therefore, we propose feature alignment and feature suppression branches, respectively. The effect of feature shift is eliminated by this strategy, which reduces the probability of false positives and errors in detection. Specifically, we exploit a distribution difference metric function to improve prediction consistency. It allows the model to focus on the object-relevant features aligned in the high-level semantic space. Experimental results on the common benchmarks indicated that our model achieves a comparable performance with advanced methods, achieving 48.4% mAP and 47.4% mAP, outperforming the next best methods by +1.8% and +4.3%, respectively. The experimental results also indicated that our detector is highly robust in different scales, which is very effective and advantageous in cross-domain object detection.

Author Contributions

Methodology, X.L. and N.L.; investigation, N.L.; writing—original draft preparation, X.L.; writing—review and editing, B.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Research and Innovation Project for Postgraduates in Tianjin (Artificial Intelligence) grant number 2020YJSZXB08, the Youth Program of Tianjin Natural Science Foundation grant number 21JCQNJC00910, the State Key Program of Tianjin Natural Science Foundation grant number 21JCZDJC00760, and the Key Training Project for Tianjin” Project plus Team” grant number XC202054.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors acknowledge the research support from the School of Computer Science and Engineering, the School of Electrical Engineering and Automation, and Tianjin Key Laboratory for Control Theory and Applications in Complicated System at Tianjin University of Technology.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. ImageNet Large Scale Visual Recognition Challenge. Int. J. Comput. Vis. (IJCV) 2015, 115, 211–252. [Google Scholar] [CrossRef] [Green Version]
  2. Doulamis, N.; Voulodimos, A. FAST-MDL: Fast Adaptive Supervised Training of multi-layered deep learning models for consistent object tracking and classification. In Proceedings of the 2016 IEEE International Conference on Imaging Systems and Techniques (IST), Chania, Greece, 4–6 October 2016; pp. 318–323. [Google Scholar] [CrossRef]
  3. Kate, S.; Brian, K.; Mario, F.; Trevor, D. Adapting Visual Category Models to New Domains. In Proceedings of the Computer Vision—ECCV 2010, Heraklion, Greece, 5–11 September 2010; Springer: Berlin/Heidelberg, Germany, 2010; pp. 213–226. [Google Scholar]
  4. Zheng, Y.; Huang, D.; Liu, S.; Wang, Y. Cross-domain Object Detection through Coarse-to-Fine Feature Adaptation. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 13763–13772. [Google Scholar] [CrossRef]
  5. Chen, Y.; Li, W.; Sakaridis, C.; Dai, D.; Van Gool, L. Domain Adaptive Faster R-CNN for Object Detection in the Wild. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3339–3348. [Google Scholar] [CrossRef] [Green Version]
  6. Saito, K.; Ushiku, Y.; Harada, T.; Saenko, K. Strong-Weak Distribution Alignment for Adaptive Object Detection. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 6949–6958. [Google Scholar] [CrossRef] [Green Version]
  7. David, F. Object Detection with Discriminatively Trained Part-Based Models. Computer 2014, 47, 6–7. [Google Scholar] [CrossRef]
  8. Viola, P.; Jones, M. Rapid object detection using a boosted cascade of simple features. In Proceedings of the 2001 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR 2001), Kauai, HI, USA, 8–14 December 2001; Volume 1, p. 511. [Google Scholar] [CrossRef]
  9. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar] [CrossRef] [Green Version]
  10. Girshick, R. Fast R-CNN. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar] [CrossRef]
  11. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed]
  12. Joseph, R.; Ali, F. YOLO9000: Better, Faster, Stronger. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 6517–6525. [Google Scholar] [CrossRef] [Green Version]
  13. Redmon, J.; Farhadi, A. YOLOv3: An Incremental Improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  14. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. SSD: Single Shot MultiBox Detector. In Proceedings of the Computer Vision—ECCV 2016, Amsterdam, The Netherlands, 8–10 October 2016; Springer International Publishing: Cham, Switzerland, 2016; pp. 21–37. [Google Scholar]
  15. Zhang, S.; Wen, L.; Lei, Z.; Li, S.Z. RefineDet++: Single-Shot Refinement Neural Network for Object Detection. IEEE Trans. Circuits Syst. Video Technol. 2021, 31, 674–687. [Google Scholar] [CrossRef]
  16. Xu, J.; Ramos, S.; Vázquez, D.; López, A.M. Domain Adaptation of Deformable Part-Based Models. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 36, 2367–2380. [Google Scholar] [CrossRef] [PubMed]
  17. Raj, A.; Namboodiri, V.P.; Tuytelaars, T. Subspace Alignment Based Domain Adaptation for RCNN Detector. arXiv 2015, arXiv:1507.05578. [Google Scholar]
  18. Chen, C.; Zheng, Z.; Huang, Y.; Ding, X.; Yu, Y. I3Net: Implicit Instance-Invariant Network for Adapting One-Stage Object Detectors. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 12571–12580. [Google Scholar] [CrossRef]
  19. Chen, Y.; Wang, H.; Li, W.; Sakaridis, C.; Dai, D.; Van Gool, L. Scale-Aware Domain Adaptive Faster R-CNN. Int. J. Comput. Vis. 2021, 129, 2223–2243. [Google Scholar] [CrossRef]
  20. Hsu, C.C.; Tsai, Y.H.; Lin, Y.Y.; Yang, M.H. Every Pixel Matters: Center-Aware Feature Alignment for Domain Adaptive Object Detector. In Proceedings of the Computer Vision—ECCV 2020, Glasgow, UK, 23–28 August 2020; Springer International Publishing: Cham, Switzerland, 2020; pp. 733–748. [Google Scholar]
  21. Rezaeianaran, F.; Shetty, R.; Aljundi, R.; Reino, D.O.; Zhang, S.; Schiele, B. Seeking Similarities over Differences: Similarity-based Domain Alignment for Adaptive Object Detection. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 9184–9193. [Google Scholar] [CrossRef]
  22. Tian, K.; Zhang, C.; Wang, Y.; Xiang, S.; Pan, C. Knowledge Mining and Transferring for Domain Adaptive Object Detection. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 9113–9122. [Google Scholar] [CrossRef]
  23. Wu, A.; Liu, R.; Han, Y.; Zhu, L.; Yang, Y. Vector-Decomposed Disentanglement for Domain-Invariant Object Detection. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 9322–9331. [Google Scholar] [CrossRef]
  24. Xu, M.; Wang, H.; Ni, B.; Tian, Q.; Zhang, W. Cross-Domain Detection via Graph-Induced Prototype Alignment. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020. [Google Scholar] [CrossRef]
  25. Zhang, Y.; Wang, Z.; Mao, Y. RPN Prototype Alignment For Domain Adaptive Object Detector. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 12420–12429. [Google Scholar] [CrossRef]
  26. Zhao, Z.; Guo, Y.; Shen, H.; Ye, J. Adaptive Object Detection with Dual Multi-label Prediction. In Proceedings of the Computer Vision—ECCV 2020, Glasgow, UK, 23–28 August 2020; Springer International Publishing: Cham, Switzerland, 2020; pp. 54–69. [Google Scholar]
  27. Mehran, K.; Arash, V.; Mani, R.; William, M. A Robust Learning Approach to Domain Adaptive Object Detection. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 480–490. [Google Scholar] [CrossRef] [Green Version]
  28. Seunghyeon, K.; Jaehoon, C.; Taekyung, K.; Changick, K. Self-Training and Adversarial Background Regularization for Unsupervised Domain Adaptive One-Stage Object Detection. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 6091–6100. [Google Scholar] [CrossRef] [Green Version]
  29. Li, S.; Huang, J.; Hua, X.S.; Zhang, L. Category Dictionary Guided Unsupervised Domain Adaptation for Object Detection. In Proceedings of the National Conference on Artificial Intelligence, Sao Paulo, Brazil, 21–25 June 2021. [Google Scholar]
  30. Ramamonjison, R.; Banitalebi-Dehkordi, A.; Kang, X.; Bai, X.; Zhang, Y. SimROD: A Simple Adaptation Method for Robust Object Detection. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, BC, Canada, 10–17 October 2021; pp. 3550–3559. [Google Scholar] [CrossRef]
  31. RoyChowdhury, A.; Chakrabarty, P.; Singh, A.; Jin, S.; Jiang, H.; Cao, L.; Learned-Miller, E. Automatic Adaptation of Object Detectors to New Domains Using Self-Training. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 780–790. [Google Scholar] [CrossRef] [Green Version]
  32. Li, X.; Chen, W.; Xie, D.; Yang, S.; Yuan, P.; Pu, S.; Zhuang, Y. A Free Lunch for Unsupervised Domain Adaptive Object Detection without Source Data. In Proceedings of the AAAI, Virtually, 2–9 February 2021; pp. 1–9. [Google Scholar]
  33. Lin, C.; Yuan, Z.; Zhao, S.; Sun, P.; Wang, C.; Cai, J. Domain-Invariant Disentangled Network for Generalizable Object Detection. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, BC, Canada, 10–17 October 2021; pp. 8751–8760. [Google Scholar] [CrossRef]
  34. Wang, X.; Huang, T.E.; Liu, B.; Yu, F.; Wang, X.; Gonzalez, J.E.; Darrell, T. Robust Object Detection via Instance-Level Temporal Cycle Confusion. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, BC, Canada, 10–17 October 2021; pp. 9123–9132. [Google Scholar] [CrossRef]
  35. He, Z.; Zhang, L. Multi-Adversarial Faster-RCNN for Unrestricted Object Detection. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 6667–6676. [Google Scholar] [CrossRef] [Green Version]
  36. Kim, T.; Jeong, M.; Kim, S.; Choi, S.; Kim, C. Diversify and Match: A Domain Adaptive Representation Learning Paradigm for Object Detection. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 12448–12457. [Google Scholar] [CrossRef] [Green Version]
  37. Zhu, X.; Pang, J.; Yang, C.; Shi, J.; Lin, D. Adapting Object Detectors via Selective Cross-Domain Alignment. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 687–696. [Google Scholar] [CrossRef]
  38. Xie, R.; Yu, F.; Wang, J.; Wang, Y.; Zhang, L. Multi-Level Domain Adaptive Learning for Cross-Domain Detection. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW), Seoul, Republic of Korea, 27–28 October 2019; pp. 3213–3219. [Google Scholar] [CrossRef]
  39. Cai, Q.; Pan, Y.; Ngo, C.W.; Tian, X.; Duan, L.; Yao, T. Exploring Object Relation in Mean Teacher for Cross-Domain Detection. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 11449–11458. [Google Scholar] [CrossRef] [Green Version]
  40. He, Z.; Zhang, L. Domain Adaptive Object Detection via Asymmetric Tri-Way Faster-RCNN. In Proceedings of the Computer Vision—ECCV 2020, Glasgow, UK, 23–28 August 2020; Springer International Publishing: Cham, Switzerland, 2020; pp. 309–324. [Google Scholar]
  41. Zhao, G.; Li, G.; Xu, R.; Lin, L. Collaborative Training Between Region Proposal Localization and Classification for Domain Adaptive Object Detection. In Proceedings of the Computer Vision—ECCV 2020, Glasgow, UK, 23–28 August 2020; Springer International Publishing: Cham, Switzerland, 2020; pp. 86–102. [Google Scholar]
  42. Everingham, M.; Van Gool, L.; Williams, C.K.I.; Winn, J.; Zisserman, A. The Pascal visual object classes challenge: A retrospective. Int. J. Comput. Vis. 2015, 111, 98–136. [Google Scholar] [CrossRef]
  43. Wilber, M.J.; Fang, C.; Jin, H.; Hertzmann, A.; Collomosse, J.; Belongie, S. BAM! The Behance Artistic Media Dataset for Recognition Beyond Photography. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 1211–1220. [Google Scholar] [CrossRef] [Green Version]
  44. Cordts, M.; Omran, M.; Ramos, S.; Rehfeld, T.; Enzweiler, M.; Benenson, R.; Franke, U.; Roth, S.; Schiele, B. The Cityscapes Dataset for Semantic Urban Scene Understanding. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 3213–3223. [Google Scholar] [CrossRef] [Green Version]
  45. Deng, J.; Li, W.; Chen, Y.; Duan, L. Unbiased Mean Teacher for Cross-domain Object Detection. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 4091–4101. [Google Scholar] [CrossRef]
  46. Shen, Z.; Maheshwari, H.; Yao, W.; Savvides, M. SCL: Towards Accurate Domain Adaptive Object Detection via Gradient Detach Based Stacked Complementary Losses. arXiv 2019, arXiv:1911.02559. [Google Scholar]
  47. Chen, C.; Zheng, Z.; Ding, X.; Huang, Y.; Dou, Q. Harmonizing Transferability and Discriminability for Adapting Object Detectors. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 8866–8875. [Google Scholar] [CrossRef]
  48. Tzeng, E.; Hoffman, J.; Saenko, K.; Darrell, T. Adversarial Discriminative Domain Adaptation. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 2962–2971. [Google Scholar] [CrossRef] [Green Version]
  49. Inoue, N.; Furuta, R.; Yamasaki, T.; Aizawa, K. Cross-domain weakly-supervised object detection through progressive domain adaptation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 5001–5009. [Google Scholar]
  50. Tarvainen, A.; Valpola, H. Mean teachers are better role models: Weight-averaged consistency targets improve semi-supervised deep learning results. arXiv 2017, arXiv:1703.01780. [Google Scholar]
  51. Li, Y.J.; Dai, X.; Ma, C.Y.; Liu, Y.C.; Chen, K.; Wu, B.; He, Z.; Kitani, K.; Vajda, P. Cross-Domain Adaptive Teacher for Object Detection. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022, New Orleans, LA, USA, 19–24 June 2016; pp. 7571–7580. [Google Scholar] [CrossRef]
  52. He, M.; Wang, Y.; Wu, J.; Wang, Y.; Li, H.; Li, B.; Gan, W.; Wu, W.; Qiao, Y. Cross Domain Object Detection by Target-Perceived Dual Branch Distillation. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 9560–9570. [Google Scholar] [CrossRef]
  53. Zhuang, C.; Han, X.; Huang, W.; Scott, M. iFAN: Image-Instance Full Alignment Networks for Adaptive Object Detection. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 13122–13129. [Google Scholar] [CrossRef]
  54. Wang, W.; Cao, Y.; Zhang, J.; He, F.; Zha, Z.J.; Wen, Y.; Tao, D. Exploring Sequence Feature Alignment for Domain Adaptive Detection Transformers. arXiv 2021, arXiv:2107.12636. [Google Scholar]
  55. Wang, H.; Liao, S.; Shao, L. AFAN: Augmented Feature Alignment Network for Cross-Domain Object Detection. IEEE Trans. Image Process. 2021, 30, 4046–4056. [Google Scholar] [CrossRef] [PubMed]
  56. Wu, Z.; Wang, X.; Xu, T.; Yang, X.; Zou, L.; Xu, L.; Weise, T. Domain-Invariant Proposals based on a Balanced Domain Classifier for Object Detection. arXiv 2022, arXiv:2202.05941. [Google Scholar]
  57. Bolya, D.; Foley, S.; Hays, J.; Hoffman, J. TIDE: A General Toolbox for Identifying Object Detection Errors. In Proceedings of the Computer Vision—ECCV 2020, Glasgow, UK, 23–28 August 2020; Springer International Publishing: Cham, Switzerland, 2020; pp. 558–573. [Google Scholar]
Figure 1. The main error sources in cross-domain object detection.
Figure 1. The main error sources in cross-domain object detection.
Sensors 23 01199 g001
Figure 2. Framework Overview. This framework consists of two parts: a target-domain teacher model and a cross-domain student model. In target-domain teacher model, the target domain images are feeding into the teacher model to generate pseudo-labels. In cross-domain student model, we propose a DAB structure. With the purpose of feature alignment branch, the features are mappes into a high-dimensional space, and they restrict the gradient using a distribution difference measure function, which minimizes the difference of domain-invariant features between two domains. With the purpose of feature suppression branch, the domain-specific features are constrainted, which eliminate the influence of domain-specific feature on cross-domain object detection.
Figure 2. Framework Overview. This framework consists of two parts: a target-domain teacher model and a cross-domain student model. In target-domain teacher model, the target domain images are feeding into the teacher model to generate pseudo-labels. In cross-domain student model, we propose a DAB structure. With the purpose of feature alignment branch, the features are mappes into a high-dimensional space, and they restrict the gradient using a distribution difference measure function, which minimizes the difference of domain-invariant features between two domains. With the purpose of feature suppression branch, the domain-specific features are constrainted, which eliminate the influence of domain-specific feature on cross-domain object detection.
Sensors 23 01199 g002
Figure 3. The feature distribution.
Figure 3. The feature distribution.
Sensors 23 01199 g003
Figure 4. The feature alignment branch.
Figure 4. The feature alignment branch.
Sensors 23 01199 g004
Figure 5. The feature suppression branch.
Figure 5. The feature suppression branch.
Sensors 23 01199 g005
Figure 6. Confusion matrix on Pascal VOC → Clipart transfer.
Figure 6. Confusion matrix on Pascal VOC → Clipart transfer.
Sensors 23 01199 g006
Figure 7. Visualization of the localization ability of multi-label classification. (ac) are heat maps of attention for the FA branch only, FS branch only, and Proposed Method, respectively.
Figure 7. Visualization of the localization ability of multi-label classification. (ac) are heat maps of attention for the FA branch only, FS branch only, and Proposed Method, respectively.
Sensors 23 01199 g007
Figure 8. Confusion Matrix on Cityscapes → Foggy Cityscapes transfer.
Figure 8. Confusion Matrix on Cityscapes → Foggy Cityscapes transfer.
Sensors 23 01199 g008aSensors 23 01199 g008b
Figure 9. The ablation study for the scales of different target domain inputs. The results are from Pascal VOC → Clipart. The scale of image is fixed to be 640 pixels in the source domain, and we resize the different scales of image in the target domain, as shown in the X-axis.
Figure 9. The ablation study for the scales of different target domain inputs. The results are from Pascal VOC → Clipart. The scale of image is fixed to be 640 pixels in the source domain, and we resize the different scales of image in the target domain, as shown in the X-axis.
Sensors 23 01199 g009
Figure 10. The ablation study of different network depths. The results are from Pascal VOC → Clipart. The image scale has been fixed at 640 pixels, and we evaluated the accuracy and speed of different depth networks.
Figure 10. The ablation study of different network depths. The results are from Pascal VOC → Clipart. The image scale has been fixed at 640 pixels, and we evaluated the accuracy and speed of different depth networks.
Sensors 23 01199 g010
Figure 11. Error analysis. Results are reported using the Pascal VOC → Clipart setup. Four types: Cls (classification error), Loc (localization error), Bkg (background error), and Miss (missed GT error). From left to right, it is source only, feature alignment branch only, feature suppression branch only, and DAB, respectively.
Figure 11. Error analysis. Results are reported using the Pascal VOC → Clipart setup. Four types: Cls (classification error), Loc (localization error), Bkg (background error), and Miss (missed GT error). From left to right, it is source only, feature alignment branch only, feature suppression branch only, and DAB, respectively.
Sensors 23 01199 g011
Figure 12. Qualitative results (Pascal VOC → Clipart). The Pascal VOC dataset is used as source domain, and the Clipart dataset is used as target domain. From left to right are the results from ground truth, source only, DA-Faster, and proposed method.
Figure 12. Qualitative results (Pascal VOC → Clipart). The Pascal VOC dataset is used as source domain, and the Clipart dataset is used as target domain. From left to right are the results from ground truth, source only, DA-Faster, and proposed method.
Sensors 23 01199 g012
Figure 13. Qualitative results (Pascal VOC→DT Clipart). The Pascal VOC dataset is used as source domain, and the DT Clipart dataset is used as target domain. From left to right are the results from ground truth, source only, DA-Faster, and proposed method.
Figure 13. Qualitative results (Pascal VOC→DT Clipart). The Pascal VOC dataset is used as source domain, and the DT Clipart dataset is used as target domain. From left to right are the results from ground truth, source only, DA-Faster, and proposed method.
Sensors 23 01199 g013
Figure 14. Qualitative results (Cityscapes→Foggy Cityscapes). The Cityscapes dataset is used as source domain, and the Foggy Cityscapes dataset is used as target domain. From left to right are the results from ground truth, source only, DA-Faster, and proposed method.
Figure 14. Qualitative results (Cityscapes→Foggy Cityscapes). The Cityscapes dataset is used as source domain, and the Foggy Cityscapes dataset is used as target domain. From left to right are the results from ground truth, source only, DA-Faster, and proposed method.
Sensors 23 01199 g014
Table 1. The details of samples in the datasets.
Table 1. The details of samples in the datasets.
DatasetsTrainValTestTotal
Pascal VOC [42]2501251049529963
Clipart [43]6002002001000
Watercolor [43]10005005002000
DT Clipart [43]2501251049529963
Cityscapes [44]297550015255000
Foggy Cityscapes [44]297550015255000
Table 2. The mean Average Precision (mAP) of different models on Clipart val set for Pascal VOC → Clipart transfer. The best results are in bold.
Table 2. The mean Average Precision (mAP) of different models on Clipart val set for Pascal VOC → Clipart transfer. The best results are in bold.
MethodsAeroBikeBirdBoatBottleBusCarCatChairCowTableDogHorseMbikePersonPlantSheepSofaTrainTVmAP
DA-Faster [5]15.034.612.411.919.821.123.23.122.126.310.610.019.639.434.629.31.017.119.724.819.8
SWDA [6]26.248.532.633.738.554.337.118.634.858.317.012.533.865.561.652.09.324.954.149.138.1
SCL [46]44.750.033.627.442.255.638.319.237.969.030.126.334.467.361.047.921.426.350.147.341.5
HTCN [47]33.658.934.023.445.657.039.812.039.751.321.120.139.172.863.043.119.330.150.251.840.3
ATF [40]41.967.027.436.441.048.542.013.139.275.133.47.941.256.261.450.642.025.053.139.142.1
UMT [45]39.659.132.435.045.161.948.47.546.067.621.429.548.275.970.556.725.928.939.443.644.1
Source only5.450.27.915.940.616.222.10.336.81.725.68.310.036.734.846.411.024.010.131.321.8
FA branch only13.767.433.338.633.163.952.315.540.354.434.828.342.153.876.256.341.925.152.656.844.1
FS branch only27.760.638.648.342.865.159.316.641.452.940.530.436.859.375.756.848.022.549.544.945.9
Proposed Method25.275.731.842.332.570.857.218.342.273.742.525.741.165.977.458.047.933.752.553.448.4
oracle55.278.351.158.160.758.461.527.360.971.760.540.756.982.582.865.949.246.159.758.159.3
Table 3. The mean Average Precision (mAP) of different models on DT Clipart test set for Pascal VOC → DT Clipart transfer. The best results are in bold.
Table 3. The mean Average Precision (mAP) of different models on DT Clipart test set for Pascal VOC → DT Clipart transfer. The best results are in bold.
MethodsAeroBikeBirdBoatBottleBusCarCatChairCowTableDogHorseMbikePersonPlantSheepSofaTrainTVmAP
ADDA [48]20.150.220.523.611.440.534.92.339.722.327.110.431.753.646.632.118.021.123.618.327.4
CDWS [49]23.360.124.941.526.453.044.04.145.351.539.511.640.462.261.137.120.939.638.436.038.0
Source only4.444.97.815.633.527.122.10.437.99.223.510.413.640.840.343.017.014.723.640.623.5
FA branch only64.677.240.729.726.061.677.858.533.548.953.044.870.577.769.827.737.838.662.241.952.1
FS branch only62.075.837.326.531.368.280.260.237.353.252.341.863.776.370.533.138.040.467.246.453.1
Proposed Method67.178.539.230.626.370.576.963.837.752.357.845.168.374.471.030.545.143.467.347.354.7
Table 4. The domain generalization ability on unknown domains. The best results are in bold.
Table 4. The domain generalization ability on unknown domains. The best results are in bold.
MethodsPascal VOC & Watercolor → Clipart
BicycleBirdCarCatDogPerson
MT [50]64.823.434.63.122.061.4
AT [51]78.630.140.310.932.672.8
Proposed Method60.430.652.914.422.466.5
Table 5. The mean Average Precision (mAP) of different models on Foggy Cityscapes val set for Cityscapes → Foggy Cityscapes transfer. The best results are in bold.
Table 5. The mean Average Precision (mAP) of different models on Foggy Cityscapes val set for Cityscapes → Foggy Cityscapes transfer. The best results are in bold.
MethodsPersonRiderCarTruckBusTrainMotorBikemAP
DA-Faster [5]31.941.646.420.132.017.523.134.630.9
SCDA [37]33.538.048.526.539.023.328.033.633.8
SWDA [6]29.942.343.524.536.232.630.035.334.3
MTOR [39]30.641.444.021.938.640.628.335.636.0
iFan [53]32.640.048.527.945.531.722.833.035.3
HTCN [47]33.247.547.931.647.440.932.337.139.8
GPA [24]32.946.754.124.745.741.132.438.739.5
SFA [54]46.548.662.625.146.229.428.344.041.3
UMT [45]33.046.748.634.156.546.830.437.341.7
AFAN [55]42.544.657.026.448.028.333.237.139.6
DIR [56]36.945.849.428.244.634.935.138.939.2
TDD [52]39.647.555.733.847.642.137.041.443.1
Source only27.631.448.921.233.816.919.923.127.9
FA branch only40.441.764.830.655.356.725.933.543.6
FS branch only41.242.164.232.054.449.929.834.143.5
Proposed Method46.146.568.935.657.150.835.238.747.4
oracle51.249.271.940.157.756.340.142.351.1
Table 6. The ablation study of the trade-off parameter. The best results are in bold.
Table 6. The ablation study of the trade-off parameter. The best results are in bold.
EXP. λ μ Scale AP 0.5 AP 0.5 : 0.95 Speed/ms
10.990.9951243.223.814.9
20.0990.09951248.426.523.7
30.00990.009951247.026.319.2
Table 7. The ablation study of Conv kernel.
Table 7. The ablation study of Conv kernel.
EXP.Conv KernelScaleParams/M AP 0.5 AP 0.5 : 0.95 Speed/msFLOPs/B
1[16,32,64,128,256]6401.7935.217.415.14.2
2[32,64,128,256,512]6407.138.819.517.315.9
3[48,96,192,384,768]64020.946.422.422.448.1
4[64,128,256,512,1024]64046.244.423.619.1108.6
5[80,160,320,640,1280]64086.346.225.937.4204.2
Table 8. The ablation study of the DAB structure.The best results are in bold.
Table 8. The ablation study of the DAB structure.The best results are in bold.
MethodScale AP 0.5 AP 0.5 : 0.95 Speed/ms
FA branch only51245.925.614.2
FS branch only51244.125.414.7
Proposed Method51248.426.523.7
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liu, X.; Zhang, B.; Liu, N. Cross-Domain Object Detection by Dual Adaptive Branch. Sensors 2023, 23, 1199. https://doi.org/10.3390/s23031199

AMA Style

Liu X, Zhang B, Liu N. Cross-Domain Object Detection by Dual Adaptive Branch. Sensors. 2023; 23(3):1199. https://doi.org/10.3390/s23031199

Chicago/Turabian Style

Liu, Xinyi, Baofeng Zhang, and Na Liu. 2023. "Cross-Domain Object Detection by Dual Adaptive Branch" Sensors 23, no. 3: 1199. https://doi.org/10.3390/s23031199

APA Style

Liu, X., Zhang, B., & Liu, N. (2023). Cross-Domain Object Detection by Dual Adaptive Branch. Sensors, 23(3), 1199. https://doi.org/10.3390/s23031199

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop