Next Article in Journal
Assessing Upper Limb Function in Breast Cancer Survivors Using Wearable Sensors and Machine Learning in a Free-Living Environment
Previous Article in Journal
An Improved Distributed Sampling PPO Algorithm Based on Beta Policy for Continuous Global Path Planning Scheme
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Domain Adaptation Based on Semi-Supervised Cross-Domain Mean Discriminative Analysis and Kernel Transfer Extreme Learning Machine

College of Information Engineering, Henan University of Science and Technology, Kaiyuan Avenue, Luoyang 471023, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(13), 6102; https://doi.org/10.3390/s23136102
Submission received: 29 May 2023 / Revised: 22 June 2023 / Accepted: 28 June 2023 / Published: 2 July 2023
(This article belongs to the Section Intelligent Sensors)

Abstract

:
Good data feature representation and high precision classifiers are the key steps for pattern recognition. However, when the data distributions between testing samples and training samples do not match, the traditional feature extraction methods and classification models usually degrade. In this paper, we propose a domain adaptation approach to handle this problem. In our method, we first introduce cross-domain mean approximation (CDMA) into semi-supervised discriminative analysis (SDA) and design semi-supervised cross-domain mean discriminative analysis (SCDMDA) to extract shared features across domains. Secondly, a kernel extreme learning machine (KELM) is applied as a subsequent classifier for the classification task. Moreover, we design a cross-domain mean constraint term on the source domain into KELM and construct a kernel transfer extreme learning machine (KTELM) to further promote knowledge transfer. Finally, the experimental results from four real-world cross-domain visual datasets prove that the proposed method is more competitive than many other state-of-the-art methods.

1. Introduction

Traditional classification tasks deal with situations where the distribution of source domain samples and target domain samples are the same, and a classifier trained from source domain samples can be directly applied to the target domain samples. Theoretical studies on classifiers are also based on this assumption [1]. However, in real-world environments, the distribution of source domain samples and target domain samples is often different due to factors such as lighting, viewpoints, weather conditions, and cameras [2]. The most advanced classifiers trained on source domain samples may dramatically degrade when applied to target domain samples. One possible solution is to annotate the new data and retrain the model. Unfortunately, labeling a large number of samples is costly and time-consuming. Another promising solution is domain adaptation (DA), which aims to minimize the distribution gap between the two domains and learn a fairly accurate shared classifier for both domains, where the target domain samples do not have available labels.
One strategy of DA is to find invariant feature representation subspaces between domains to minimize distribution divergence. A large number of existing methods applying this strategy learn a shared feature subspace where the distribution mismatch between the two domains is reduced, and then employ standard classification methods in this subspace [3]. Metrics that measure the mismatch in distribution between domains include maximum mean discrepancy (MMD) [4], Kullback–Leibler (KL) divergence [5], Bregman divergence [6], and Wasserstein distance [7], among others. Due to the fact that MMD is a nonparametric estimation criterion for distance, researchers have proposed various adaptive models commonly with this metric. Pan et al. [8] proposed a classical DA method maximum mean difference embedding (MMDE) by defining a semidefinite program (SDP) and introducing the MMD criterion to match marginal distribution. To further reduce the computational burden of the MMDE and to preserve the key features of the data, [9] proposed an improved method, transfer component analysis (TCA), which reduces the differences in marginal distribution between domains by projecting the data into a latent feature space. Gong et al. [10] developed geodesic flow kernel (GFK), which is another classical method. GFK establishes connections between domains based on a series of intermediate subspaces instead of using only one subspace. In addition to marginal distributions, the joint distribution adaptive (JDA) [11] approach also considers conditional distribution matching, where target samples are iteratively assigned pseudo-labels. Based on JDA, various variants have been developed. Lu et al. [12] introduced weight adaptation into JDA, paid attention to the feature representation of a single sample, emphasized the difference between samples in the target domain learning, and proposed the weights adaptation joint distribution adaptation (W-JDA). A joint distribution adaptation based transfer network with diverse feature aggregation (JDFA) is proposed by Jia et al. [13], which enhances feature extraction capability across large domain gaps through the diverse feature aggregation module; then, MMD was employed to reduce distribution differences.
Huang et al. [14] proposed the extreme learning machine (ELM) model, which is a “generalized” single-hidden-layer feedforward neural network that has been employed to address classification and regression problems. The hidden layer nodes of the ELM are randomly initialized and without tuning, and the output weights between the hidden and output layers can be analytically determined [15]. Many researchers have extended it in theory and application, and proposed numerous new ELMs, such as AdaBoost ELM [16], kernel-based multilayer ELM (ML-KELM) [17], voting-based ELM (V-ELM) [18], Kalman filter ELM (KA-ELM) [19], and others. However, the above algorithms require a prerequisite assumption that there is consistency in the distribution of labeled training samples and testing samples, and therefore these classical ELM algorithms cannot handle domain adaptation problem. To address this issue, some improvements have been made to ELM. Zang et al. [20] proposed a two-stage transfer limit learning machine (TSTELM) framework. The framework uses MMD to reduce the distribution differences of output layers between domains in the statistical matching stage. In the subspace alignment stage, the source and target model parameters are aligned, and output weight approximation is added to further promote knowledge transfer across domains. Chen et al. [21] developed a DST-ELM method to learn a new feature space. With the target domain discrimination information fixed in the feature space, the source domain samples are adapted to the target domain and the MMD distance between them is minimized. The MMD metric has been used in the above methods to measure the distribution differences between the source domain samples and the target domain samples as a whole, but they ignore the impact of individual samples on the distribution differences.
To sum up, there are several challenges to DA. The first issue is the similarity between source domain and target domain. To cope with this issue, suitable distribution distance measurement criteria are found to eliminate this dissimilarity. Secondly, there is class separability and shared feature extraction between source and target domains in the feature subspace. The third challenge is the knowledge transfer ability of the classifier.
We believe that a well-developed DA algorithm is able to address these three challenges. Therefore, we want to tackle all these problems simultaneously. We propose a domain adaptation approach which consists of two stages: feature-based adaptation and classifier-based adaptation. At the former stage, we develop semi-supervised cross-domain mean discriminative analysis (SCDMDA). It introduces cross-domain mean approximation (CDMA) [17] into SDA to minimize the marginal and conditional interdomain distribution discrepancy. Meanwhile, the category separability of LDA and the original structure information of graph regularizers are used together to capture shared features between the source and target domains. At the latter stage, we first choose KELM as the classification model. Then, a kernel transfer extreme learning machine (KTELM) is designed by adding a cross-domain mean constraint term into KELM to further enhance the knowledge transfer ability of our method, as shown in Figure 1. Finally, we investigate the performance of our approach on public domain adaptation datasets, and the experimental results show that it can learn better domain-invariant features with high accuracy and outperform other existing shallow and deep domain adaptation methods.
In this paper, we make the following contributions:
  • We introduce CDMA into SDA and then propose SCDMDA. It extracts shared discriminative features across domains by using CDMA to minimize the marginal and conditional discrepancies between domains and applying SDA to exploit the label information and original structure information.
  • We present KTELM by designing a cross-domain mean approximation constraint into KELM for classification in domain adaptation.
  • We obtain a classifier with the ability of knowledge transfer by combining SCDMDA and KTELM and implement a classification task on public image datasets. The results show the superiority of our approach.
The rest of this paper is as follows: In Section 2, we briefly describe SDA, ELM, KELM, and DA. Section 3 provides SCDMDA and KTELM algorithms. The experimental results and analysis are presented in Section 4. Finally, Section 5 is the conclusion of this paper.

2. Preliminary

In this section, we will briefly introduce SDA, ELM, KELM, and DA.

2.1. Semi-Supervised Discriminant Analysis (SDA)

As a classical semi-supervised feature extraction algorithm, SDA extends LDA by adding a geometrical regularizer to prevent overfitting. Given a sample set X = X l X u = x r , y r r = 1 n l x v v = 1 n u R d × N , it is divided into two parts: a labeled subset X l and an unlabeled subset X u . x r is a labeled sample with a label y r , and x v is an unlabeled sample. n l and n u are the number of labeled and unlabeled samples, respectively, N = n l + n u . The dimension of the sample is d . We can define the objective function of SDA as follows:
J 1 ( P ) = arg   max P T S b P P T S w P + α P T X Q X T P
where P is the projection matrix; Q = D G is Laplacian matrix. G is the graph neighbor similarity matrix of X . G i j = 1   ,   i f   x i N p ( x j )   o r   x j N p ( x i ) 0   ,   o t h e r w i s e , where x i and x j are among p nearest neighbors of each other, and N p ( x i ) denotes the set of p nearest neighbors of x i . D is a diagonal matrix; its entries are a column (or a row, since G is symmetric) sum of G , D i i = j G i j . α is the parameter balancing the graph regularization. S b and S w represent the interclass scatter matrix and between-class scatter matrix, respectively, which are calculated as follows:
S b = c = 1 C N c ( m ( c ) u ) ( m ( c ) u ) T
S w = c = 1 C i = 1 N c ( x i ( c ) m ( c ) ) ( x i ( c ) m ( c ) ) T
where u = 1 N i = 1 N x i , and m ( c ) = 1 N c i = 1 N c x i ( c ) . C is the number of class, x i ( c ) is the i -th sample belonging to class c , N c represents the number x i ( c ) , and m ( c ) is the mean value of samples of class c .

2.2. Extreme Learning Machine (ELM) and Kernel Extreme Learning Machine (KELM)

As a high-performance classifier, ELM with a single hidden layer has a fast training process and high classification performance. Given a dataset ( x i , y i ) i = 1 N with N samples x i and its label y i , we could construct an ELM network with L hidden nodes as follows:
y i T = j = 1 L W j T g ( w j x i + b j )
where y i is the output prediction result of sample x i , and W j is the output weight connecting the hidden layer and output layer. g ( x ) is an activation function to improve the nonlinear fitting ability of the network. w j and b j are input weight and bias, between the input layer and the hidden layer, which can be set randomly. It can be seen that W j is the key parameter of an ELM network. If we want to obtain an optimal result, the following objective function needs to be solved:
J 2 ( W )   = min W 1 2 W 2 + λ 2 Y H W 2
In Equation (5), H = [ g ( x i ) T , , g ( x N ) T ] T , W = W 1 , , W L T , and Y = y 1 , , y N T . W 2 is a regularizer to avoid model overfitting, and Y H W 2 is the training error. λ is a tradeoff parameter between minimizing training errors and the regularizer.
We consider Equation (5) to be a regular least square problem and set J 2 W = 0 . The optimal solution of W is calculated as follows:
W = ( H T H + I L λ ) 1 H T Y , N > L H T ( H T H + I N λ ) 1 Y , N L
Then, the output of the trained ELM model corresponding to test sample x T e is
y T e =   sign ( h T e T W ) arg max ( h T e T W ) ,   f o r   b i n a r y   c l a s s i f i c a t i o n ,   f o r   m u l t i c l a s s i f i c a t i o n
where h T e = g ( x T e ) .
If the activation function g ( x ) is unknown, we let K = H H T , K ( i j ) = g ( x i ) g ( x j ) T = Ω ( x i , x j ) , Ω ( x i , x j ) is a kernel function. The output of KELM can be defined as:
y T e =   sign g x T e H T K + I L λ 1 Y   ,   f o r   b i n a r y   c l a s s i f i c a t i o n arg max g x T e H T K + I N λ 1 Y , f o r   m u l t i c l a s s i f i c a t i o n
where g x T e H T = Ω ( x , x 1 ) Ω ( x , x N ) .

2.3. Domain Adaptation (DA)

In practical application scenarios, it is difficult to collect enough training samples with the same distribution as the testing samples. For example, in photography, it is hard to avoid changes in pixel distribution caused by changes in lighting, posture, weather, and background. If a large number of samples with the same distribution are selected and manually marked for model training, the cost will be high [22]. The above distribution changes will generate a domain bias between the training dataset and the testing dataset, which could degrade the performance of traditional machine learning methods. DA can not only apply knowledge from other domains (source domains) that contain substantial labeled samples and are related to the target domain for model training, but also eliminate domain bias. Existing DA methods are mainly divided into instance-based adaptation, feature-based adaptation, and classifier-based adaptation.
The instance-based adaptation methods seek a strategy which selects “good” samples in the source domain to participate in model training and suppresses “bad” samples to prevent negative transfer. Kernel mean matching (KMM) [23] minimizes the maximum mean difference [24] and weighs the source data and target data in the reproducing kernel Hilbert space (RKHS), which can correct the inconsistent distribution between domains. As a classic instance-based adaptation method, transfer adaptive boosting (TrAdaBoost) [25] extends the AdaBoost algorithm to weigh source-labeled samples and target-labeled samples to match the distributions between domains. In the work [26], a simple two-stage algorithm was proposed to reweight the results of testing data from the training classifier using their signed distance to the domain separator. Moreover, it applied manifold regularization to propagate the labels of target instances with larger weights to those with smaller weights. Instance-based adaptation methods are more efficient for knowledge transfer, but negative transfer [27] can easily occur when there are no shared samples between domains.
The feature-based adaptation methods attempt to learn a subspace with a better shared features representation, in which the marginal or conditional distribution divergences between domains are minimized to facilitate knowledge transfer. Transfer component analysis (TCA), joint distribution adaptation (JDA), balanced distribution adaptation (BDA), and multiple kernel variant of maximum mean discrepancy (MK-MMD) apply the maximum mean discrepancy (MMD) strategy to perform distribution matching, thus minimizing the marginal or conditional distribution divergences between domains [28]. In [29], Wang et al. use the locality preserving projection (LPP) to learn a joint subspace from the source and target domains.
In recent years, deep neural networks (DNNs) have performed well in high-level semantic feature extraction with strong expression ability, and have broad application prospects in domain adaptation. Domain adaptation networks (DAN) [30], faster domain adaptation networks (FDAN) [31], residual transfer networks (RTN) [32], domain adversarial neural networks (DANN) [33], and collaborative and adversarial networks (CAN) [34] use MMD to align distribution discrepancies in an end-to-end learning paradigm for feature extraction across domains. However, DNNs also have their shortcomings, for example, the models contain huge parameters and require a large amount of data for training, which is not suitable for small sample data. DNNs have a strong ability to fit data and can approximate any complex objective function, but the computational cost is high. Our method in this paper has a small number of parameters and fast calculation speed, and it is suitable for small sample data classification.
The goal of the classifier-based adaptation methods is to adjust the parameters of the classifier so that the classifier trained on the source domain has good performance in the target task. Adaptive support vector machine (Adapt-SVM) [35] introduced a regularizer into SVM to minimize the classification error and the discrepancy between two classifiers trained on the source- and target-labeled samples. Based on LS-SVM, Tommasi et al. [36] presented an adaptation model called multi-model knowledge transfer (Multi-KT), in which the objective function imposed the requirement that the target classifier and a linear combination of the source classifiers be close to each other for knowledge transfer. Considering the simplicity and efficiency of ELM, many adaptation models based on ELM have been proposed for domain adaptation, such as joint domain adaptation semi-supervised extreme learning machine (JDA-S2ELM) [37], cross-domain ELM (CDELM) [38], domain space transfer ELM (DST-ELM) [21], domain adaptation extreme learning machine (DAELM) [39], and so on.
In addition, the adaptation model based on the generative adversarial network (adversarial-based adaptation) [34,40,41,42] has recently shown exceptional performance. Many target samples with transferable and domain-invariant features are produced through the generator of adversarial learning and then applied to confuse the domain discriminator training on the source domain and optimize the generator. It can effectively reduce the distribution differences between domains and transfer knowledge efficiently. Since generating samples with domain-invariant features is the main task of adversarial-based adaptation, it is commonly attributed to feature-based adaptation.
In this paper, our method is a combination of feature- and classifier-based adaptation. SCDMDA belongs to a shallow feature-based adaptation method, and KTELM is a classifier-based adaptation method. At the feature-based adaptation stage, SCDMDA finds domain-invariant features. It not only applies SDA to maximize between-class scatter, minimize within-class scatter, and keep the original structure information by pseudo-labeled target samples and labeled source samples, but it also reduces the distribution differences between domains using CDMA. At the classifier-based adaptation stage, a cross-domain mean constraint term is added to KELM to further enhance its knowledge transfer ability.

3. Proposed Method

In a domain adaptation environment, given a source domain with labeled samples X S = x i , y i i = 1 n S R d × n S and a target domain with unlabeled samples X T = x j j = 1 n T R d × n T , where the distributions between X S and X T are not the same but are related and the label space Y = y c c = 1 C is shared, the number of categories is denoted by C . To address the domain adaptation problem, our proposed method is divided into two stages: feature-based adaptation and classifier-based adaptation.
At the feature-based adaptation stage, we design SCDMDA, and its goal is to find an optimal transformation matrix P = ( p 1 , p 2 , , p k ) R d × k , which embeds X S and X T into low-dimensional subspaces z S = P T x S R k and z T = P T x T R k , where k is the subspace dimension. In this subspace, the distributions between Z S and Z T are closer than those between X S and X T .
At the classifier-based adaptation stage, KTELM is constructed by adding a cross-domain mean constraint term into KELM, which improves the knowledge transfer ability of KELM.

3.1. Feature-Based Adaptation of SCDMDA

3.1.1. Cross-Domain Mean Approximation (CDMA)

From Equation (1) in Section 2.1, we can see that although SDA has better feature extraction performance than LDA, it can obtain bad feature representation when the training samples and testing samples have inconsistent data distributions. To address this issue, we need to reduce the distribution discrepancy of training samples and testing samples with the help of domain adaptation technology based on CDMA. It is expressed as follows:
L C D M A P = min P i = 1 n S + n T P x i P μ 2 2 + c = 1 C i = 1 n S i ( c ) + n T i ( c ) P x i ( c ) P μ ( c ) 2 2 = min P tr P T M 0 P + P T M c P s . t . P T X T G X P = I
where μ = μ T ,           if       x i X S μ S ,           if       x i X T and μ ( c ) = μ T ( c ) ,           if       x i ( c ) X S ( c ) μ S ( c ) ,           if       x i ( c ) X T ( c ) . μ T S is the mean vector of the target (source) domain samples, and μ T S ( c ) is the mean vector of the target (source) domain X T S ( c ) with c category. n S i ( c ) and n T i ( c ) are the number of samples in X S ( c ) and X T ( c ) . M 0 = X T X X T μ X μ T + μ T μ , M c = c = 1 C   ( X ( c ) T ( X ( c ) ) ( X ( c ) ) T μ ( c ) X ( c ) ( μ ( c ) ) T + μ ( c ) T ( μ ( c ) ) ) , X = X S X T , and X ( c ) = X S ( c ) X T ( c ) . G = I 1 n q q T is the center matrix, where q R n is the column vector with element 1, n = n S + n T . In Equation (9), CDMA adapts both the marginal distribution in the first term and the conditional distribution in the second term.

3.1.2. Semi-supervised Cross-Domain Mean Discriminative Analysis (SCDMDA)

To enhance the knowledge transfer ability of SDA, we combine Equation (9) and Equation (1) and design the objective function of SCDMDA for shared feature extraction across domains as follows:
J 3 ( P ) = min P   tr P T β S w + α X Q X T P β P T S b P + P T X M 0 + M c X T P + λ 1 P F 2 s . t .                   P T X T G X P = I
Furthermore, the above formula is equivalent to:
J 3 ( P ) = min P   tr P T X T M 0 + M c X + α X T Q X + β S w S b P + λ 1 P F 2 s . t .                             P T X T G X P = I
where P F 2 with tradeoff parameter λ 1 is introduced into the objective function to ensure the sparsity of matrix P . α and β balance the influence of the graph regularizer and LDA, respectively.
We set J 3 P = 0 , and Equation (11) becomes a generalized eigen-decomposition problem:
X T M 0 + M c + α Q X + β S w S b + λ 1 Ι P = G P Φ
Finally, Equation (12) is used to find k , the smallest eigenvectors for the optimal adaptation matrix P , to reduce the difference distribution between domains.
To address the issue of the nonlinear data, we transform Equations (11) and (12) into a high-dimensional space by kernel mapping, namely x m a p p i n g φ ( x ) . The kernel matrix corresponding to the dataset X is K = φ T ( X ) φ ( X ) , then we obtain:
J 3 ( P ) = min P   tr P T K T M 0 + M c K + α K T Q K + β S w S b P + λ 1 P F 2 s . t .                             P T K T G K P = I
K T M 0 + M c + α Q K + β S w S b + λ 1 Ι P = G P Φ
where S w φ and S b φ are the nonlinear forms of S w and S b under the kernel map, respectively, which can be obtained through Ref. [43]. A complete procedure of SCDMDA is summarized in Algorithm 1.
Algorithm 1. SCDMDA
Input: Dataset X S and X T , subspace dimension k , parameters α , β , λ 1 , θ , and λ 2 , classifier KTELM, maximum number of iterations T max .
Output: Projection matrix P and target prediction Y ˜ T .
Step1: According to Equations (1) and (9), construct Q , G , and M 0 , and set S w and S b to 0.
Step2: Let t = 1 .
Step3: Solve Equation (12) or Equation (14) to obtain the projection matrix P .
Step4: Project X S and X T by P into k -dimensional subspace to obtain Z S and Z T .
Step5: Learn a KTELM on { Z S , Y S } , and classify Z T to obtain the label set of the target domain data Y ˜ T .
Step6: Use { X S , Y S } and { X T , Y ˜ T } , construct M c , and solve S w and S b ( S w φ and S b φ in the nonlinear case) according to Equations (2) and (3).
Step7: Let t t + 1 .
Step8: If t T max or Y ˜ T does not change, output P , otherwise, go to Step3.

3.2. Classifier-Based Adaptation of KTELM

After feature adaptation of SCDMADA, we can obtain P and obtain Z S = P T X S and Z T = P T X T . Then, it can be seen from Section 2.2 that we can train a KELM on { Z S , Y S } and effectively predict Z T . However, we can see from Equation (5) that KELM has no capacity for knowledge transfer. In this section, we design a cross-domain mean constraint on the source domain as follows:
J 4 ( W ) = min W H S W H Tav W 2 + c = 1 C H S ( c ) W H Tav ( c ) W 2
where H S ( c ) is the samples dataset with c category in H S , and H Tav ( c ) denotes the mean of the samples dataset with c category in H T . Then, we add Equation (15) into Equation (5) and obtain
J 5 ( W )   = min W 1 2 W 2 + λ 2 2 Y S H S W 2 + θ H S W H Tav W 2 + c = 1 C H S ( c ) W H Tav ( c ) W 2
where λ 2 and θ balance the influence of training errors and cross-domain mean constraint on the source domain, respectively.
Similar to ELM, we set J 5 W = 0 and obtain
W = H S T H S T H S + θ H S H Tav H S H Tav T + c = 1 C H S ( c ) H Tav ( c ) H S ( c ) H Tav ( c ) T + I N λ 2
We kernelize Equation (17) and let K S = H S T H S , K S ( i j ) = g ( x S i ) g ( x S j ) T = Ω ( x S i , x S j ) , Ω ( x i , x j ) is a kernel function. The output of KTELM can be defined as:
y T =   sign g x T H T K S + θ φ S φ Tav T φ S φ Tav + c = 1 C φ S ( c ) φ Tav ( c ) T φ S ( c ) φ Tav ( c ) + I L λ 2 1 Y   ,   f o r   b i n a r y   c l a s s i f i c a t i o n arg max g x T H T K S + θ φ S φ Tav T φ S φ Tav + c = 1 C φ S ( c ) φ Tav ( c ) T φ S ( c ) φ Tav ( c ) + I N λ 2 1 Y   ,   f o r   m u l t i - c l a s s i f i c a t i o n
where g x T H T = Ω ( x T j , x 1 ) Ω ( x T j , x n T ) , φ S = Ω ( x S i , x 1 ) Ω ( x S i , x n S ) , and φ Tav is the mean of Ω ( x T j , x 1 ) Ω ( x T j , x n T ) . φ S ( c ) = Ω ( x S i ( c ) , x 1 ( c ) ) Ω ( x S i ( c ) , x n S ( c ) ) , and φ Tav ( c ) is the mean of Ω ( x T j ( c ) , x 1 ( c ) ) Ω ( x T j ( c ) , x n T ( c ) ) .

3.3. Discussion

In this paper, our method is proposed for domain adaptation, which includes feature adaptation based on SCDMDA and classifier adaptation based on KTELM.
  • From Equation (11) and Equation (18), it can be seen that, compared with SDA, SCDMDA adopts CDMA to reduce the distribution discrepancy between domains, which is better for domain adaptation than SDA. Moreover, as a semi-supervised feature extraction method, SCDMDA focuses more attention on individual information with the help of the category separability of LDA and the original structure information of graph regularizers.
  • Compared with MMD, CDMA reflects individual differences. In our method, CDMA mines individual information through M 0 and M c , which is a more effective interdomain distribution difference measurement mechanism than MMD. In addition, we verify the improved method on k-NN, KELM, and KTELM classifiers.
  • In the classical ELM, solving the output weight W that connects the hidden layer and the output layer is the key, and the optimal solution is obtained by solving Equation (5). However, for samples with interdomain distribution differences, the solution obtained by Equation (5) is not the optimal. The domain adaptation is added to the ELM to obtain Equation (16), and the optimal W * can be obtained. By adopting the cross-domain mean constraint on the source domain to achieve the cross-domain transfer of knowledge, the interdomain distribution difference can be reduced, which shows that KTELM has higher domain adaptation accuracy.

4. Experiments and Analysis

In this section, extensive experiments were conducted on four widely adopted datasets, including Office+Caltech [10], Office-31 [44] object recognition, USPS [45] and MNIST [46] digital handwriting, and PIE face [47], to test SCDMDA+KETLM. Table 1 shows some descriptions of these. All the experiments were conducted on a PC equipped with Win10, an Intel i5 10400F, 2.9 GHz CPU, and 8 GB RAM, with software MATLAB 2017b. In the experiment, we repeated the experiment 20 times and recorded the average results.

4.1. Dataset Description

Office + Caltech: It is a visual domain adaptation benchmark dataset, which contains two sub-datasets of Office and Caltech (C), shown in Figure 2. Office contains a total of 4652 pictures with 31 categories from 3 different domain datasets: Amazon (A), DSLR (D), and Webcam (W). In Amazon, each class typically contains an average of 90 images, with an average of 30 images in DSLR or Webcam. Additionally, Caltech is a benchmark dataset for target recognition, containing 30,607 images in 256 categories. In our experiments, we selected 10 shared categories in 4 domains with a total of 2533 images with 800 SURF features. During the experiment, two different domains were selected from the four databases as source and target domain datasets. Thus, we had 12 cross-domain tasks, i.e., C vs. A, C vs. W, C vs. D... and D vs. W.
Office-31: This is a public dataset commonly used to investigate domain adaptation algorithms, which includes three distinct domains: amazon, webcam, and dslr. It contains 4652 images with 31 categories. In our experiments, we carried out a classification task in a 2048-dimensional ResNet-50 feature space and performed six groups of experiments: amazon vs. dslr, amazon vs. webcam, dslr vs. amazon, dslr vs. webcam, webcam vs. amazon, and webcam vs. dslr.
PIE: It is a standard facial dataset collected by Carnegie Mellon University, which contains 41,368 face images with a pixel resolution of 32 × 32 from 68 people. These images depict facial postures, lighting, and expressions (as shown in Figure 3a). To verify the performance of the domain adaptation algorithm, we selected five subsets from the dataset: PIE1 (left pose), PIE2 (upward pose), PIE3 (downward pose), PIE4 (frontal pose), and PIE5 (right pose). In each subset (posture), all face images with different expressions and illuminations are changed. In this paper, we selected the source data and the target data, respectively, and constructed 20 cross-domain tasks: PIE1 vs. PIE2, PIE1 vs. PIE3... PIE5 vs. PIE4.
USPS+MNIST: USPS and MNIST are two closely correlated, yet differently distributed, handwritten datasets, sharing 10 numeric categories of 0–9 (as shown in Figure 3b,c). The USPS dataset contains 7291 training samples and 2007 test samples, each with 16 × 16 pixels. There are 60,000 training images and 10,000 test images with 28 × 28 pixels in the MNIST database. During this experiment, 1800 pictures from USPS and 2000 pictures from MNIST were selected to construct source and target domain datasets. All the pictures were converted into grayscale images with 16 × 16 pixels. We constructed two recognition tasks, i.e., USPS as the source domain and MNIST as the target domain (USPS vs. MNIST) and vice versa (MNIST vs. USPS).

4.2. Experiment Setting

To evaluate the performance of our algorithm, we compared our approach with several state-of-the-art domain adaptation approaches categorized into three classes: the traditional methods, e.g., 1-NN [48], KELM, and SDA; the shallow feature adaptation methods, e.g., GFK, JDA, STDA-CMC [49], W-JDA, and JDA-CDMAW [50]; and the deep transfer methods, e.g., DAN, DANN, and CAN. In addition, our method is divided into three cases: SCDMDA0 represents SCDMDA+1NN, SCDMDA1 represents SCDMDA+KELM, and SCDMDA2 represents SCDMDA+KTELM. In addition to KELM, SCDMDA1, and SCDMDA2, other algorithms all choose 1-NN as the basic classifier. We applied classification accuracy as the evaluation metric for each algorithm and its formula is as follows:
A c c u r a c y = c o r r e c t l y _ c l a s s i f i e d _ s a m p l e s t o t a l _ s a m p l e s × 100 %
To achieve optimal functionality of SCDMDA and KTELM, we set k = 100 , 100 , 100 , 512 on Office+Caltech, USPS+MNIST, PIE, and Office-31 datasets, respectively, T max = 20 , α 10 2 , 10 0 , β 10 2 , 10 1 , λ 1 10 1 , 10 3 , θ 10 3 , 10 0 , and λ 2 10 2 , 10 0 for all datasets. For comprehensive comparison, the results of some methods are provided in their original papers or their publicly available codes. We tuned parameters according to their default parameters in released codes or their respective literature.

4.3. Results and Analysis

We applied our method, compared algorithms, and performed classification experiments on the Office+Caltech, USPS+MNIST, PIE, and Office-31 datasets. The results are presented in Table 2 and Table 3. From Table 2, conclusions can be obtained as follows:
  • Table 2 summarizes the accuracies of all methods on Office+Caltech, PIE, and USPS+MNIST datasets with shallow feature representation, and the optimal result of each row in the table among all the methods is presented in bold. Our method SCDMDA (0–2) outperforms any other compared methods. The total average accuracy of SCDMDA2 on the 34 tasks is 74.9%, which achieves a 28.75% improvement compared with the baseline SDA. This verifies that CDMA is effective in reducing the between-domain distribution discrepancy and improving the knowledge transfer ability of SDA.
  • SCDMDA2 outperforms SCDMDA0 and SCDMDA1, indicating that KTELM and KELM have higher accuracy than 1-NN, and that the cross-domain mean constraint on the source domain is effective for domain adaptation.
  • Similarly, the semi-supervised feature extraction method SCDMDA0 works better than SWTDA-CMC. An explanation is that CDMA is a better distribution discrepancy measurement criterion than MMD, so our method works quite well in reducing domain bias. SCDMDA0 and SWTDA-CMC are better than JDA, JDA-CDMAW, and W-JDA. This illustrates that the category separability of LDA and the original structure information of graph regularizers are important for the classification task.
  • The accuracy of SDA, GFK, JDA, STDA-CMC, W-JDA, JDA-CDMAW, and SCDMDA (0–2) outperforms 1-NN in most cases, showing the importance of feature extraction in classification tasks. Most domain adaptation techniques such as JDA, STDA-CMC, W-JDA, JDA-CDMAW, and SCDMDA (0–2) achieve higher accuracy than SDA, due to cross-domain shared feature extraction with few or no same distribution samples in the target domain. The accuracy of JDA, STDA-CMC, W-JDA, JDA-CDMAW, and SCDMDA (0–2) generally improves shared feature extraction through joint distribution adaptation.
In addition, results on the Office-31 dataset with ResNet-50 features are shown in Table 3. We observe the classification results obtained by 1-NN, KELM, JDA, STDA-CMC, JDA-CDMAW, and SCDMDA (0–2) and some deep domain adaptation methods such as DAN, DANN, and CAN to verify the effectiveness of our approach on this more complex dataset. SCDMDA2 outperforms other methods on all six groups of domain adaptation tasks. In particular, for the amazon vs. dslr domain adaptation task, it achieves a higher accuracy (91.16%). It further validates that our method, equipped with deep generic features, can further reduce the cross-domain distribution discrepancy and achieve the best adaptation performance, demonstrating the potential of our method.
Table 4 shows the running-time comparisons of SCDMDA (0–2) with 1-NN, KELM, JDA, STDA-CMC, and JDA-CDMAW on the PIE1 vs. PIE2 dataset. The following can be seen: (1) Because JDA, STDA-CMC, JDA-CDMAW, and SCDMDA (0–2) require 20 iterations for label refinement, these approaches consume more time than 1-NN, and KELM. (2) JDA-CDMAW needs less running time than JDA due to no requirement for the construction of the interdomain distribution divergence matrix. (3) SCDMDA (0–2) and STDA-CMC take more time than JDA-CDMAW to obtain the within- and between-class scatter matrices and graph Laplacian matrix (4) The time cost of SCDMDA2 requires more time than SCDMDA1 to compute the cross-domain mean constraint term for domain adaptation. Because KELM is faster than 1-NN, SCDMDA1 needs less time than SCDMDA0.

4.4. Sensitivity Analysis

In this section, we conducted some experiments to investigate the influence on the accuracy of SCDMDA2, including the number of dimensions of the subspace, the graph tradeoff parameter α , the SDA tradeoff parameter β , and the sparse regularization parameter λ 1 in SCDMDA; and the cross-domain mean constraint control parameter θ and the model regular tradeoff parameter λ 2 of KETLM. In addition, we needed to observe the convergence of the designed algorithm. Then, we carried out experimental analysis on four cross-domain tasks, including A vs. D, MNIST vs. USPS, PIE1 vs. PIE2, and amazon vs. dslr datasets, and the results are shown in Figure 4.
In Figure 4a, we observe that the accuracy of SCDMDA2 grows with the increasing number of iterations on the four sets of datasets and gradually stabilizes after the 8th iteration, verifying its good convergence.
Figure 4b illustrates the experiments with the subspace dimension k 20 , 40 , , 200 on 3 datasets, which indicated the impact of the subspace dimension on our method. It can be observed that SCDMDA2 obtains strong robustness when k 80 , 120 . Moreover, we did not perform this experiment on the amazon vs. dslr datasets, because the Office-31 dataset with ResNet-50 feature is robust with regard to different dimensions for feature transfer learning.
We ran SCDMDA2 with other parameters fixed α [ 10 6 , 10 1 ] . From Figure 4c, we can see that the accuracy rises first and then falls with α increasing and obtains optimal performance when α = 10 2 on 4 datasets. When α is too small, the original structure information cannot be exploited. On the contrary, if it is too large, other terms in Equation (13) will not work and hurt the proposed algorithm.
Similar situations also occur in Figure 4d–g. The accuracy curve increases first and then decreases with the increase of β , λ 1 , θ , and λ 2 . When β [ 10 2 , 10 1 ] , λ 1 [ 10 1 , 10 2 ] ,
θ [ 10 2 , 10 0 ] , and λ 2 [ 10 2 , 10 1 ] , SCDMDA2 works well. If β , λ 1 , θ , and λ 2 are too small, label information will be lost, the sparsity of the projection matrix fails to be controlled to reach the optimum, the cross-domain mean constraint term does not’ work, and KTELM will decline. In other words, the control of these parameters is beneficial to our approach when the parameter values are reasonable.
We ran SCDMDA with KTELM on A vs. D, MNIST vs. USPS, PIE1 vs. PIE2, and amazon vs. dslr datasets with the optimal parameter settings, and computed the CDMA distance (shown in Figure 5) according to its definition as follows:
CDMA _ dist Z S , Z T = i = 1 n S + n T z i μ 2 2 n S + n T   ,   n o     l a b e l       i n     t a r g e t   d o m a i n c = 1 C z S ( c ) μ T ( c ) 2 2 + c = 1 C z T ( c ) μ S ( c ) 2 2 n S + n T   ,   o t h e r
We can see that as the iteration number increases, SCDMDA reduces the between-domain marginal and conditional distribution difference, leading to a smaller CDMA distance and higher accuracy. Therefore, SCDMDA can extract better shared features across domains.

4.5. Visualization Analysis

We further display the t-distributed stochastic neighbor embedding (t-SNE) visualization plots of feature embedding for the PIE1 vs. PIE2 classification task. t-SNE, as a dimension-reduction method, transforms high-dimensional data into a low-dimensional space for visualization. In t-SNE visualization plots, clusters of the same color are closer, clusters of different colors are more distant, and the extracted features are more discriminative. The visualized source and target features derived from raw data, JDA, STDA-CMC, and SCDMA were drawn in the feature scatterplot Figure 6a–h, in which each color represents one of the 68 categories.
As we can see from Figure 6a,b, the original data of the source and target domains before adaptation are highly mixed and difficult to distinguish. At the moment, we can easily obtain a poor classifier trained on the source domain. It can be seen from Figure 6c–f that JDA and STDA-CMC can unite samples with the same category and partly separate samples with different categories. Compared with JDA and STDA-CMC from Figure 6d,f, the small number of easily misclassified samples of SCDMDA in the boxes of Figure 6h can be especially observed, showing that SCDMDA can effectively obtain more discriminative and domain-invariant feature representation.

5. Conclusions

In this paper, we develop a semi-supervised domain adaptation approach which contains feature-based adaptation and classifier-based adaptation. At the feature-based adaptation stage, we introduce CDMA into SDA to reduce the cross-domain distribution discrepancy and develop SCDMDA for domain-invariant feature extraction. At the classifier-based adaptation stage, we select KELM as the baseline classifier. To further enhance knowledge transfer ability of the proposed approach, we design a cross-domain mean constraint term and add it to KELM to construct KTELM for domain adaptation. Through feature and classifier adaptation, we can learn more discriminatory feature representations and obtain higher accuracy. Comprehensive experiments on several visual cross-domain datasets show that SCDMDA with KTELM significantly outperforms many other state-of-the-art shallow and deep domain adaptation methods.
In the future, we plan to explore the interdomain metric criterion in more depth. At present, CDMA is superior to MMD in reducing cross-domain distribution differences; however, its computational complexity is greater than that of MMD. To solve this problem, better metrics can be developed to improve the distribution differences between domains.

Author Contributions

Methodology, software, writing—original draft preparation, X.L.; validation, writing—review and editing, funding acquisition, J.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China under Grants 62073124 and 51805149, in part by the Key Scientific Research Projects of Universities in Henan Province under Grant 22A120005, the National Aviation Fund Projects under Grant 201701420002, and Henan Province Key Scientific and Technological Projects under Grant 222102210095.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets generated during and/or analyzed during the current study are available from the corresponding author on reasonable request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wang, X.Z.; Xing, H.J.; Li, Y.; Hua, Y.L.; Dong, C.R.; Pedrycz, W. A study on relationship between generalization abilities and fuzziness of base classifiers in ensemble learning. IEEE Trans. Fuzzy Syst. 2014, 23, 1638–1654. [Google Scholar] [CrossRef]
  2. Li, S.; Huang, J.Q.; Hua, X.S.; Zhang, L. Category dictionary guided unsupervised domain adaptation for object detection. In Proceedings of the AAAI Conference on Artificial Intelligence, Palo Alto, CA, USA, 2–9 February 2021; Volume 35, pp. 1949–1957. [Google Scholar]
  3. Zhang, B.Y.; Liu, Y.Y.; Yuan, H.W.; Sun, L.J.; Ma, Z. A joint unsupervised cross-domain model via scalable discriminative extreme learning machine. Cogn. Comput. 2018, 10, 577–590. [Google Scholar] [CrossRef]
  4. Gretton, A.; Borgwardt, K.M.; Rasch, M.J.; Schölkopf, B.; Smola, A. A kernel two-sample test. J. Mach. Learn. Res. 2012, 13, 723–773. [Google Scholar]
  5. Zhuang, F.; Cheng, X.; Luo, P. Supervised representation learning with double encoding-layer autoencoder for transfer learning. ACM Trans. Intell. Syst. Technol. (TIST) 2017, 9, 1–17. [Google Scholar] [CrossRef]
  6. Shi, Q.; Zhang, Y.; Liu, X. Regularised transfer learning for hyperspectral image classification. IET Comput. Vis. 2019, 13, 188–193. [Google Scholar] [CrossRef]
  7. Lee, C.Y.; Batra, T.; Baig, M.H. Sliced wasserstein discrepancy for unsupervised domain adaptation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 10285–10295. [Google Scholar]
  8. Pan, S.J.; Kwok, J.T.; Yang, Q. Transfer learning via dimensionality reduction. In Proceedings of the Twenty-Third AAAI Conference on Artificial Intelligence, Chicago, IL, USA, 13–17 July 2008; Volume 8, pp. 677–682. [Google Scholar]
  9. Pan, S.J.; Tsang, I.W.; Kwok, J.T.; Yang, Q. Domain adaptation via transfer component analysis. IEEE Trans. Neural Netw. 2010, 22, 199–210. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  10. Gong, B.Q.; Shi, Y.; Sha, F.; Grauman, K. Geodesic flow kernel for unsupervised domain adaptation. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 2066–2073. [Google Scholar]
  11. Long, M.S.; Wang, J.M.; Ding, G.G.; Sun, J.G.; Yu, P. Transfer feature learning with joint distribution adaptation. In Proceedings of the IEEE International Conference on Computer Vision, Sydney, Australia, 3–6 December 2013; pp. 2200–2207. [Google Scholar]
  12. Lu, N.N.; Chu, F.; Qi, H.R.; Xia, S. A new domain adaption algorithm based on weights adaption from the source domain. IEEJ Trans. Electr. Electron. Eng. 2018, 13, 1769–1776. [Google Scholar] [CrossRef]
  13. Jia, S.; Deng, Y.F.; Lv, J.; Du, S.C.; Xie, Z.Y. Joint distribution adaptation with diverse feature aggregation: A new transfer learning framework for bearing diagnosis across different machines. Measurement 2022, 187, 110332. [Google Scholar] [CrossRef]
  14. Huang, G.B.; Zhu, Q.Y.; Siew, C.K. Extreme learning machine: Theory and applications. Neurocomputing 2006, 70, 489–501. [Google Scholar] [CrossRef]
  15. Huang, G.; Song, S.J.; Gupta, J.N.D.; Wu, C. Semi-supervised and unsupervised extreme learning machines. IEEE Trans. Cybern. 2014, 44, 2405–2417. [Google Scholar] [CrossRef]
  16. Sharifmoghadam, M.; Jazayeriy, H. Breast cancer classification using AdaBoost-extreme learning machine. In Proceedings of the 5th Iranian Conference on Signal Processing and Intelligent Systems (ICSPIS), IEEE, Shahrood, Iran, 18–19 December 2019; pp. 1–5. [Google Scholar]
  17. Zhang, J.; Li, Y.J.; Xiao, W.D.; Zhang, Z.Q. Non-iterative and fast deep learning: Multilayer extreme learning machines. J. Frankl. Inst. 2020, 357, 8925–8955. [Google Scholar] [CrossRef]
  18. Min, M.C.; Chen, X.F.; Xie, Y.F. Constrained voting extreme learning machine and its application. J. Syst. Eng. Electron. 2021, 32, 209–219. [Google Scholar]
  19. Li, D.Z.; Li, S.; Zhang, S.B.; Sun, J.R.; Wang, L.C.; Wang, K. Aging state prediction for supercapacitors based on heuristic kalman filter optimization extreme learning machine. Energy 2022, 250, 123773. [Google Scholar] [CrossRef]
  20. Zang, S.F.; Li, X.H.; Ma, J.W.; Yan, Y.Y.; Gao, J.W.; Wei, Y. TSTELM: Two-stage transfer extreme learning machine for unsupervised domain adaptation. Comput. Intell. Neurosci. 2022, 2022, 1582624. [Google Scholar] [CrossRef] [PubMed]
  21. Chen, Y.M.; Song, S.J.; Li, S.; Yang, L.; Wu, C. Domain space transfer extreme learning machine for domain adaptation. IEEE Trans. Cybern. 2018, 49, 1909–1922. [Google Scholar] [CrossRef]
  22. Li, X.; Zhang, W.; Ding, Q.; Sun, J.Q. Multi-layer domain adaptation method for rolling bearing fault diagnosis. Signal Process. 2019, 157, 180–197. [Google Scholar] [CrossRef] [Green Version]
  23. Cheng, M.; You, X. Adaptive matching of kernel means. In Proceedings of the 25th International Conference on Pattern Recognition (ICPR), Milan, Italy, 13–18 September 2021; pp. 2498–2505. [Google Scholar]
  24. Wang, W.; Li, H.J.; Ding, Z.M.; Nie, F.P.; Chen, J.Y.; Dong, X.; Wang, Z.H. Rethinking maximum mean discrepancy for visual domain adaptation. IEEE Trans. Neural Netw. Learn. Syst. 2021, 34, 264–277. [Google Scholar] [CrossRef]
  25. Yan, L.; Zhu, R.X.; Liu, Y.; Mo, N. TrAdaBoost based on improved particle swarm optimization for cross-domain scene classification with limited samples. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 3235–3251. [Google Scholar] [CrossRef]
  26. Li, S.; Song, S.; Huang, G. Prediction reweighting for domain adaptation. IEEE Trans. Neural Netw. Learn. Syst. 2016, 28, 1682–1695. [Google Scholar] [CrossRef] [PubMed]
  27. Li, J.Q.; Sun, T.; Lin, Q.Z.; Jiang, M.; Tan, K.C. Reducing negative transfer learning via clustering for dynamic multiobjective optimization. IEEE Trans. Evol. Comput. 2022, 26, 1102–1116. [Google Scholar] [CrossRef]
  28. Farahani, A.; Voghoei, S.; Rasheed, K.; Arabnia, H.R. A brief review of domain adaptation. In Advances in Data Science and Information Engineering, Transactions on Science and Computational Intelligence; Springer: Cham, Switzerland, 2021; pp. 877–894. [Google Scholar]
  29. Wang, Q.; Breckon, T.P. Cross-domain structure preserving projection for heterogeneous domain adaptation. Pattern Recognit. 2022, 123, 108362. [Google Scholar] [CrossRef]
  30. Wei, D.D.; Han, T.; Chu, F.L.; Zuo, M.J. Weighted domain adaptation networks for machinery fault diagnosis. Mech. Syst. Signal Process. 2021, 158, 107744. [Google Scholar] [CrossRef]
  31. Li, J.J.; Jing, M.M.; Su, H.Z.; Lu, K.; Zhu, L.; Shen, H.T. Faster domain adaptation networks. IEEE Trans. Knowl. Data Eng. 2021, 34, 5770–5783. [Google Scholar] [CrossRef]
  32. Fang, X.H.; Bai, H.L.; Guo, Z.Y.; Shen, B.; Hoi, S.; Xu, Z.L. DART: Domain-adversarial residual-transfer networks for unsupervised cross-domain image classification. Neural Netw. 2020, 127, 182–192. [Google Scholar] [CrossRef] [Green Version]
  33. Sicilia, A.; Zhao, X.C.; Hwang, S.J. Domain adversarial neural networks for domain generalization: When it works and how to improve. Mach. Learn. 2023, 1–37. [Google Scholar] [CrossRef]
  34. Zhang, W.C.; Ouyang, W.L.; Li, W.; Xu, D. Collaborative and adversarial network for unsupervised domain adaptation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 3801–3809. [Google Scholar]
  35. Yang, J.; Yan, R.; Hauptmann, A.G. Adapting SVM classifiers to data with shifted distributions. In Proceedings of the Seventh IEEE International Conference on Data Mining Workshops (ICDMW 2007), Omaha, NE, USA, 28–31 October 2007; pp. 69–76. [Google Scholar]
  36. Tommasi, T.; Orabona, F.; Caputo, B. Learning categories from few examples with multi model knowledge transfer. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 36, 928–941. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  37. Wu, Y.P.; Lv, W.J.; Li, Z.R.; Chang, J.; Li, X.C.; Liu, S. Unsupervised domain adaptation for vibration-based robotic ground classification in dynamic environments. Mech. Syst. Signal Process. 2022, 169, 108648. [Google Scholar] [CrossRef]
  38. Li, S.; Song, S.J.; Huang, G.; Wu, C. Cross-domain extreme learning machines for domain adaptation. IEEE Trans. Syst. Man Cybern. Syst. 2019, 49, 1194–1207. [Google Scholar] [CrossRef]
  39. Liu, Y.; Yang, C.; Liu, K.X.; Chen, B.C.; Yao, Y. Domain adaptation transfer learning soft sensor for product quality prediction. Chemom. Intell. Lab. Syst. 2019, 192, 103813. [Google Scholar] [CrossRef]
  40. Tang, H.; Jia, K. Discriminative adversarial domain adaptation. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 5940–5947. [Google Scholar]
  41. Li, S.; Liu, C.H.; Xie, B.H.; Su, L.M.; Ding, Z.M.; Huang, G. Joint adversarial domain adaptation. In Proceedings of the 27th ACM International Conference on Multimedia, Nice, France, 21–25 October 2019; pp. 729–737. [Google Scholar]
  42. Xu., M.H.; Zhang, J.; Ni, B.B.; Li, T.; Wang, C.J.; Tian, Q.; Zhang, W.J. Adversarial domain adaptation with domain mixup. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 6502–6509. [Google Scholar]
  43. Cai, D.; He, X.; Han, J. Speed up kernel discriminant analysis. VLDB J. 2011, 20, 21–33. [Google Scholar] [CrossRef] [Green Version]
  44. Saenko, K.; Kulis, B.; Fritz, M.; Darrell, T. Adapting visual category models to new domains. In Proceedings of the Computer Vision–ECCV 2010: 11th European Conference on Computer Vision, Heraklion, Greece, 5–11 September 2010; pp. 213–226. [Google Scholar]
  45. Gentile, C. A new approximate maximal margin classification algorithm. In Advances in Neural Information Processing Systems 13, Proceedings of the Neural Information Processing Systems (NIPS) 2000, Denver, USA, 1 January 2000; MIT Press: Cambridge, MA, USA, 2001; pp. 479–485. [Google Scholar]
  46. LeCun, Y.; Bottou, L.; Bengio, Y.; Haffber, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef]
  47. Gross, R.; Matthews, I.; Cohn, J.; Kanade, T.; Baker, S. Multi-pie. Image Vis. Comput. 2010, 28, 807–813. [Google Scholar] [CrossRef] [PubMed]
  48. Guo, G.; Wang, H.; Bell, D.; Bi, Y.; Greer, K. KNN model-based approach in classification. In Proceedings of the On the Move to Meaningful Internet Systems 2003: CoopIS, DOA, and ODBASE: OTM Confederated In-ternational Conferences, Catania, Italy, 3–7 November 2003; pp. 986–996. [Google Scholar]
  49. Zang, S.F.; Cheng, Y.H.; Wang, X.S.; Yu, Q. Semi-supervised transfer discriminant analysis based on cross-domain mean constraint. Artif. Intell. Rev. 2018, 49, 581–595. [Google Scholar] [CrossRef]
  50. Zang, S.F.; Cheng, Y.H.; Wang, X.; Yu, Q.; Xie, G.S. Cross domain mean approximation for unsupervised domain adaptation. IEEE Access 2020, 8, 139052–139069. [Google Scholar] [CrossRef]
Figure 1. Samples of the same color are of the same class in the source and target domains. An illustration of our method: (1) SCDMDA is designed by introducing CDMA into SDA for shared feature extraction across domains. (2) KTELM is constructed by adding a cross-domain mean approximation constraint into a kernel extreme learning machine for classification. (3) SCDMDA and KTELM are combined for cross-domain feature extraction and classification.
Figure 1. Samples of the same color are of the same class in the source and target domains. An illustration of our method: (1) SCDMDA is designed by introducing CDMA into SDA for shared feature extraction across domains. (2) KTELM is constructed by adding a cross-domain mean approximation constraint into a kernel extreme learning machine for classification. (3) SCDMDA and KTELM are combined for cross-domain feature extraction and classification.
Sensors 23 06102 g001
Figure 2. Office + Caltech dataset.
Figure 2. Office + Caltech dataset.
Sensors 23 06102 g002
Figure 3. Datasets PIE, USPS, and MNIST. (a) PIE; (b) USPS; (c) MNIST.
Figure 3. Datasets PIE, USPS, and MNIST. (a) PIE; (b) USPS; (c) MNIST.
Sensors 23 06102 g003
Figure 4. Effect of number of iterations, dimensions of subspace, α, β, λ 1 , θ, and λ 2 on accuracy.
Figure 4. Effect of number of iterations, dimensions of subspace, α, β, λ 1 , θ, and λ 2 on accuracy.
Sensors 23 06102 g004
Figure 5. CDMA distances versus number of iterations.
Figure 5. CDMA distances versus number of iterations.
Sensors 23 06102 g005
Figure 6. Scatter diagram of feature visualization on PIE1 vs. PIE2. (a) Source domain feature. (b) Target domain feature. (c) Source domain JDA feature. (d) Target domain JDA feature. (e) Source domain STDA-CMC feature. (f) Target domain STDA-CMC feature. (g) Source domain SCDMDA feature. (h) Target domain SCDMDA feature.
Figure 6. Scatter diagram of feature visualization on PIE1 vs. PIE2. (a) Source domain feature. (b) Target domain feature. (c) Source domain JDA feature. (d) Target domain JDA feature. (e) Source domain STDA-CMC feature. (f) Target domain STDA-CMC feature. (g) Source domain SCDMDA feature. (h) Target domain SCDMDA feature.
Sensors 23 06102 g006aSensors 23 06102 g006b
Table 1. Description of image datasets.
Table 1. Description of image datasets.
DatasetTypeNo. of SamplesFeature DimensionType NumberSubsets
OfficeObject141080010A, W, D
CaltechObject112380010C
USPSDigit180025610USPS
MNISTDigit200025610MNIST
Office-31Image4652204831amazon, webcam, dslr
PIEFace11,554102468PIE1,..., PIE5
Table 2. Accuracy (%) comparison of different algorithms on Office+Caltech, PIE, and USPS+MNIST datasets.
Table 2. Accuracy (%) comparison of different algorithms on Office+Caltech, PIE, and USPS+MNIST datasets.
Dataset/Methods1-NNKELMSDAGFKJDASTDA-CMCW-JDAJDA-CDMAWSCDMDA0SCDMDA1SCDMDA2
USPS vs. MNIST44.7046.7027.5046.4559.6563.9062.3560.3566.7576.8076.85
MNIST vs. USPS65.9468.2862.8367.2267.2879.2276.1173.0676.0683.4483.89
Average55.3257.4945.1756.8463.4771.5669.2366.7071.4080.1280.37
PIE1 vs. PIE2 (1)26.0926.4627.6926.1558.8172.4458.8777.7286.8684.5384.53
PIE1 vs. PIE3 (2)26.5927.0828.5527.2754.2373.5358.1567.7181.6882.6682.72
PIE1 vs. PIE4 (3)30.6731.0941.0031.1584.5093.9386.1993.1895.7696.0396.03
PIE1 vs. PIE5 (4)16.6717.8915.3817.5949.7563.8556.5660.1174.9476.6576.65
PIE2 vs. PIE1 (5)24.4926.8631.7825.2457.6274.9763.7875.3081.4284.8184.84
PIE2 vs. PIE3 (6)46.6346.6351.4147.3762.9369.0664.9575.3180.0280.6482.60
PIE2 vs. PIE4 (7)54.0754.4677.0554.2575.8288.3780.7183.8490.3993.6993.81
PIE2 vs. PIE5 (8)26.5326.9633.2127.0839.8954.4740.3266.5472.0069.3669.67
PIE3 vs. PIE1 (9)21.3722.0924.3721.8250.9672.3359.6672.3978.7281.0081.15
PIE3 vs. PIE2 (10)41.0140.9546.5943.1657.9567.3461.0275.6381.8382.8782.93
PIE3 vs. PIE4 (11)46.5347.7377.2046.4168.4585.4679.0685.1090.1593.6394.38
PIE3 vs. PIE5 (12)26.2326.8441.1826.7839.9560.3648.4758.0374.1477.0277.70
PIE4 vs. PIE1 (13)32.9534.3346.4934.2480.5894.2484.2491.3996.1396.9797.03
PIE4 vs. PIE2 (14)62.6862.9280.9162.9282.6391.0484.5390.4295.4096.2696.26
PIE4 vs. PIE3 (15)73.2273.6586.2773.3587.2591.1887.5089.5893.2095.1095.10
PIE4 vs. PIE5 (16)37.1938.1756.3137.3854.6675.4959.1368.9384.6284.5685.60
PIE5 vs. PIE1 (17)18.4920.2325.0920.3546.4670.0252.7661.8576.1175.9075.90
PIE5 vs. PIE2 (18)24.1924.8043.9524.6242.0557.7043.2266.3673.4279.5679.31
PIE5 vs. PIE3 (19)28.3128.9853.0028.4953.3166.8555.5171.2678.7486.5886.58
PIE5 vs. PIE4 (20)31.2433.7055.6931.3357.0178.6458.5277.0283.9087.4487.47
Average34.7635.5947.1635.3560.2475.0664.1675.3883.4785.2685.51
C vs. A (1)23.7054.4945.7241.0244.7846.0348.1244.6850.2160.7562.32
C vs. W (2)25.7648.1435.5940.6841.6942.3744.0741.6946.1055.2557.29
C vs. D (3)25.4843.9543.3138.8545.2249.6847.1345.8650.9654.7855.41
A vs. C (4)26.0045.0637.7640.2539.3641.4141.4138.5642.8348.2648.44
A vs. W (5)29.8342.0338.3138.9837.9739.3240.0038.3143.0549.1552.54
A vs. D (6)25.4844.5931.2136.3139.4938.2238.2243.3138.8545.2251.59
W vs. C (7)19.8635.7133.6630.7231.1732.2431.8832.3233.3037.3138.47
W vs. A (8)22.9638.9431.6329.7532.7833.9233.0933.6136.8541.0241.13
W vs. D (9)59.2476.4387.2680.8989.1787.9089.1789.8189.8192.9992.99
D vs. C (10)26.2734.7331.8830.2831.5232.1531.3430.4533.5736.9537.49
D vs. A (11)28.5036.0132.3632.0533.0932.6732.8832.6732.4646.0346.03
D vs. W (12)63.3966.4487.1275.5989.4990.1791.1991.1989.8391.8691.86
Average31.3747.2144.6542.9546.3147.1747.3846.8748.9954.9756.30
Total Average34.7740.9846.1539.2955.5165.0158.5364.8170.5974.2774.90
Table 3. Accuracy (%) comparison of different algorithms on Office-31 (Resnet-50).
Table 3. Accuracy (%) comparison of different algorithms on Office-31 (Resnet-50).
Datasets/Algorithms1-NNKELMJDADANDANNCANSTDA-CMCJDA-CDMAWSCDMDA0SCDMDA1SCDMDA2
amazon vs. dslr79.12 79.12 85.54 78.6079.7085.5082.73 85.5482.1390.5691.16
amazon vs. webcam75.85 75.8582.52 80.5082.0081.5079.75 82.8982.6486.4288.55
dslr vs. amazon60.17 60.2466.56 63.6068.2065.9068.09 67.1369.0873.5273.62
dslr vs. webcam95.97 95.9797.36 97.1096.9098.2098.62 97.3698.4998.9998.99
webcam vs. amazon59.92 59.9668.83 62.8067.4063.4066.31 69.0868.4473.7373.84
webcam vs. dslr99.40 99.4099.20 99.6099.1099.7099.8099.2099.8099.8099.80
Average78.41 78.4283.33 80.4082.2082.4082.55 83.5383.4387.1787.66
Table 4. Consuming time of different approaches on PIE1 vs. PIE 2.
Table 4. Consuming time of different approaches on PIE1 vs. PIE 2.
Algorithm1-NNKELMJDASTDA-CMCJDA-CDMAWSCDMDA0SCDMDA1SCDMDA2
Time(s)2.722.02187.26221.3496.69137.60134.78173.93
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, X.; Ma, J. Domain Adaptation Based on Semi-Supervised Cross-Domain Mean Discriminative Analysis and Kernel Transfer Extreme Learning Machine. Sensors 2023, 23, 6102. https://doi.org/10.3390/s23136102

AMA Style

Li X, Ma J. Domain Adaptation Based on Semi-Supervised Cross-Domain Mean Discriminative Analysis and Kernel Transfer Extreme Learning Machine. Sensors. 2023; 23(13):6102. https://doi.org/10.3390/s23136102

Chicago/Turabian Style

Li, Xinghai, and Jianwei Ma. 2023. "Domain Adaptation Based on Semi-Supervised Cross-Domain Mean Discriminative Analysis and Kernel Transfer Extreme Learning Machine" Sensors 23, no. 13: 6102. https://doi.org/10.3390/s23136102

APA Style

Li, X., & Ma, J. (2023). Domain Adaptation Based on Semi-Supervised Cross-Domain Mean Discriminative Analysis and Kernel Transfer Extreme Learning Machine. Sensors, 23(13), 6102. https://doi.org/10.3390/s23136102

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop