Next Article in Journal
Performance of Bearing Ball Defect Classification Based on the Fusion of Selected Statistical Features
Previous Article in Journal
Detecting the Critical States of Type 2 Diabetes Mellitus Based on Degree Matrix Network Entropy by Cross-Tissue Analysis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Adapting Multiple Distributions for Bridging Emotions from Different Speech Corpora

1
Key Laboratory of Child Development and Learning Science of Ministry of Education, Southeast University, Nanjing 210096, China
2
School of Biological Science and Medical Engineering, Southeast University, Nanjing 210096, China
3
School of Information Science and Engineering, Southeast University, Nanjing 210096, China
*
Author to whom correspondence should be addressed.
Entropy 2022, 24(9), 1250; https://doi.org/10.3390/e24091250
Submission received: 11 August 2022 / Revised: 1 September 2022 / Accepted: 2 September 2022 / Published: 5 September 2022
(This article belongs to the Topic Machine and Deep Learning)

Abstract

:
In this paper, we focus on a challenging, but interesting, task in speech emotion recognition (SER), i.e., cross-corpus SER. Unlike conventional SER, a feature distribution mismatch may exist between the labeled source (training) and target (testing) speech samples in cross-corpus SER because they come from different speech emotion corpora, which degrades the performance of most well-performing SER methods. To address this issue, we propose a novel transfer subspace learning method called multiple distribution-adapted regression (MDAR) to bridge the gap between speech samples from different corpora. Specifically, MDAR aims to learn a projection matrix to build the relationship between the source speech features and emotion labels. A novel regularization term called multiple distribution adaption (MDA), consisting of a marginal and two conditional distribution-adapted operations, is designed to collaboratively enable such a discriminative projection matrix to be applicable to the target speech samples, regardless of speech corpus variance. Consequently, by resorting to the learned projection matrix, we are able to predict the emotion labels of target speech samples when only the source label information is given. To evaluate the proposed MDAR method, extensive cross-corpus SER tasks based on three different speech emotion corpora, i.e., EmoDB, eNTERFACE, and CASIA, were designed. Experimental results showed that the proposed MDAR outperformed most recent state-of-the-art transfer subspace learning methods and even performed better than several well-performing deep transfer learning methods in dealing with cross-corpus SER tasks.

1. Introduction

Speech is one of the most natural behaviors through which emotional information is communicated in the daily life of human beings [1,2]. Hence, research into speech emotion recognition (SER), which seeks to enable machines to learn how to automatically understand emotional states, e.g., H a p p y , F e a r f u l , and S a d , from speech signals, has attracted attention among affective computing, pattern recognition, and speech signal processing research communities. Over recent decades, many well-performing SER methods have been proposed and have achieved promising levels of performance for widely-used speech emotion corpora [3,4,5,6,7,8]. However, the existing SER methods are far from being practically applicable. One of the major reasons is that such methods do not consider real-world scenarios, in which the training and testing speech signals may be recorded by different acoustic sensors. For example, the audio data of EmoDB [9], a widely-used speech emotion corpus, were recorded using a Sennheiser MKH40-P48 microphone and a Tascam DA-P1 portable DAT recorder. However, as for another popular speech emotion corpus, CASIA [10], its samples were recorded using a RODE K2 (a large membrane microphone) and Fireface 800 (sound card). When using these two speech emotion corpora to alternatively serve training and testing purposes, an evident feature distribution mismatch inevitably exists between their corresponding feature sets due to the acoustic sensor difference. Hence, the performance of an initially well-performing SER method will drop significantly.
The above example highlights a challenging, but interesting, task in SER, i.e., cross-corpus SER. Formally, in the task of cross-corpus SER, the training and testing speech sample sets belong to different corpora. The emotion label information of the training sample sets is provided, while the target sample sets’ labels are not entirely given. We need to enable a classifier guided by the source emotion label information to accurately predict the emotions of the unlabeled testing speech samples. Note that, in what follows, we follow the custom in the research concerning transfer learning and domain adaptation [11,12,13], which are closely related to cross-corpus SER, and refer to the training and testing speech samples/signals/corpora/feature sets as the source and target sets, respectively, such that readers can better understand this paper.
In this paper, we try to deal with cross-corpus SER tasks from the perspective of transfer learning and domain adaptation and propose a straightforward transfer subspace learning method called multiple distribution-adapted regression (MDAR). As with most existing transfer subspace learning methods [14,15,16,17,18], MDAR aims to learn a projection matrix to find a common subspace bridging the source and target speech samples from different corpora. However, we pay more attention to designing an emotion wheel knowledge-guided regularization term to help MDAR better eliminate the feature distribution difference between the source and target speech samples. Specifically, instead of directly measuring and improving both corpora’s marginal feature distribution gaps, our MDAR incorporates the idea of joint distribution adaption (JDA) [17] and joint alleviation of marginal distribution mismatch and fine emotion class-aware conditions. More importantly, unlike existing JDA-based methods [16,17,19,20], MDAR extends the JDA operation to a multiple distribution adaption (MDA) method by additionally introducing a well-designed rough emotion class-aware conditional distribution adaption to improve the feature distribution difference alleviation between the speech samples from different corpora. By resorting to MDA, MDAR can learn both corpus invariant and emotion discriminative feature representations for cross-corpus SER.
To evaluate the proposed MDAR, we carried out extensive cross-corpus SER experiments on three widely used speech emotion corpora, including EmoDB [9], eNTERFACE [21], and CASIA [10]. The experimental results showed that, compared with existing state-of-the-art transfer subspace learning, and several well-performing deep transfer learning methods, our MDAR achieved more promising performance when dealing with cross-corpus SER tasks. In summary, the main contributions of this paper are three-fold:
  • We propose a novel transfer subspace learning method called MDAR to deal with cross-corpus SER tasks. The basic idea of MDAR is very straightforward, i.e., learning corpus invariant and emotion discriminative representations for both source and target speech samples belonging to different corpora such that the classifier learning based on the labeled source speech samples is also applicable to predicting the emotions of target speech signals.
  • We present a new distribution difference alleviation regularization term called MDA for MDAR to guide the corpus invariant feature learning for the recognition of the emotions of speech signals. MDA collaboratively aligns marginal, fine emotion class-aware conditional, and rough emotion class-aware feature distributions between source and target speech samples.
  • Three widely used speech emotion corpora, i.e., EmoDB, eNTERFACE, and CASIA, were used to design the cross-corpus SER tasks to evaluate the proposed MDAR. Extensive experiments were conducted to demonstrate the effectiveness and superior performance of MDAR in coping with cross-corpus SER tasks.
The remainder of this paper is organized as follows: Section 2 reviews progress in cross-corpus SER. Section 3 provides details of the proposed MDAR method. In Section 4, extensive cross-corpus SER experiments, conducted to evaluate the proposed MDAR method, are described. Finally, we conclude this paper in Section 5.

2. Related Works

In this section, we briefly review recent advances in research concerning cross-corpus SER. To deal with cross-corpus SER tasks, considerable effort has been applied by researchers to focus on solving its key problem, i.e., relieving the feature distribution difference between the source and target speech samples belonging to different corpora. In what follows, we first describe the progress of cross-corpus SER based on transfer subspace learning methods. Moreover, we also introduce recent research into the use of deep transfer learning methods to deal with cross-corpus SER tasks.

2.1. Transfer Subspace Learning for Cross-Corpus SER

The earliest investigations into cross-corpus SER may be traced to [22], in which Schuller et al. proposed the adoption of different normalization methods, including speaker normalization (SN), corpus normalization (CN), and speaker-corpus normalization (SCN) to balance the source and target speech corpora. Then, the classifier which absorbs only the emotion discriminant information from the source speech corpus can also be applied to the target speech corpus. Subsequently, transfer subspace learning methods have been used to address the cross-corpus SER problem. For example, Hassan et al. [23] built an importance-weighted support vector machine (IW-SVM) classifier integrating three typical IW methods, i.e., kernel mean matching (KMM) [24], unconstrained least-squares importance fitting (uLSIF) [25], and the Kullback–Leibler importance estimation procedure (KLIEP) [26], to compensate for source speech samples such that the feature distribution gap between two different speech emotion corpora can be better removed. Recently, Song et al. [27] and Zhang et al. [16] designed transfer subspace learning models to learn a shared projection matrix to jointly build the relationship between the emotion labels and transformed speech features, and to align the source and target speech samples’ feature distributions.

2.2. Deep Transfer Learning for Cross-Corpus SER

Apart from the above subspace learning methods, inspired by the success of deep transfer learning and deep domain adaptation in many cross-domain visual recognition tasks, researchers have also designed domain invariant deep neural networks to deal with the cross-corpus SER problem. For example, Deng et al. [28,29] proposed a series of unsupervised deep domain adaptation methods using autoencoder (AE) networks instead of projection matrices to seek a common subspace for both source and target speech signals such that their new representations in the common subspace are similarly distributed. Gideon et al. [30] were motivated by the idea of generative adversarial networks (GANs) [31] and presented an adversarial discriminative domain generalization (ADDoG) model to cope with cross-corpus SER tasks. ADDoG consists of three major modules, i.e., a feature encoder, an emotion classifier, and a critic. Among these, the critic is one of the major modules aiming to remove the bias between the source and target speech corpora by estimating their earth-mover or Wasserstein distance. In addition, it is also of note that, unlike most existing methods, ADDoG made use of speech spectrums rather than hand-crafted speech features to serve as the inputs of networks. Hence, it is an end-to-end learning method.

3. Proposed Method

3.1. Notations

In this section, we address the proposed MDAR method in detail and describe how to use MDAR to deal with cross-corpus SER tasks. To begin with, we give several notations which are needed in formulating MDAR. Suppose we have a set of labeled source speech samples from one corpus whose feature matrix is denoted by X s R d × n s , where d is the dimension of the speech feature vectors and n s is the source speech sample number. Their corresponding emotion ground truth information is denoted by a label matrix Y s R c × n s , where c is the emotion class number and its ith column y i = [ y 1 i , , y c i ] T describes its corresponding speech sample’s emotion information. As for y i , only the jth entry is set as 1 while the others are set as 0 if this speech sample’s label is the jth emotion.
Simliarly, let the target speech feature matrix corresponding to the other corpus and its corresponding unknown label matrix be X t R d × n t and Y t R c × n t , where n t is the target sample number. According to the emotion class, we divide the source and target speech feature matrices X s and X t into { X s f ( 1 ) , , X s f ( c ) } and { X t f ( 1 ) , , X t f ( c ) } , where X s f ( i ) and X t f ( i ) denote the source and target speech feature matrices corresponding to the ith emotion among the fine emotion class set { 1 , , c } . Accordingly, several fine emotion class feature matrix sets can further merge to obtain the rough emotion class feature matrix set for source and target speech samples, which can be expressed as { X s r ( 1 ) , , X s r ( c r ) } and { X t r ( 1 ) , , X t r ( c r ) } , where X s r ( i ) and X t r ( i ) represent the feature matrices corresponding to the ith rough emotion class and c r is the rough emotion class number.

3.2. Formulation of MDAR

As described previously, the basic idea of MDAR is to build a subspace learning model to learn emotion discriminative and corpus invariant representations for both source and target speech samples belonging to different corpora. To achieve this goal, we propose to use the label-information-guided feature space to serve as the subspace and then learn a projection matrix to build the relationship between this subspace and the original feature space, which can be formulated as a simple linear regression optimization problem:
min U Y s U T X s F 2 ,
where U is such a satisfactory projection matrix and · F denotes the Frobenius norm of a matrix. Using U , we can easily transform the speech samples from the original feature space to the emotion label space. In other words, this learned projection matrix is endowed with emotion discriminative ability.
Subsequently, we need to further enable the projection matrix U to be robust to the variance of speech corpora such that it is applicable to the problem of cross-corpus SER. To this end, we design a regularization term to help MDAR learn such an expectative projection matrix, whose corresponding optimization problem can be expressed as follows:
min U U T 2 , 1 + λ 1 ( 1 n s U T X s 1 s 1 n t U T X t 1 t 2 + i = 1 c 1 n s f ( i ) U T X s f ( i ) 1 s f ( i ) 1 n t f ( i ) U T X t f ( i ) 1 t f ( i ) 2 + i = 1 c r 1 n s r ( i ) U T X s r ( i ) 1 s r ( i ) 1 n t r ( i ) U T X t r ( i ) 1 t r ( i ) 2 ) ,
where λ 1 is a trade-off parameter controlling the balance between different terms, and 1 s , 1 t , 1 s f ( i ) , 1 t f ( i ) , 1 s r ( i ) , and 1 t r ( i ) are all the one-valued vectors, and their dimensions are the numbers of source and target samples denoted by n s and n t , target and target samples corresponding to ith fine emotion class denoted by n s f ( i ) and n t f ( i ) , and source and target samples corresponding to ith rough emotion class denoted by n s r ( i ) and n t r ( i ) , respectively.
From Equation (2), it is clear that the objective function designed for the corpus robustness of the projection matrix consists of a l 2 , 1 norm and a combination of marginal, fine emotion class-aware conditional, and rough emotion class-aware conditional distributions aligned functions with respect to U , respectively. These two terms correspond to two aspects of our efforts regarding MDAR:
  • U T 2 , 1 can be called the feature selection term. Minimizing U T 2 , 1 helps the MDAR learn a row-sparse projection matrix, which suppresses the speech features contributing less to the distinction of different emotions, while highlighting the features contributing most to distinction.
  • The other aspect is the multiple distribution adaption (MDA), which corresponds to the resting three terms. Among these, the first two terms are so-called joint distribution adaptions (JDA) [16,17,19,20]. JDA is a combination of the marginal distribution adaption and the fine emotion class-aware conditional adaption and has been demonstrated the effectiveness in coping with domain adaptation and other cross-domain recognition tasks. Our MDA can be viewed as an extension of JDA incorporating an additional rough emotion class-aware conditional distribution-adapted term, which enables further enhancement of the corpus invariant ability of the proposed MDAR.
Finally, by combining Equations (1) and (2), we arrive at the eventual optimization problem of the proposed MDAR method, which can be formulated as follows:
min U Y s U T X s F 2 + λ U T 2 , 1 + μ ( 1 n s U T X s 1 s 1 n t U T X t 1 t 2 + i = 1 c 1 n s r ( i ) U T X s f ( i ) 1 s f ( i ) 1 n t f ( i ) U T X t ( i ) 1 t f ( i ) 2 + i = 1 c r 1 n s r ( i ) U T X s r ( i ) 1 s r ( i ) 1 n t r ( i ) U T X t r ( i ) 1 t r ( i ) 2 ) ,
where λ and μ = λ × λ 1 are the trade-off parameters to balance all the terms.

3.3. Disturbance Strategy for Constructing Rough Emotion Groups in MDA

The major inspiration for designing the rough emotion class-aware conditional distribution adapted term to obtain MDA was the recent work of [32], in which a modified 2D arousal-valence emotion wheel consisting of two dimensions, i.e., valence and arousal, is presented. To better understand our motivation, we repost Yang et al.’s emotion wheel in Figure 1. From Figure 1, it is clear that each typical discrete emotion, e.g., A n g r y , H a p p y , and S u r p r i s e , can be mapped to one point in the emotion wheel based on its corresponding valence and arousal degrees. As the emotion wheel shows, there is an intrinsic distance between two emotions according to their positions on the emotion wheel. Several typical emotions, e.g., F e a r vs. D i s g u s t , and S u r p r i s e vs. H a p p y , are very similar and difficult to distinguish from their distance measured with respect to the valence and arousal. In other words, it may be hard to directly align the fine class-aware conditional distribution associated with these emotions due to the unavailability of target speech sample emotion labels. Although we can predict their pseudo emotion labels to calculate statistics for the fine class-aware conditional distribution, the emotion discriminative ability of MDAR is limited in the initial iterations of optimization.
To relieve this tension, in this paper, we introduce the rough emotion class-aware conditional distribution-adapted term and present a disturbance strategy to construct its rough emotion class groups. Specifically, along the valence dimension, we first divide the emotions into two rough emotion class groups including P o s i t i v e - V a l e n c e ( S u r p r i s e , H a p p y , and N e u t r a l ) and N e g a t i v e - V a l e n c e ( A n g r y , D i s g u s t , F e a r , and S a d ). Then, regarding the specific cross-corpus SER task, we make several modifications to the original rough emotion groups to break the inseparability of some emotions which have a close distance with respect to the degree of valence and arousal. For example, we can switch A n g r y and S u r p r i s e for H i g h - V a l e n c e and L o w - V a l e n c e groups. Finally, following the modified mixed emotion groups, we calculate the rough emotion class-aware conditional distribution-adapted term i = 1 c r 1 n s U T X s r ( i ) 1 s r ( i ) 1 n t U T X t r ( i ) 1 t r ( i ) 2 .
Note that, introducing the above rough emotion class-aware conditional distribution-adapted term under the disturbance strategy for MDAR has two expectative benefits. First, the modification of the mixed emotion groups alleviates the inseparability of the emotion elements in P o s i t i v e - V a l e n c e or N e g a t i v e - V a l e n c e groups and, hence, assists fine emotion class-aware conditional distribution adaption in MDAR. Second, unlike the fine emotion class-aware conditional distribution adaption, performing a rough adaption does not require over-precise target pseudo-labels, which affects the fine emotion class-aware conditional distribution adaption. However, the proposed rough adaption does not have this drawback because it only needs rough emotion labels of target speech samples, the prediction of which is an easier task.

3.4. Predicting the Target Emotion Label Using MDAR

Once the optimal projection matrix of MDAR denoted by U ^ is learned, we are able to predict the emotion label of the target speech samples according to the following criterion:
emo _ label = arg max i y t t e ( i ) .
Note that y t t e denotes the target emotion label vector and can be computed by y t t e = U ^ T x t t e , where x t t e is its corresponding feature vector and y t t e ( i ) is its ith entry.

3.5. Optimization of MDAR

The optimization of MDAR can be solved by the alternated direction method (ADM) and inexact augmented Lagrangian multiplier (IALM) [33]. Specifically, we first initialize the projection matrix U and then repeat the following two major steps until convergence:
  • Predict the target emotion labels based on the projection matrix U and Equation (4). Then compute the original marginal and two aware conditional feature distribution gaps denoted by Δ m , Δ f ( i ) , and Δ r ( i ) according to the predicted target emotion labels using the following Equations (5)–(7):
    Δ m = 1 n s X s 1 s 1 n t X t 1 t .
    Δ f ( i ) = 1 n s f ( i ) X s f ( i ) 1 t f ( i ) 1 n t f ( i ) X t f ( i ) 1 t f ( i ) ,
    where i = { 1 , , c } .
    Δ r ( i ) = 1 n s r ( i ) X s r ( i ) 1 s r ( i ) 1 n t r ( i ) X t r ( i ) 1 t r ( i ) ,
    where i = { 1 , , c r } .
  • Solve the following optimization problem:
    min U [ Y s , 0 ] U T [ X s , μ Δ ] F 2 + λ U T 2 , 1 ,
    where 0 R c × ( c + c r + 1 ) is a zero matrix, and Δ = [ Δ m , Δ f ( 1 ) , , Δ f ( c ) , Δ r ( 1 ) , , Δ f ( c r ) ] R d × ( c + c r + 1 ) .
As for Equation (8), IALM can be used to efficiently optimize it. More specifically, we introduce an auxiliary variable P satisfying P = U . Thus, we can convert the original optimization problem to a constrained problem as follows:
min U , P L P T Z F 2 + λ U T 2 , 1 , s . t . P = U ,
where L = [ Y s , 0 ] and Z = [ X s , μ Δ ] .
Subsequently, we can write its corresponding Lagrangian function as follows:
L ( U , P , T , κ ) = L P T Z F 2 + λ U T 2 , 1 + T r [ T T ( P U ) ] + κ 2 P U F 2 ,
where T r ( · ) denotes the trace of a square matrix, T is the multiplier matrix and κ is the trade-off parameter. By alternatively minimizing the Lagrangian function with respect to the variables, we can obtain the optimal U . We summarize the detailed updating rules in Algorithm 1.
Algorithm 1 Complete updating rule for learning the optimal U in Equation (10).
Repeat the following steps until convergence:
  • Fix U , T , and κ , update P : min P L P T Z F 2 + T r ( T T P ) + κ 2 P U F 2 , which results in P = ( 2 Z Z T + κ I ) 1 ( κ U T Z L T ) .
  • Fix P , T , and κ , update U : min U λ κ U T 2 , 1 + 1 2 U T ( P T + T T κ ) F 2 , whose solution is obtained by
    c i = p i + t i κ λ κ p i + t i κ ( p i + t i κ ) , if λ κ < p i + t i κ , where p i and t i are the ith row of P and T , respectively. Otherwise, c i = 0 .
  • Update T and κ : T = T + κ ( P U ) , and κ = min { ρ κ , κ m a x } .
  • Check convergence: P U F < ϵ .

4. Experiments

4.1. Speech Emotion Corpora and Experimental Protocol

In this section, we describe cross-corpus SER experiments to evaluate the proposed MDAR method. In what follows, we give the detail of the evaluation experiments.
  • Speech Emotion Corpora: Three widely-used speech emotion corpora, i.e., EmoDB (Berlin) [9], eNTERFACE [21], and CASIA [10], were adopted to design cross-corpus SER tasks. EmoDB is a German corpus and was collected by Burkhardt et al. from TU Berlin, Germany. It consists of 535 acted speech samples from 10 speakers, including five females and five males. Each speech sample is assigned one of seven basic emotion labels, i.e., N e u t r a l ( N E ) , A n g r y ( A N ) , F e a r ( F E ) , H a p p y ( H A ) , S a d ( S A ) , D i s g u s t ( D I ) , and B o r e d o m . eNTERFACE is an English audio-visual emotion database consisting of 42 speakers from 14 different nationalities. The emotions involved are A N , D I , F E , H A , S A , and S u r p r i s e ( S U ) . In the experiments, we only adopted its audio subset. CASIA is a Chinese speech emotion corpus designed by the Institute of Automation, Chinese Academy of Science, China. It includes 1200 speech samples covering six basic emotions, i.e., A N , S A , F E , H A , N E , and S U .
  • Task Detail: We used two of the above speech emotion corpora to serve as the source and target corpora, alternatively, and thus derived six typical cross-corpus SER tasks, i.e., B E , E B , B C , C B , E C , and C E , where B, E, and C are short for EmoDB, eNTERFACE, and CASIA, and the left and right corpora of the arrow correspond to the source and target corpora, respectively. It is of note that, since these corpora have different emotions, in each cross-corpus SER task, we extracted speech samples sharing the same emotion labels to ensure label consistency. The detailed sample statistical information of the selected speech emotion corpora is given in Table 1.
  • Performance Metric: As for the performance metric, the unweighted average recall (UAR) [22], defined as the accuracy per class averaged by the total emotion class number, was chosen.

4.2. Comparison Methods and Implementation Detail

For comparison, we included recent well-performing transfer subspace learning methods, i.e., transfer component analysis (TCA) [14], geodesic flow kernel (GFK) [15], subspace alignment (SA) [34], domain-adaptive subspace learning (DoSL) [35], and joint distribution adaptive regression (JDAR) [16]. Linear SVM was used as the classifier and we report its results for all the cross-corpus SER tasks to serve as the baseline. Since subspace learning methods are not end-to-end methods, they need a hand-crafted speech feature set to describe speech signals. In the experiments, we adopted IS09 [36] and IS10 [37] feature sets provided by the INTERSPEECH 2009 Emotion Challenge and the INTERSPEECH 2010 Paralinguistic Challenge, respectively, for all the subspace learning methods. The IS09 feature set consists of 384 elements produced by 32 low-level descriptors (LLDs), e.g., fundamental frequency (F0), Mel-frequency cepstrum coefficient (MFCC), their first-order difference, and their 12 corresponding functions, e.g., mean, maximal, and minimal value. Compared with IS09, the IS10 feature set contains more LLDs and functions such that its element number increases to 1582. Both feature sets can be conveniently extracted using the openSMILE toolkit [38]; detailed information is available in [36,37].
Furthermore, we also compared our MDAR method with several recent state-of-the-art deep transfer learning methods including the deep adaptation network (DAN) [39], the domain-adversarial neutral network (DANN) [40], deep-CORAL [41], the deep subdomain adaptation network (DSAN) [42], and the deep transductive transfer regression network (DTTRN) [20]. For these deep learning methods, AlexNet was chosen as the CNN backbone and we also used AlexNet to conduct the experiments to serve as the baseline. The speech spectrums served as the network inputs instead of the hand-crafted speech feature sets. Specifically, the frame size and overlap were first set as 350 and 175 sampling points, respectively. Then, for each speech signal, all the frames were windowed using the Hamming function and subsequently transformed to individual spectrums by resorting to Fourier transformation. Finally, these individual spectrums composed the spectrum of the speech signal. Note that due to the unavailability of target label information in cross-corpus SER, a cross-validation method cannot be used to determine the optimal hyper-parameters for all the methods. Hence, following most existing studies [16,35,39,42], in our experiments, we searched the hyper-parameters for all the methods from a preset interval and then reported their best UAR corresponding to the best optimal hyper-parameter. The details of the hyper-parameter setting for all the transfer learning methods were as follows:
  • TCA, GFK, and SA: For these three methods, the hyper-parameter, i.e., the reduced dimension, needed to be set. In the experiments, we searched it from [ 5 : 5 : d m a x ] , where d m a x is the maximal dimension.
  • DoSL and JDAR: DoSL and JDAR have two trade-off parameters controlling the balance between the original loss function and two regularization terms, in which one corresponds to the sparsity and the other corresponds to feature distribution adaption. We searched them both from [ 5 : 5 : 200 ] in the experiments.
  • DAN and DSAN: DAN and DSAN both have a trade-off parameter to balance the original loss and the MMD regularization term. In the experiments, we set it by searching from { 0.001 , 0.005 , 0.01 , 0.05 , 0.1 , 0.5 } .
  • DANN: As for DANN, it also has only one trade-off parameter. We searched it from the parameter set { 0.001 , 0.003 , 0.005 , 0.01 , 0.05 , 0.1 , 0.5 } throughout the experiments.
  • DTTRN: Since the protocol in [20] was identical to ours, we used the results reported in their experiments for comparison.
  • MDAR: Similar to DoSL and JDAR, our MDAR also had two hyper-parameters, i.e., λ and μ . They were used to control the balance between the original regression loss function and the two regularization terms, including the feature selection and feature distribution difference alleviation terms. In the experiments, they were also both searched from the parameter interval [ 5 : 5 : 200 ] . In addition, the rough emotion class number c r was set to 2 ( H i g h - V a l e n c e and L o w - V a l e n c e ). The disturbance strategy for the two mixed rough emotion groups was performed as follows: Reassign D i s g u s t from the L o w - V a l e n c e group to the H i g h - V a l e n c e group for B E and E B , and F e a r from the L o w - V a l e n c e group to the H i g h - V a l e n c e group for B C and C B . Switch A n g r y and S u r p r i s e for E C and C E .

4.3. Results and Discussion

4.3.1. Comparison with Transfer Subspace Learning Methods

The experimental results are shown in Table 2, Table 3 and Table 4. Among these, Table 2 and Table 3 correspond to the comparison among the transfer subspace learning methods using IS09 and IS10 as the feature sets, respectively. From Table 2 and Table 3, several interesting observations can be made:
First, it is clear that the proposed MDAR method achieved the best UAR averaged by the results of all the six cross-corpus SER tasks among all the transfer subspace learning methods when using both IS09 and IS10 feature sets to describe the speech signals. Specifically, the average UAR achieved by our MDAR reached 42.26% and 36.69% in the experiments using IS09 and IS10 as the feature sets, respectively, with promising increases of 0.50% and 0.42% over the second best results (41.76% obtained by JDAR [16] + IS10 and 36.33 % obtained by SA [34] + IS09). This indicates that our MDAR demonstrated superior overall performance compared to recent state-of-the-art transfer subspace learning methods when dealing with cross-corpus SER tasks.
Second, it was also evident that, using IS10 as the speech feature set, our MDAR achieved more promising results in terms of UAR than all the comparison methods for the four cross-corpus SER tasks ( B E , E B , B C , and E C ) among all the six tasks. Although in the resting tasks our MDAR did not beat the other transfer subspace learning methods, the performance of MDAR was very competitive against the best-performing transfer subspace learning methods, e.g., 37.30 % (MDAR) vs. 37.58 % (JDAR) in the task of B E .
Last, but not least, from the comparison between Table 2 and Table 3, it is clear that the performance of all the transfer subspace learning methods varied with respect to the feature set used to describe speech signals. Specifically, the IS10 feature set included more low-level acoustic descriptors and statistical functions than IS09, which provided more emotion discriminative information when recognizing the emotions of speech signals. Hence, the performance of all the transfer subspace learning methods with the IS10 feature set increased remarkably compared to IS09. This remarkable performance increase indicates that, when dealing with cross-corpus SER tasks, the capacity of the hand-crafted speech feature set chosen to describe the speech signals is very important for the transfer of subspace learning methods.

4.3.2. Comparison with Deep Transfer Learning Methods

Table 4 shows the comparison between our MDAR and several recent state-ot-the-art deep transfer learning methods. From Table 4, it can be seen that, in terms of the average UAR, all the deep transfer learning methods outperformed our MDAR using IS09 as the feature set to describe the speech signals. However, when using the IS10 feature set, the performance of our MDAR increased from 36.69% to 42.26% in terms of the average UAR, beating the deep transfer learning methods. More importantly, our MDAR, together with the IS10 feature set, showed superior performance compared with the comparison deep learning methods in five of six cross-corpus SER tasks. These observations further confirmed the effectiveness and satisfactory performance of the proposed MDAR in coping with cross-corpus SER tasks, which would otherwise lose to the deep transfer learning methods if the hand-crafted speech feature set has adequate ability to describe the speech signals adopted.

4.3.3. Going Deeper into Disturbance Strategy in MDAR

As Equation (2) shows, our MDAR absorbs the knowledge of the emotion wheel to design a rough emotion class-aware conditional distribution-adapted term to help corpus-invariant feature learning. In this distribution-adapted term, two rough emotion groups are obtained in advance, according to the valence dimension under the guidance of the disturbance strategy, i.e., switching several emotion elements in both groups. Therefore, it is interesting to consider whether the proposed strategy (denoted by the Proposed Modification) is effective for improving MDAR in coping with cross-corpus SER tasks. To this end, we conducted additional experiments choosing tasks using the IS09 feature set as the representatives, and then adopted the original valence-based rough emotion groups to compute this well-designed term (denoted by the Original Version) for MDAR. Table 5 presents the experimental results. From Table 5, it can be seen that MDAR achieved better performance when using the proposed disturbance strategy to modify the rough emotion groups and to compute its corresponding conditional distribution-adapted term compared with using the original method.

4.3.4. Sensitivity Analysis of Trade-Off Parameters in MDAR

From the optimization problem of MDAR shown in Equation (3), it is known that our MDAR has two major trade-off parameters, including λ and μ , controlling the balance between the original regression loss and the distribution-adapted regularization terms. This generates an interesting problem, i.e., how the performance of the proposed MDAR changes with respect to these two parameters. To investigate this, we conducted additional experiments choosing the tasks B E and C B , using the IS09 feature set as the representatives. Specifically, we alternatively fixed one parameter at the optimal value and varied the other for a parameter interval centered at its optimal value, and then performed MDAR for each task. The experimental results are shown in Figure 2, in which the fixed parameter and the varying parameter interval are also provided. From Figure 2, it is clear that the performance of the proposed MDAR varied slightly with respect to the change in both trade-off parameters, which indicates that our MDAR was less sensitive to the choice of its trade-off parameters.

5. Conclusions

In this paper, we investigated the problem of cross-corpus SER and proposed a novel effective transfer subspace learning method called MDAR. Unlike most existing transfer subspace learning methods, the proposed MDAR absorbs the emotion wheel knowledge and adopts a well-designed distribution-adapted regularization term which considers the marginal distribution adaption and two-scale emotion-aware conditional adaption to jointly alleviate the feature distribution mismatch between the source and target speech corpora. Extensive cross-corpus SER experiments were carried out to evaluate the performance of the proposed MDAR method. The experimental results demonstrated the effectiveness of MDAR and its superior performance over recent state-of-the-art transfer subspace learning methods, including several high-performing deep transfer learning methods, in coping with cross-corpus SER tasks.

Author Contributions

Conceptualization, methodology, writing—original draft preparation, and funding acquisition, Y.Z.; software and validation, Y.Z. and H.L.; formal analysis, Y.Z. and H.C.; investigation, H.L., H.C. and C.L.; resources, and data curation, H.L.; writing—review and editing, Y.Z., H.L. and C.T. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported, in part, by the National Natural Science Foundation of China (NSFC) under Grants U2003207 and 61902064, in part, by the Jiangsu Frontier Technology Basic Research Project under Grant BK20192004, and, in part, by the Zhishan Young Scholarship of Southeast University.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Publicly available datasets were analyzed in this study. This data can be found here: [EmoDB] [http://emodb.bilderbar.info/start.html] (Accessed on 22 May 2022), [eNTERFACE] [http://www.enterface.net/enterface05] (Accessed on 22 May 2022), and [CASIA] [http://www.chineseldc.org] (Accessed on 22 May 2022).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. El Ayadi, M.; Kamel, M.S.; Karray, F. Survey on speech emotion recognition: Features, classification schemes, and databases. Pattern Recognit. 2011, 44, 572–587. [Google Scholar] [CrossRef]
  2. Schuller, B.W. Speech emotion recognition: Two decades in a nutshell, benchmarks, and ongoing trends. Commun. ACM 2018, 61, 90–99. [Google Scholar]
  3. Zong, Y.; Zheng, W.; Cui, Z.; Li, Q. Double sparse learning model for speech emotion recognition. Electron. Lett. 2016, 52, 1410–1412. [Google Scholar]
  4. Chen, M.; He, X.; Yang, J.; Zhang, H. 3-D Convolutional Recurrent Neural Networks With Attention Model for Speech Emotion Recognition. IEEE Signal Process. Lett. 2018, 25, 1440–1444. [Google Scholar] [CrossRef]
  5. Zhang, S.; Tao, X.; Chuang, Y.; Zhao, X. Learning deep multimodal affective features for spontaneous speech emotion recognition. Speech Commun. 2021, 127, 73–81. [Google Scholar]
  6. Li, D.; Zhou, Y.; Wang, Z.; Gao, D. Exploiting the potentialities of features for speech emotion recognition. Inf. Sci. 2021, 548, 328–343. [Google Scholar]
  7. Zhang, S.; Zhao, X.; Tian, Q. Spontaneous speech emotion recognition using multiscale deep convolutional LSTM. IEEE Trans. Affect. Comput. 2022, 13, 680–688. [Google Scholar]
  8. Lu, C.; Zong, Y.; Zheng, W.; Li, Y.; Tang, C.; Schuller, B.W. Domain Invariant Feature Learning for Speaker-Independent Speech Emotion Recognition. IEEE/ACM Trans. Audio Speech Lang. Process. 2022, 30, 2217–2230. [Google Scholar]
  9. Burkhardt, F.; Paeschke, A.; Rolfes, M.; Sendlmeier, W.F.; Weiss, B. A database of German emotional speech. In Proceedings of the INTERSPEECH 2005, Lisbon, Portugal, 4–8 September 2005; pp. 1517–1520. [Google Scholar]
  10. Tao, J.; Liu, F.; Zhang, M.; Jia, H. Design of speech corpus for mandarin text to speech. In Proceedings of the Blizzard Challenge 2008 Workshop, Brisbane, Australia, 22–26 September 2008; pp. 1–4. [Google Scholar]
  11. Pan, S.J.; Yang, Q. A survey on transfer learning. IEEE Trans. Knowl. Data Eng. 2009, 22, 1345–1359. [Google Scholar]
  12. Wang, M.; Deng, W. Deep visual domain adaptation: A survey. Neurocomputing 2018, 312, 135–153. [Google Scholar]
  13. Zhuang, F.; Qi, Z.; Duan, K.; Xi, D.; Zhu, Y.; Zhu, H.; Xiong, H.; He, Q. A comprehensive survey on transfer learning. Proc. IEEE 2020, 109, 43–76. [Google Scholar]
  14. Pan, S.J.; Tsang, I.W.; Kwok, J.T.; Yang, Q. Domain Adaptation via Transfer Component Analysis. IEEE Trans. Neural Netw. 2011, 22, 199–210. [Google Scholar] [CrossRef] [PubMed]
  15. Gong, B.; Shi, Y.; Sha, F.; Grauman, K. Geodesic flow kernel for unsupervised domain adaptation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 2066–2073. [Google Scholar]
  16. Zhang, J.; Jiang, L.; Zong, Y.; Zheng, W.; Zhao, L. Cross-Corpus Speech Emotion Recognition Using Joint Distribution Adaptive Regression. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Toronto, ON, Canada, 6–11 June 2021; pp. 3790–3794. [Google Scholar] [CrossRef]
  17. Long, M.; Wang, J.; Ding, G.; Sun, J.; Yu, P.S. Transfer feature learning with joint distribution adaptation. In Proceedings of the IEEE International Conference on Computer Vision, Sydney, Australia, 1–8 December 2013; pp. 2200–2207. [Google Scholar]
  18. Zhang, T.; Zong, Y.; Zheng, W.; Chen, C.P.; Hong, X.; Tang, C.; Cui, Z.; Zhao, G. Cross-database micro-expression recognition: A benchmark. IEEE Trans. Knowl. Data Eng. 2022, 34, 544–559. [Google Scholar]
  19. Tan, Y.; Guo, L.; Gao, H.; Zhang, L. Deep coupled joint distribution adaptation network: A method for intelligent fault diagnosis between artificial and real damages. IEEE Trans. Instrum. Meas. 2020, 70, 1–12. [Google Scholar]
  20. Zhao, Y.; Wang, J.; Ye, R.; Zong, Y.; Zheng, W.; Zhao, L. Deep Transductive Transfer Regression Network for Cross-Corpus Speech Emotion Recognition. In Proceedings of the INTERSPEECH, Incheon, Korea, 18–22 September 2022. [Google Scholar]
  21. Martin, O.; Kotsia, I.; Macq, B.; Pitas, I. The eNTERFACE’05 Audio-Visual Emotion Database. In Proceedings of the 22nd International Conference on Data Engineering Workshops, Washington, DC, USA, 3–7 April 2006; p. 8. [Google Scholar]
  22. Schuller, B.; Vlasenko, B.; Eyben, F.; Wöllmer, M.; Stuhlsatz, A.; Wendemuth, A.; Rigoll, G. Cross-Corpus Acoustic Emotion Recognition: Variances and Strategies. IEEE Trans. Affect. Comput. 2010, 1, 119–131. [Google Scholar] [CrossRef]
  23. Hassan, A.; Damper, R.; Niranjan, M. On Acoustic Emotion Recognition: Compensating for Covariate Shift. IEEE Trans. Audio Speech Lang. Process. 2013, 21, 1458–1468. [Google Scholar] [CrossRef]
  24. Gretton, A.; Smola, A.; Huang, J.; Schmittfull, M.; Borgwardt, K.; Schölkopf, B. Covariate shift by kernel mean matching. Dataset Shift Mach. Learn. 2009, 3, 5. [Google Scholar]
  25. Kanamori, T.; Hido, S.; Sugiyama, M. A least-squares approach to direct importance estimation. J. Mach. Learn. Res. 2009, 10, 1391–1445. [Google Scholar]
  26. Tsuboi, Y.; Kashima, H.; Hido, S.; Bickel, S.; Sugiyama, M. Direct density ratio estimation for large-scale covariate shift adaptation. J. Inf. Process. 2009, 17, 138–155. [Google Scholar]
  27. Song, P. Transfer Linear Subspace Learning for Cross-Corpus Speech Emotion Recognition. IEEE Trans. Affect. Comput. 2019, 10, 265–275. [Google Scholar] [CrossRef]
  28. Deng, J.; Zhang, Z.; Eyben, F.; Schuller, B. Autoencoder-based Unsupervised Domain Adaptation for Speech Emotion Recognition. IEEE Signal Process. Lett. 2014, 21, 1068–1072. [Google Scholar] [CrossRef]
  29. Deng, J.; Xu, X.; Zhang, Z.; Frühholz, S.; Schuller, B. Universum Autoencoder-Based Domain Adaptation for Speech Emotion Recognition. IEEE Signal Process. Lett. 2017, 24, 500–504. [Google Scholar] [CrossRef]
  30. Gideon, J.; McInnis, M.G.; Provost, E.M. Improving Cross-Corpus Speech Emotion Recognition with Adversarial Discriminative Domain Generalization (ADDoG). IEEE Trans. Affect. Comput. 2021, 12, 1055–1068. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  31. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. Adv. Neural Inf. Process. Syst. 2014, 27, 1–11. [Google Scholar]
  32. Yang, L.; Shen, Y.; Mao, Y.; Cai, L. Hybrid Curriculum Learning for Emotion Recognition in Conversation. In Proceedings of the AAAI, Online. 22 February–1 March 2022. [Google Scholar]
  33. Lin, Z.; Liu, R.; Su, Z. Linearized alternating direction method with adaptive penalty for low-rank representation. Adv. Neural Inf. Process. Syst. 2011, 24, 1–9. [Google Scholar]
  34. Fernando, B.; Habrard, A.; Sebban, M.; Tuytelaars, T. Unsupervised visual domain adaptation using subspace alignment. In Proceedings of the IEEE International Conference on Computer Vision, Sydney, Australia, 1–8 December 2013; pp. 2960–2967. [Google Scholar]
  35. Liu, N.; Zong, Y.; Zhang, B.; Liu, L.; Chen, J.; Zhao, G.; Zhu, J. Unsupervised Cross-Corpus Speech Emotion Recognition Using Domain-Adaptive Subspace Learning. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; pp. 5144–5148. [Google Scholar] [CrossRef] [Green Version]
  36. Schuller, B.; Steidl, S.; Batliner, A. The interspeech 2009 emotion challenge. In Proceedings of the INTERSPEECH, Brighton, UK, 6–10 September 2009. [Google Scholar]
  37. Schuller, B.; Steidl, S.; Batliner, A.; Burkhardt, F.; Devillers, L.; Müller, C.; Narayanan, S.S. The INTERSPEECH 2010 paralinguistic challenge. In Proceedings of the INTERSPEECH, Makuhari, Japan, 26–30 September 2010. [Google Scholar]
  38. Eyben, F.; Wöllmer, M.; Schuller, B. Opensmile: The munich versatile and fast open-source audio feature extractor. In Proceedings of the 18th ACM International Conference on Multimedia, Firenze, Italy, 25–29 October 2010; pp. 1459–1462. [Google Scholar]
  39. Long, M.; Cao, Y.; Wang, J.; Jordan, M. Learning transferable features with deep adaptation networks. In Proceedings of the International Conference on Machine Learning, PMLR, Lille, France, 7–9 July 2015; pp. 97–105. [Google Scholar]
  40. Ajakan, H.; Germain, P.; Larochelle, H.; Laviolette, F.; Marchand, M. Domain-adversarial neural networks. arXiv 2014, arXiv:1412.4446. [Google Scholar]
  41. Sun, B.; Saenko, K. Deep coral: Correlation alignment for deep domain adaptation. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2016; pp. 443–450. [Google Scholar]
  42. Zhu, Y.; Zhuang, F.; Wang, J.; Ke, G.; Chen, J.; Bian, J.; Xiong, H.; He, Q. Deep subdomain adaptation network for image classification. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 1713–1722. [Google Scholar]
Figure 1. The 2D arousal-valence emotion wheel proposed by Yang et al. [32]. This is a reduced version involving only the emotions used in this paper.
Figure 1. The 2D arousal-valence emotion wheel proposed by Yang et al. [32]. This is a reduced version involving only the emotions used in this paper.
Entropy 24 01250 g001
Figure 2. Experimental results of trade-off parameter sensitivity analysis for MDAR, where the λ and μ are set as λ = [ 1 : 10 ] , μ = 50 for the left and λ = 5 , μ = [ 1 : 10 ] for the right in (a), and λ = [ 1 : 10 ] , μ = 80 for the left and λ = 5 , μ = [ 70 : 90 ] for the right in (b).
Figure 2. Experimental results of trade-off parameter sensitivity analysis for MDAR, where the λ and μ are set as λ = [ 1 : 10 ] , μ = 50 for the left and λ = 5 , μ = [ 1 : 10 ] for the right in (a), and λ = [ 1 : 10 ] , μ = 80 for the left and λ = 5 , μ = [ 70 : 90 ] for the right in (b).
Entropy 24 01250 g002
Table 1. The sample statistics of the selected speech corpora used in cross-corpus SER tasks.
Table 1. The sample statistics of the selected speech corpora used in cross-corpus SER tasks.
TasksSpeech Corpus (# Samples from Each Emotion)#Sample
B→EEmoDB (AN: 127, SA: 62, FE: 69, HA: 71, DI: 46)375
E→BeNTERFACE (AN: 211, SA: 211, FE: 211, HA: 208, DI: 211)1052
B→CEmoDB (AN: 127, SA: 62, FE: 69, HA: 71, NE: 79)408
C→BCASIA (AN: 200, Sad: 200, FE: 200, HA: 200, NE: 200)1000
E→CeNTERFACE (AN: 211, SA: 211, FE: 211, HA: 208, SU: 211)1052
C→ECASIA (AN: 200, SA: 200, FE: 200, HA: 200, SU: 200)1000
Table 2. The comparison results among all the transfer subspace learning methods of using IS09 as feature set, in which the best results are highlighted in bold.
Table 2. The comparison results among all the transfer subspace learning methods of using IS09 as feature set, in which the best results are highlighted in bold.
MethodB→ EE→BB→CC→BE→CC→EAverage
SVM28.9323.5829.6035.0126.1025.1428.06
TCA30.5244.0333.4045.0731.1032.3236.07
GFK32.1142.4833.1048.0832.8028.1336.17
SA33.5043.8935.8049.0332.6028.1736.33
DoSL36.1238.9534.4045.7530.4031.5936.20
JDAR36.3339.9731.1046.2932.4031.5036.27
MDAR36.5240.2933.1047.3231.7031.2136.69
Table 3. The comparison results among all the transfer subspace learning methods of using IS10 as feature set, in which the best results are highlighted in bold.
Table 3. The comparison results among all the transfer subspace learning methods of using IS10 as feature set, in which the best results are highlighted in bold.
MethodB→ EE→BB→CC→BE→CC→EAverage
SVM34.5028.1335.3035.2924.3026.8130.73
TCA32.6044.5340.5051.4733.2029.7738.68
GFK36.0140.1140.0045.9333.0029.0937.35
SA35.6543.9237.5047.0632.1030.6137.80
DoSL36.8243.3336.8048.4535.6033.9139.15
JDAR37.9547.8042.7048.9735.6037.5841.76
MDAR38.9048.9543.0049.5235.8037.3042.26
Table 4. The comparison results between our MDAR of using IS09 and IS10 as feature sets and all the deep transfer learning methods, in which the best results are highlighted in bold.
Table 4. The comparison results between our MDAR of using IS09 and IS10 as feature sets and all the deep transfer learning methods, in which the best results are highlighted in bold.
MethodB→ EE→BB→CC→BE→CC→EAverage
AlexNet29.4931.0333.2041.9127.8027.2531.78
DAN36.1340.4139.0049.8529.0031.4737.64
DANN33.3843.6839.2053.7129.8029.2538.05
Deep-CORAL35.0343.3838.3048.2831.0030.8937.81
DSAN36.2546.9040.3050.6928.7032.6139.41
DTTRN37.7048.2040.4055.2031.2033.6041.10
MDAR + IS0936.5240.2933.1047.3231.7031.6836.69
MDAR + IS1038.9048.9543.0049.5235.8037.3042.26
Table 5. Comparison between the MDAR model, with and without the guidance of the disturbance strategy, in the cross-corpus SER experiments using the IS09 speech feature set. The best result in each task are highlighted in bold.
Table 5. Comparison between the MDAR model, with and without the guidance of the disturbance strategy, in the cross-corpus SER experiments using the IS09 speech feature set. The best result in each task are highlighted in bold.
Rough Emotion GroupsB→ EE→BB→CC→BE→CC→EAverage
Proposed Modification36.5240.2933.1047.3231.7031.6836.69
Original Version36.3339.8931.5046.5132.0031.5036.28
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zong, Y.; Lian, H.; Chang, H.; Lu, C.; Tang, C. Adapting Multiple Distributions for Bridging Emotions from Different Speech Corpora. Entropy 2022, 24, 1250. https://doi.org/10.3390/e24091250

AMA Style

Zong Y, Lian H, Chang H, Lu C, Tang C. Adapting Multiple Distributions for Bridging Emotions from Different Speech Corpora. Entropy. 2022; 24(9):1250. https://doi.org/10.3390/e24091250

Chicago/Turabian Style

Zong, Yuan, Hailun Lian, Hongli Chang, Cheng Lu, and Chuangao Tang. 2022. "Adapting Multiple Distributions for Bridging Emotions from Different Speech Corpora" Entropy 24, no. 9: 1250. https://doi.org/10.3390/e24091250

APA Style

Zong, Y., Lian, H., Chang, H., Lu, C., & Tang, C. (2022). Adapting Multiple Distributions for Bridging Emotions from Different Speech Corpora. Entropy, 24(9), 1250. https://doi.org/10.3390/e24091250

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop