Next Article in Journal
Deep Learning-Enabled Improved Direction-of-Arrival Estimation Technique
Next Article in Special Issue
Research on Lightweight-Based Algorithm for Detecting Distracted Driving Behaviour
Previous Article in Journal
Identity Recognition System Based on Multi-Spectral Palm Vein Image
Previous Article in Special Issue
A Deep Learning-Enhanced Multi-Modal Sensing Platform for Robust Human Object Detection and Tracking in Challenging Environments
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multimodal Sentiment Analysis in Realistic Environments Based on Cross-Modal Hierarchical Fusion Network

1
Hubei Key Laboratory of Intelligent Vision Based Monitoring for Hydroelectric Engineering, China Three Gorges University, Yichang 443002, China
2
College of Computer and Information Technology, China Three Gorges University, Yichang 443002, China
3
College of Information Science and Technology, Hangzhou Normal University, Hangzhou 310000, China
*
Author to whom correspondence should be addressed.
Electronics 2023, 12(16), 3504; https://doi.org/10.3390/electronics12163504
Submission received: 12 July 2023 / Revised: 11 August 2023 / Accepted: 14 August 2023 / Published: 18 August 2023

Abstract

:
In the real world, multimodal sentiment analysis (MSA) enables the capture and analysis of sentiments by fusing multimodal information, thereby enhancing the understanding of real-world environments. The key challenges lie in handling the noise in the acquired data and achieving effective multimodal fusion. When processing the noise in data, existing methods utilize the combination of multimodal features to mitigate errors in sentiment word recognition caused by the performance limitations of automatic speech recognition (ASR) models. However, there still remains the problem of how to more efficiently utilize and combine different modalities to address the data noise. In multimodal fusion, most existing fusion methods have limited adaptability to the feature differences between modalities, making it difficult to capture the potential complex nonlinear interactions that may exist between modalities. To overcome the aforementioned issues, this paper proposes a new framework named multimodal-word-refinement and cross-modal-hierarchy (MWRCMH) fusion. Specifically, we utilized a multimodal word correction module to reduce sentiment word recognition errors caused by ASR. During multimodal fusion, we designed a cross-modal hierarchical fusion module that employed cross-modal attention mechanisms to fuse features between pairs of modalities, resulting in fused bimodal-feature information. Then, the obtained bimodal information and the unimodal information were fused through the nonlinear layer to obtain the final multimodal sentiment feature information. Experimental results on the MOSI-SpeechBrain, MOSI-IBM, and MOSI-iFlytek datasets demonstrated that the proposed approach outperformed other comparative methods, achieving Has0-F1 scores of 76.43%, 80.15%, and 81.93%, respectively. Our approach exhibited better performance, as compared to multiple baselines.

1. Introduction

In the research field of intelligent human–computer interaction, the ability to recognize, analyze, understand, and express emotions is essential for intelligent machines. Therefore, the utilization of computer technology to automatically recognize, understand, analyze, classify, and respond to emotion holds significant value for establishing a harmonious human–machine interaction environment, improving interaction efficiency, and enhancing user experience [1,2,3]. Previous studies [4,5] have primarily focused on sentiment analysis using textual data and have achieved remarkable accomplishments. However, as compared to unimodal analysis, MSA can effectively leverage the coordinated and complementary information from different modalities to enhance emotional understanding and expression capabilities and provide richer information that is more consistent with human behavior.
In recent years, there has been a growing interest in multimodal data for sentiment analysis. MSA aims to utilize the information interaction between texts, images, speech, etc., enabling machines to automatically utilize comprehensive multimodal emotional information for the identification of users’ sentiment tendencies. Early research often employed multimodal fusion [6] through early fusion, directly combining multiple sources of raw features [7,8], or late fusion, aggregating the decisions of multiple sentiment classifiers [9,10,11]. However, the former approach may result in a large number of redundant input vectors, leading to increased computational complexity, while the latter approach may struggle to capture the correlations between different modalities. Therefore, various methods have been proposed for feature fusion in multimodal sentiment analysis. Existing fusion methods include those based on simple operations [12,13], attention-based methods [14,15,16,17], tensor-based methods [18], translation-based methods [19], GAN-based methods [20], routing-based methods [21], and hierarchical fusion [22,23,24]. Although there is a wide range of fusion methods, attention-based fusion methods have shown superior efficiency and performance [25]. However, weighting and summing the features of each modality in the attention mechanism alone may not be able to effectively adapt to the differences in the features across different modalities. Consequently, certain modal features might be disregarded or underestimated, ultimately impacting the accuracy of the fused feature representation. Additionally, complex nonlinear interactions may exist between different modalities, and the attention mechanism may struggle to model such relationships accurately, thereby impacting the effectiveness of feature fusion. Furthermore, previous methods have rarely considered the simultaneous utilization of the interaction information within a single modality and between modalities.
In addition to fusion mechanisms, handling noise in the data is crucial for MSA. While many excellent MSA models have been proposed, they are rarely deployed in the real world. The reason is that during the process of human–machine interaction in real environments, text acquisition can only be achieved through ASR models, which convert speech into texts. Through an in-depth analysis of ASR output, we have discovered that errors in recognizing emotional words in ASR-generated texts directly impair MSA models.
To address these challenges, we propose a MWRCMH framework that reduces speech recognition errors and incorporates cross-modal hierarchical fusion. Specifically, we utilized a language model to capture syntactic and semantic information, enabling us to determine the most likely positions of sentiment words. Furthermore, we leveraged multimodal sentiment information to dynamically complement the emotional semantics in the representation of these positions. The corrected representation of sentiment words was then combined with images and sounds, allowing us to capture the interaction information between pairs of modalities using a cross-modal attention mechanism. The fused bimodal feature information was then concatenated with the feature information from each mode, resulting in a vector that encompassed all the feature information. This vector served as the input, passing through the nonlinear layers for fusion. Finally, we employed nonlinear layers to map the concatenated feature vector to a higher-dimensional space, utilizing the output as the final feature representation for training and prediction in subsequent tasks.
The remaining structure of this paper is as follows: Section 2 provides a review of related literature. In Section 3, we present a detailed description of our proposed method. Subsequently, in Section 4, we elaborate the experimental results. Finally, Section 5 concludes the paper, and we discuss future work.

2. Related Work

Sentiment computing, as an emerging interdisciplinary research field, has been widely studied and explored since its introduction in 1995 [26]. Previous research has primarily focused on unimodal data representation and multimodal fusion. In terms of unimodal data representation, Pang et al. [4] were the first to employ machine learning-based methods to address textual sentiment classification, achieving better results than traditional manual annotation, by using movie reviews as the dataset. Yue et al. [5] proposed a hybrid model called Word2vec-BiLSTM-CNN, which leveraged the feature extraction capability of convolutional neural networks (CNN) and the ability of bi-directional long short-term memory (Bi-LSTM) to capture short-term bidirectional dependencies in text. Their results demonstrated that hybrid network models outperformed single-structure neural networks in the context of short texts. Colombo et al. [27] segmented different regions in image and video data based on features such as color, warmth, position, and size, enabling their method to obtain higher semantic levels beyond the objects themselves. They applied this approach to a sentiment analysis of art-related images. Wang et al. [28] utilized neural networks for the facial feature extraction of images. Bonifazi et al. [29] proposed a space–time framework that leveraged the emotional context inherent in a presented situation. They employed this framework to extract the scope of emotional information concerning users’ sentiments on a given subject. However, the use of unimodal analysis in sentiment analysis has had some limitations since humans express emotions through various means, including sound, content, facial expressions, and body language, all of which are collectively employed to convey emotions.
Multimodal data describe objects from different perspectives, providing richer information, as compared to unimodal data. Different modalities of information can complement each other in terms of content. In the context of multimodal fusion, previous research can be categorized into three stages: early feature fusion, mid-level model fusion, and late decision fusion. Wollmer et al. [30] and Rozgic et al. [31] integrated data from audio, visual, and text sources to extract emotions and sentiments. Metallinou et al. [32] and Eyben et al. [33] combined audio and text patterns for emotional recognition. These methods relied on early feature fusion, which mapped them to the same embedding space through simple concatenation, and the lack of interaction between different modalities. For late-stage decision-fusion methods, internal representations were initially learned within each modality, followed by learning the fusion between modalities. Zadeh et al. [18] utilized tensor-fusion networks to calculate the outer product between unimodal representations, yielding tensor representations. Liu et al. [34] introduced a low-rank multimodal-fusion method to reduce the computational complexity of tensor-based approaches. These methods aimed to enhance efficiency by decomposing the weights of high-dimensional fusion tensors, reducing redundant information, yet they struggled to effectively simulate intermodal or specific modality dynamics. Intermediate model fusion amalgamates the advantages of both early feature fusion and late-stage decision fusion, facilitating the selection of fusion points and enabling multimodal interaction. Poria et al. [35] further extended the combination of convolutional neural networks (CNNs) and multiple kernel learning (MKL). In contrast to Ghosal et al. [36], Poria et al. utilized a novel fusion method to effectively enhance fusion features. Zhang et al. [37] introduced a quantum-inspired framework for the sentiment analysis of bimodal data (texts and images) to address semantic gaps and model the correlation between the two modalities using density matrices. However, these methods exhibited limited adaptability to feature differences and suffered from significant feature redundancy. Concerning hierarchical fusion, Majumder et al. [22] employed a hierarchical-fusion strategy, initially combining two modalities and, subsequently, integrating all three modalities. However, this approach struggled to adequately capture intramodal dynamics. Georgiou et al. [23] introduced a deep hierarchical-fusion framework, applying it to sentiment-analysis problems involving audio and text modalities. Yan et al. [24] introduced a hierarchical attention-fusion network for geographical localization. Nevertheless, these methods overlooked the potential existence of complex nonlinear interactions between modalities. Moreover, many fusion approaches have seldomly considered simultaneously harnessing intramodal and intermodal interactions. Verma et al. [38] emphasized that each modality possesses unique intramodality features, and multimodal sentiment analysis methods should capture both common intermodality information and distinctive intramodality signals.
In addition to considering fusion strategies in the MSA model, it is crucial to address the noise present in modal data. Pham et al. [39] proposed the MCTN model to handle the potential absence of visual and acoustic data. Liang et al. [40] and Mittal et al. [41] also focused on addressing noise introduced by visual and acoustic data, relying on word-level features there were obtained by aligning the audio with the actual text. Xue et al. [42] introduced a multi-level attention-graph network to reduce noise within and between modalities. Cauteruccio et al. [43] introduced a string-comparison metric that could be employed to enhance the processing of heterogeneous audio samples, mitigating modality-related noise. However, these models did not investigate the impact of ASR errors on the MSA model. Notably, Wu et al. [44] utilized a sentiment word position-detection module to determine the most likely positions of sentiment words in text. They dynamically refined sentiment-word embeddings using a multimodal sentiment-word-refinement module that incorporated the improved embeddings as the textual input for the multimodal feature-fusion module. This approach reduced the influence of ASR errors on the MSA model. The sentiment word position-detection module and multimodal sentiment-word-refinement module have proven highly effective, achieving state-of-the-art performance on real-world datasets. However, the original SWRM simply concatenated modalities in feature fusion without capturing intramodal and intermodal features, even when genuine correlations existed.
In this work, we employed a cross-modal hierarchical fusion network for feature fusion. This enabled the model to focus on important features, reduce noisy impacts, and achieve superior fusion results. Thus, our approach enhanced the robustness and performance of the model.

3. Methodology

To enhance the adaptability to feature differences across different modalities and effectively model complex nonlinear interactions, we designed a cross-modal, hierarchical-fusion network that considered the presence of sentiment-word-recognition errors in the texts generated by ASR. The model’s architecture is depicted in Figure 1, and it comprises 4 main components: unimodal-feature extraction, a multimodal-word-refinement module, a multimodal-feature-fusion module, and a prediction network. Initially, we employed BERT to extract textual feature information while Bi-GRU was utilized to extract audio and visual feature information. BERT, rooted in the transformer architecture, served as a pre-trained language model capable of capturing the nuanced semantic information of words within sentences through bidirectional contextual modeling [45]. Bi-GRU, a variant of a recurrent neural network (RNN), was particularly well suited for modeling sequential data [46]. Subsequently, the multimodal-word-refinement module dynamically supplemented missing sentiment semantics by leveraging multimodal-sentiment information, resulting in the generation of new word embeddings. These new word embeddings were then input into the multimodal-feature-fusion module, which utilized a cross-modal hierarchical-fusion network to perform feature fusion across different modalities. Finally, a nonlinear layer was employed to predict the final sentiment-regression labels, facilitating accurate sentiment judgment.

3.1. Multimodal Word-Refinement Module (MWRM)

We drew inspiration from the SWRM model of Wu et al. [44]. Firstly, we employed a language model to estimate the potential positions of sentiment words in the ASR transcripts by capturing syntactic and semantic information. For example, given a sentence w 1 , w 2 , , w n t , let us assume that the i word in the sentence was a sentiment word; we replaced the word with a special word (MASK). Then, we used BERT and Bi-GRU to model the word embeddings and the visual, audio, and visual–audio features separately, resulting in h i = { h l i : 1 l n i , h l i d h i } , i { t , v , a } . In addition, we also used a Bi-GRU network to fuse acoustic and visual features in order to capture advanced sentiment semantics and obtained h v a = { h l v a : 1 l n i , h l v a d h v a } . The formulas are presented below:
h t = BERT t ( x t )
h v = B i G R U v ( u v )
h a = B i G R U a ( u a )
h v a = B i G R U v a ( [ u v ; u a ] )
where the three modalities of unaligned features, word embeddings, visual features, and audio features are represented as x m = { x l m : 1 l n i , x l m d x m } , m { t , v , a } . To obtain the corresponding multimodal information, we used a pseudo-alignment method to align the features. We divided the acoustic and visual features into non-overlapping feature groups with lengths of n a / n t and n v / n t , respectively, and took the average of the features in each group to obtain the aligned feature representation u i = { u l i : 1 l n l , u l i d x i } , i { v , a } .
Then, we used a multimodal, gated network to filter the representation information at the locations of the sentiment word to generate r v . To make the model ignore impossible solutions, we used the gate mask p . Therefore, g v is the gate value obtained by connecting the unimodal contextual perception representation h s t , h s v , h s a , and the bimodal representation h s v a in the position s , using a nonlinear network. We then used the word embedding x c l s of the candidate word c l s and the multimodal representation h s v , h s a and h s v a , to obtain the attention score through linear fusion. We fed the attention score to the softmax function to obtain the attention weight, which was then applied to the candidate word embeddings to obtain the sentiment embedding r e , as follows:
g v = S i g mo i d ( W 1 ( [ h s t ; h s v ; h s a ; h s v a ] ) + b 1 )
r v = ( 1 g v p ) x s t
g l e = W 2 ( [ x c l s ; h s v ; h s a ; h s v a ] ) + b 2
r e = s o f t max ( g l e ) x c l s ,
where W 1 1 × Σ i { t , v , a , v a } d h i , b 1 1 are the parameters of the multimodal, gated network, W 2 1 × ( d x t + Σ i { v , a , v a } d h i ) , b 2 1 are the parameters of the multimodal sentiment-word attention network, and ⊙ represents element-wise multiplication.
As the sentiment word we wanted may not have been included in the candidate word list, we also introduced the representation of (MASK) when correcting the word representation, that is x m a s k . Finally, we combined the above representations to obtain the corrected word vector r t , as follows:
g m a s k = S i g m o i d ( W 3 ( [ r e ; x m a s k ] ) + b 3 )
r t = ( g v p ) ( g m a s k r e + ( 1 g m a s k ) x m a s k ) + r v ,
where W 3 1 × 2 d x t , b 3 1 are the trainable parameters of the nonlinear network.

3.2. Unimodal Feature Extraction

We utilized the BERT model to extract features from the improved word embeddings z t = { x 1 t , x 2 t , , r t , , x n l t } , which are represented as s t :
s t = B E R T t e x t u a l ( z t ) .
For the visual and audio modalities, we employed two separate Bi-GRU networks to encode the visual and audio features. The representation of the first word was used as the visual representation s v and acoustic representation s a :
s v = B i G R U v i s u a l ( x v )
s a = B i G R U a u d i o ( x a )
Unimodal features can have different dimensions, i.e., d t d a d v . Therefore, they were mapped to the same dimension, such as d m , using a fully connected layer, and computed as follows:
ω t = r e l u ( W t s t + b t )
ω a = r e l u ( W a s a + b a )
ω v = r e l u ( W v s v + b v )
where W t d m × d t , b t d m are the trainable parameters of the textual feature-extraction network, W a d m × d a , b a d m are the trainable parameters of the audio feature-extraction network, and W v d m × d v , b v d m are the trainable parameters of the visual feature-extraction network.

3.3. Cross-Modal Attention (CMA)

In general, the three modalities of word-embedding, vision, and audio had some content that was related to sentiment information. If the interaction of the modality information was not considered, the final multimodal features obtained would inevitably contain a large amount of redundant information that would be irrelevant to the detection task. Therefore, we used a cross-modal attention framework, where the attention function was applied to the representations of the two modalities, i.e., visual–textual, textual–audio, and audio–visual. The cross-modal attention calculation for visual–textual ( ω v , ω t ) was as follows:
The feature matrices for visual ω v and textual ω t were multiplied to obtain the cross-modal information matrix B 1 , B 2 u × u , where · represents matrix multiplication.
B 1 = ω v ( ω t ) T , B 2 = ω t ( ω v ) T
Through a line-by-line softmax function, the attention distribution was calculated, N 1 , N 2 u × u , where N 1 ( i , j ) represents the correlation between the i feature of the word-embedding modality and the j feature of the visual modality: the larger the value, the stronger the interaction between the two features and the more important the fusion information.
N 1 = softmax ( B 1 ) , N 2 = softmax ( B 2 )
The attention distribution was then multiplied by the feature matrix to obtain the final attention representation matrix, O 1 , O 2 u × d m .
O 1 = N 1 ω t , O 2 = N 2 ω v
Finally, a multiplication gate mechanism was used to obtain the mutual-attention-information matrix of the two modalities A 1 , A 2 u × d m , where ⊙ represents element-wise matrix multiplication.
A 1 = O 1 ω v , A 2 = O 2 ω t
By transposing A 1 and A 2 to the desired dimensions and concatenating them, the fused information representation of the word embedding and visual modalities was obtained: C m a t v u × 2 d m , where ⊕ represents the concatenation operation of vectors.
C m a t v = ( A 1 ) T ( A 2 ) T
Using the same method, the fused features of the textual–audio modality C m a t a and the audio–visual modality could be obtained: C m a a v .

3.4. Trimodal Fusion

After concatenating the aforementioned cross-modal attention-fused features C m a t v , C m a t a , and C m a a v , as well as the unimodal features ω t , ω a , and ω v , we used fully connected layers to fuse features from different sources, enabling the model to fully utilize various feature information. The nonlinear tanh activation function was used to enhance the expressive power of the features. The equation was as follows:
s f = C m a t v C m a t a C m a a v ω t ω a ω v
s F = tanh ( W f s f + b f )
where W f d F × d f , b f d F are the trainable parameter of the trimodal fusion network.

3.5. Regression Analysis

In the regression analysis, we utilized a nonlinear layer to predict the final sentiment regression labels, as shown in the following equation:
s c o r e = tanh ( W F s F + b F )
where W F 1 × d F , b F 1 are the trainable parameters of the prediction network.

4. Experiments

4.1. Dataset

We used three real-world datasets constructed from the CMU-MOSI [47] dataset, including MOSI-SpeechBrain, MOSI-IBM, and MOSI-iFlytek. The CMU-MOSI dataset is one of the most popular multimodal sentiment analysis datasets and contains 93 monologue clips of YouTube movie-review videos. These clips were divided into 2199 annotated video segments with emotion labels. Sentiment scores ranging from −3 (strongly negative) to 3 (strongly positive) were also annotated. For the multimodal features, we used Facet to extract the visual features. Facet can extract facial-action units, facial landmarks, head poses, and gaze-tracking data from each frame, resulting in a comprehensive representation across 34 dimensions. Facet achieves the visualization of relationships and trends among multiple dimensions by mapping various dimensions of the data onto distinct visual channels. The acoustic features were obtained by applying COVAREP [48], which includes 12 Mel-frequency cepstral coefficients (MFCC) and other low-level features.
However, the text data in the MOSI dataset was sourced from the subtitles of movies and TED Talks as transcriptions and have been expertly corrected and processed for speech. However, obtaining text for real-world applications in this way is unrealistic. To evaluate models in the real world, we manually replaced the raw text data in the dataset with text generated by ASR models. We used SpeechBrain’s ASR model and two widely used commercial APIs to generate text. The ASR model used, as published by Ravanelli et al. [49], was built on a transformer encoder–decoder framework and trained on the Librispeech dataset [50]. The commercial APIs we used were IBM and iFlytek speech-to-text, which are widely used by researchers and software developers. Finally, we transcribed the videos into text using the three ASR models and then built three new datasets: MOSI-SpeechBrain, MOSI-IBM, and MOSI-iFlytek. The percentages of the emotional-word-replacement errors for the MOSI dataset using the 3 ASR APIs are listed in Table 1. The percentage of emotional-word-replacement errors for MOSI-IBM was 17.6%, meaning that about 17 out of 100 utterances had this type of error.

4.2. Implementation Details

To ensure fairness, we used the same parameters for the experiments conducted on all 3 datasets. We used Adam as the optimizer, and the learning rate was 5 × 10−5. The batch size was 32. The dropout was 0.1. In addition, in Section 3, the values d t , d a , and d v were 768,128 and 128, respectively. We implemented our entire method based on PyTorch, and all experiments were run on an NVIDIA GeForce RTX 3080 GPU.

4.3. Baseline

To validate the effectiveness of the model, we conducted a comparative analysis of established and representative multimodal sentiment-analysis models, including:
TFN (Zadeh et al., 2017) [18], which uses a tensor-fusion network to create a multi-dimensional tensor to learn unimodal, bimodal, and trimodal information in multimodal data;
LMF (Liu et al., 2018) [34], which decomposes the weight tensor using a low-rank multimodal-fusion approach, which reduces the computational complexity of tensor-based methods, and efficiently fuses multimodal interactions with modality-specific low-rank factors;
MulT (Tsai et al., 2019) [15], which proposes a multimodal transformer that utilizes cross-modal attention to enable interactions between multimodal sequences across different time steps;
MISA (Hazarika et al., 2020) [51], which uses multitask learning to partition each modality into different subspaces and learn related sentiment representations that aid the fusion process;
Self-MM (Yu et al., 2021) [52], which learns rich unimodal representations by training on 1 multimodal task and 3 unimodal subtasks; and
SWRM (Wu et al., 2022) [44], which proposes a multimodal word correction model based on sentiment word perception and utilizes multimodal sentiment information to dynamically complete damaged sentiment semantics.

4.4. Results and Analysis

In this section, we evaluated the performance of our model on the MOSI-SpeechBrain, MOSI-IBM, and MOSI-iFlytek datasets using various metrics, including accuracy (Acc2), F1-score (F1) in binary classification, mean absolute error (MAE), and correlation (Corr). Specifically, we calculated the Acc2 and F1-scores in 2 ways: negative/non-negative (Non0-Acc, Non0-F1) and negative/positive (Has0-Acc, Has0-F1). Since the predicted results were continuous values, we converted them into emotion classification labels by mapping the emotion scores to the labels. The experimental results for different models are presented in Table 2.
Based on the Has0-F1 values in the table, we drew the following conclusions: Compared to the tensor-fusion methods TFN and LMF, our model demonstrated a significant improvement of 7.48% and 6.65% on the MOSI-SpeechBrain dataset, respectively; 8.37% and 7.06% on the MOSI-IBM dataset, respectively; and 9.92% and 9.90% on the MOSI-iFlytek dataset, respectively, highlighting the superiority of our model. This superiority stemmed from the fact that TFN and LMF performed tensor fusion at an overall level without adequately considering the interaction modeling between the modalities, whereas our fusion network took this aspect into account.
Additionally, as compared to the attention-based cross-modal model MulT, our model achieved improvements of 4.73%, 4.61%, and 4.88% on the MOSI-SpeechBrain, MOSI-IBM, and MOSI-iFlytek datasets, respectively. The reason behind this performance gain was that MulT overlooked the feature differences between the different modalities and the potential nonlinear interaction relationships. In contrast, our hierarchical fusion approach incorporated a nonlinear layer for feature fusion, leading to better results. Furthermore, as compared to the modality-invariant and specificity-based fusion model MISA, our model outperformed it by achieving improvements of 2.58%, 3.16%, and 2.31% on the MOSI-SpeechBrain, MOSI-IBM, and MOSI-iFlytek datasets, respectively. This improvement could be attributed to our model’s focus on capturing the semantic representations of different modalities, resulting in enhanced performance. Moreover, our model, named Self-MM, surpassed the self-supervised and multitasking-representation learning model Self-MM with improvements of 2.71%, 2.78%, and 1.67% on the MOSI-SpeechBrain, MOSI-IBM, and MOSI-iFlytek datasets, respectively. These findings validated the superior performance of our model, which benefited from its unique design and approach. Lastly, in comparison with the baseline model SWRM, our model achieved improvements of 1.81%, 1.68%, and 1.64% on the MOSI-SpeechBrain, MOSI-IBM, and MOSI-iFlytek datasets, respectively. This improvement could be attributed to the utilization of our cross-modal hierarchical-fusion network, which enabled better and more effective multimodal fusion, leading to enhanced overall performance.
Figure 2, Figure 3 and Figure 4 show the accuracy of the model regression on the MOSI-SpeechBrain, MOSI-IBM, and MOSI-iFlytek datasets. Train-Has0-Acc and Train-Has0-F1 were correlated to Has0-Acc and Has0-F1, as generated by the training datasets. Val-Has0-Acc and Val-Has0-F1 were generated by the validation datasets. The models were trained in 10 epochs using 70% of the samples as the training set, and 30% of samples were used as test data.

4.5. Ablation Study

4.5.1. Modal Ablation

The literature [15,24] has demonstrated the superiority of multimodal analysis over unimodal analysis. We also observed the same trend in our experiments, where the trimodal and bimodal classifiers outperformed the unimodal classifier. Among the different modalities, the textual modality exhibited the highest accuracy in the unimodal classification, reaching 73.32%. While the other modalities contributed to enhancing the performance of the multimodal classifier, their impact was relatively small, as compared to the textual modality.
To validate the significant influence of the multimodal information fusion on sentiment-analysis accuracy, we conducted experiments using eight different combinations of modality features: unimodal (T, A, V), bimodal (V + T, T + A, A + V), trimodal (T + V + A), and composite hierarchical fusion (TVA + T + V + A). For the unimodal feature information, we processed the data through a feature-extraction layer and directly utilized the results for sentiment analysis. Similarly, for the bimodal features, we initially fused pairwise combinations of the different modalities and then applied the same processing for the sentiment analysis. Regarding the trimodal feature information, we fused the bimodal features T + A, T + V, and V + T to obtain the trimodal fusion T + A + V, followed by applying the same processing for sentiment analysis. Finally, we employed the fusion processing method proposed in this paper, merging the trimodal information with the unimodal information to obtain the final trimodal features (TVA + T + V + A) for sentiment analysis. The detailed results of sentiment analysis are presented in Table 3.
Considering the evaluation metrics for the regression results, the bimodal sentiment analysis generally outperformed the unimodal sentiment analysis, and the best performance was achieved when utilizing the fused trimodal features. Therefore, the effective fusion of textual, acoustic, and visual features contributed to enhancing the performance of sentiment classification.

4.5.2. Model Ablation

We conducted ablation experiments to assess the contributions of each component in our model. Our architecture comprised several different variants. MWRCMH represented the complete model proposed by us. MWRCMH without fusion excluded the cross-modal hierarchical-fusion module and simply concatenated the feature vectors from the different modalities. MWRCMH without MWRM did not employ the multimodal word-correction module and did not consider the impact of the recognition errors of the sentiment words in the ASR transcripts. The results of the model variants are presented in Table 4. Without the multimodal word-correction module, MWRCMH without MWRM yielded inferior results, as compared to MWRCMH, highlighting the significance of identifying appropriate words for refinement. Nevertheless, the results outperformed those of the models in the baseline experiment, thereby demonstrating the effectiveness and superiority of our fusion network. Moreover, the results of MWRCMH without fusion, which lacks the cross-modal hierarchical-fusion module, further substantiated this conclusion.

5. Conclusions

This paper introduced a cross-modal hierarchical fusion network-based multimodal sentiment-analysis architecture. Through the incorporation of the multimodal word-correction module and the cross-modal hierarchical-fusion module, our architecture effectively mitigated the sentiment word-recognition errors stemming from speech recognition, augmented the adaptability across diverse modal characteristics, and enhanced the nonlinear modeling capabilities. Our proposed architecture demonstrated strong performance across the MOSI-SpeechBrain, MOSI-IBM, and MOSI-iFlytek datasets. However, the architecture disregarded the fine-grained relationships between the textual, audio, and visual information, and it also exhibited certain issues related to feature redundancy.
In forthcoming work, we intend to address the alignment challenges presented by various modalities while concurrently minimizing modal-feature-information redundancy. Furthermore, we will place greater emphasis on rectifying contribution imbalances within multimodal learning.

Author Contributions

Conceptualization, J.H.; methodology, J.H. and F.W.; software, J.H.; validation, J.H.; formal analysis, J.H.; investigation, F.W. and S.S.; resources, J.H. and P.L.; data curation, J.H.; writing—original draft preparation, J.H.; writing—review and editing, F.W. and S.S.; visualization, J.H.; supervision, S.S. and F.W.; project administration, S.S.; funding acquisition, S.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China, grant number U1703261.

Data Availability Statement

Publicly available datasets were analyzed in this study. The data can be found at: https://github.com/huangju1/MWRCMH (accessed on 1 August 2023) dataset.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Qin, Z.; Zhao, P.; Zhuang, T.; Deng, F.; Ding, Y.; Chen, D. A survey of identity recognition via data fusion and feature learning. Inf. Fusion 2023, 91, 694–712. [Google Scholar] [CrossRef]
  2. Tu, G.; Liang, B.; Jiang, D.; Xu, R.J.I.T.o.A.C. Sentiment-Emotion-and Context-guided Knowledge Selection Framework for Emotion Recognition in Conversations. IEEE Trans. Affect. Comput. 2022, 1–14. [Google Scholar] [CrossRef]
  3. Noroozi, F.; Corneanu, C.A.; Kamińska, D.; Sapiński, T.; Escalera, S.; Anbarjafari, G. Survey on emotional body gesture recognition. IEEE Trans. Affect. Comput. 2018, 12, 505–523. [Google Scholar] [CrossRef]
  4. Pang, B.; Lee, L.; Vaithyanathan, S. Thumbs up? Sentiment classification using machine learning techniques. arXiv 2002, arXiv:cs/0205070. [Google Scholar]
  5. Yue, W.; Li, L. Sentiment analysis using Word2vec-CNN-BiLSTM classification. In Proceedings of the 2020 Seventh International Conference on Social Networks Analysis, Management and Security (SNAMS), Paris, France, 14–16 December 2020; pp. 1–5. [Google Scholar]
  6. Atrey, P.K.; Hossain, M.A.; El Saddik, A.; Kankanhalli, M. Multimodal fusion for multimedia analysis: A survey. Multimed. Syst. 2010, 16, 345–379. [Google Scholar] [CrossRef]
  7. Mazloom, M.; Rietveld, R.; Rudinac, S.; Worring, M.; Van Dolen, W. Multimodal popularity prediction of brand-related social media posts. In Proceedings of the Proceedings of the 24th ACM International Conference on Multimedia, Amsterdam, The Netherlands, 15–19 October 2016; pp. 197–201. [Google Scholar]
  8. Pérez-Rosas, V.; Mihalcea, R.; Morency, L.-P. Utterance-level multimodal sentiment analysis. In Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Sofia, Bulgaria, 4–9 August 2013; pp. 973–982. [Google Scholar]
  9. Poria, S.; Cambria, E.; Gelbukh, A. Deep convolutional neural network textual features and multiple kernel learning for utterance-level multimodal sentiment analysis. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, Lisbon, Portugal, 17–21 September 2015; pp. 2539–2544. [Google Scholar]
  10. Liu, N.; Dellandréa, E.; Chen, L.; Zhu, C.; Zhang, Y.; Bichot, C.-E.; Bres, S.; Tellez, B. Multimodal recognition of visual concepts using histograms of textual concepts and selective weighted late fusion scheme. Comput. Vis. Image Underst. 2013, 117, 493–512. [Google Scholar] [CrossRef]
  11. Yu, W.; Xu, H.; Meng, F.; Zhu, Y.; Ma, Y.; Wu, J.; Zou, J.; Yang, K. Ch-sims: A chinese multimodal sentiment analysis dataset with fine-grained annotation of modality. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Online, 5–10 July 2020; pp. 3718–3727. [Google Scholar]
  12. Poria, S.; Chaturvedi, I.; Cambria, E.; Hussain, A. Convolutional MKL based multimodal emotion recognition and sentiment analysis. In Proceedings of the 2016 IEEE 16th International Conference on Data Mining (ICDM), Barcelona, Spain, 12–15 December 2016; pp. 439–448. [Google Scholar]
  13. Nguyen, D.; Nguyen, K.; Sridharan, S.; Dean, D.; Fookes, C. Deep spatio-temporal feature fusion with compact bilinear pooling for multimodal emotion recognition. Comput. Vis. Image Underst. 2018, 174, 33–42. [Google Scholar] [CrossRef]
  14. Lv, F.; Chen, X.; Huang, Y.; Duan, L.; Lin, G. Progressive modality reinforcement for human multimodal emotion recognition from unaligned multimodal sequences. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 2554–2562. [Google Scholar]
  15. Tsai, Y.-H.H.; Bai, S.; Liang, P.P.; Kolter, J.Z.; Morency, L.-P.; Salakhutdinov, R. Multimodal transformer for unaligned multimodal language sequences. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Florence, Italy, 27 July–2 August 2019; p. 6558. [Google Scholar]
  16. Cheng, J.; Fostiropoulos, I.; Boehm, B.; Soleymani, M. Multimodal phased transformer for sentiment analysis. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, Virtual, 7–11 November 2021; pp. 2447–2458. [Google Scholar]
  17. Rahman, W.; Hasan, M.K.; Lee, S.; Zadeh, A.; Mao, C.; Morency, L.-P.; Hoque, E. Integrating multimodal information in large pretrained transformers. In Proceedings of the Conference Association for Computational Linguistics Meeting, Seattle, WA, USA, 5–10 July 2020; p. 2359. [Google Scholar]
  18. Zadeh, A.; Chen, M.; Poria, S.; Cambria, E.; Morency, L.-P. Tensor fusion network for multimodal sentiment analysis. arXiv 2017, arXiv:1707.07250. [Google Scholar]
  19. Wang, Z.; Wan, Z.; Wan, X. Transmodality: An end2end fusion method with transformer for multimodal sentiment analysis. In Proceedings of the Web Conference 2020, Taipei, Taiwan, 20–24 April 2020; pp. 2514–2520. [Google Scholar]
  20. Peng, Y.; Qi, J. CM-GANs: Cross-modal generative adversarial networks for common representation learning. ACM Trans. Multimed. Comput. Commun. Appl. (TOMM) 2019, 15, 1–24. [Google Scholar] [CrossRef]
  21. Tsai, Y.-H.H.; Ma, M.Q.; Yang, M.; Salakhutdinov, R.; Morency, L.-P. Multimodal routing: Improving local and global interpretability of multimodal language analysis. In Proceedings of the Conference on Empirical Methods in Natural Language Processing, Online, 16–20 November 2020; p. 1823. [Google Scholar]
  22. Majumder, N.; Hazarika, D.; Gelbukh, A.; Cambria, E.; Poria, S. Multimodal sentiment analysis using hierarchical fusion with context modeling. Knowl.-Based Syst. 2018, 161, 124–133. [Google Scholar] [CrossRef]
  23. Georgiou, E.; Papaioannou, C.; Potamianos, A. Deep Hierarchical Fusion with Application in Sentiment Analysis. In Proceedings of the INTERSPEECH, Graz, Austria, 15–19 September 2019; pp. 1646–1650. [Google Scholar]
  24. Yan, L.; Cui, Y.; Chen, Y.; Liu, D. Hierarchical attention fusion for geo-localization. In Proceedings of the ICASSP 2021–2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Toronto, ON, Canada, 6–11 June 2021; pp. 2220–2224. [Google Scholar]
  25. Fu, Z.; Liu, F.; Xu, Q.; Qi, J.; Fu, X.; Zhou, A.; Li, Z. NHFNET: A non-homogeneous fusion network for multimodal sentiment analysis. In Proceedings of the 2022 IEEE International Conference on Multimedia and Expo (ICME), Taipei, Taiwan, 18–22 July 2022; pp. 1–6. [Google Scholar]
  26. Poria, S.; Cambria, E.; Bajpai, R.; Hussain, A. A review of affective computing: From unimodal analysis to multimodal fusion. Inf. Fusion 2017, 37, 98–125. [Google Scholar] [CrossRef]
  27. Colombo, C.; Del Bimbo, A.; Pala, P. Semantics in visual information retrieval. IEEE Multimed. 1999, 6, 38–53. [Google Scholar] [CrossRef]
  28. Wang, K.; Peng, X.; Yang, J.; Lu, S.; Qiao, Y. Suppressing uncertainties for large-scale facial expression recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 6897–6906. [Google Scholar]
  29. Bonifazi, G.; Cauteruccio, F.; Corradini, E.; Marchetti, M.; Sciarretta, L.; Ursino, D.; Virgili, L. A Space-Time Framework for Sentiment Scope Analysis in Social Media. Big Data Cogn. Comput. 2022, 6, 130. [Google Scholar] [CrossRef]
  30. Wöllmer, M.; Weninger, F.; Knaup, T.; Schuller, B.; Sun, C.; Sagae, K.; Morency, L.-P. Youtube movie reviews: Sentiment analysis in an audio-visual context. IEEE Intell. Syst. 2013, 28, 46–53. [Google Scholar] [CrossRef]
  31. Rozgić, V.; Ananthakrishnan, S.; Saleem, S.; Kumar, R.; Prasad, R. Ensemble of SVM Trees for Multimodal Emotion Recognition. In Proceedings of the 2012 Asia Pacific Signal and Information Processing Association Annual Summit and Conference, Hollywood, CA, USA, 3–6 December 2012; pp. 1–4. [Google Scholar]
  32. Metallinou, A.; Lee, S.; Narayanan, S. Audio-visual emotion recognition using gaussian mixture models for face and voice. In Proceedings of the 2008 Tenth IEEE International Symposium on Multimedia, Berkeley, CA, USA, 15–17 December 2008; pp. 250–257. [Google Scholar]
  33. EyEyben, F.; Wöllmer, M.; Graves, A.; Schuller, B.; Douglas-Cowie, E.; Cowie, R. On-line emotion recognition in a 3-D activation-valence-time continuum using acoustic and linguistic cues. J. Multimodal User Interfaces 2010, 3, 7–19. [Google Scholar] [CrossRef]
  34. Liu, Z.; Shen, Y.; Lakshminarasimhan, V.B.; Liang, P.P.; Zadeh, A.; Morency, L.P. Efficient low-rank multimodal fusion with modality-specific factors. arXiv 2018, arXiv:1806.00064. [Google Scholar]
  35. Poria, S.; Peng, H.; Hussain, A.; Howard, N.; Cambria, E.J.N. Ensemble application of convolutional neural networks and multiple kernel learning for multimodal sentiment analysis. Neurocomputing 2017, 261, 217–230. [Google Scholar] [CrossRef]
  36. Ghosal, D.; Akhtar, M.S.; Chauhan, D.; Poria, S.; Ekbal, A.; Bhattacharyya, P. Contextual inter-modal attention for multi-modal sentiment analysis. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, 31 October–4 November 2018; pp. 3454–3466. [Google Scholar]
  37. Zhang, Y.; Song, D.; Zhang, P.; Wang, P.; Li, J.; Li, X.; Wang, B. A quantum-inspired multimodal sentiment analysis framework. Theor. Comput. Sci. 2018, 752, 21–40. [Google Scholar] [CrossRef]
  38. Verma, S.; Wang, C.; Zhu, L.; Liu, W. Deepcu: Integrating both common and unique latent information for multimodal sentiment analysis. In Proceedings of the International Joint Conference on Artificial Intelligence, Macao, China, 10–16 August 2019; pp. 3627–3634. [Google Scholar]
  39. Pham, H.; Liang, P.P.; Manzini, T.; Morency, L.P.; Póczos, B. Found in Translation: Learning Robust Joint Representations by Cyclic Translations between Modalities. In Proceedings of the 33rd AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; pp. 6892–6899. [Google Scholar]
  40. Liang, P.P.; Liu, Z.; Tsai, Y.H.H.; Zhao, Q.; Salakhutdinov, R.; Morency, L.P. Learning Representations from Imperfect Time Series Data via Tensor Rank Regularization. arXiv 2019, arXiv:1907.01011. [Google Scholar]
  41. Mittal, T.; Bhattacharya, U.; Chandra, R.; Bera, A.; Manocha, D. M3ER: Multiplicative Multimodal Emotion Recognition using Facial, Textual, and Speech Cues. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 1359–1367. [Google Scholar]
  42. Xue, X.; Zhang, C.; Niu, Z.; Wu, X. Multi-level attention map network for multimodal sentiment analysis. IEEE Trans. Knowl. Data Eng. 2022, 35, 5105–5118. [Google Scholar] [CrossRef]
  43. Cauteruccio, F.; Stamile, C.; Terracina, G.; Ursino, D.; Sappey-Mariniery, D. An automated string-based approach to White Matter fiber-bundles clustering. In Proceedings of the 2015 International Joint Conference on Neural Networks (IJCNN), Killarney, Ireland, 12–17 July 2015; pp. 1–8. [Google Scholar]
  44. Wu, Y.; Zhao, Y.; Yang, H.; Chen, S.; Qin, B.; Cao, X.; Zhao, W. Sentiment Word Aware Multimodal Refinement for Multimodal Sentiment Analysis with ASR Errors. arXiv 2022, arXiv:2203.00257. [Google Scholar]
  45. Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv 2018, arXiv:1810.04805. [Google Scholar]
  46. Chung, J.; Gulcehre, C.; Cho, K.H.; Bengio, Y.J.E.A. Empirical evaluation of gated recurrent neural networks on sequence modeling. arXiv 2014, arXiv:1412.3555. [Google Scholar]
  47. Zadeh, A.; Zellers, R.; Pincus, E.; Morency, L.-P. Mosi: Multimodal corpus of sentiment intensity and subjectivity analysis in online opinion videos. arXiv 2016, arXiv:1606.06259. [Google Scholar]
  48. Degottex, G.; Kane, J.; Drugman, T.; Raitio, T.; Scherer, S. COVAREP—A collaborative voice analysis repository for speech technologies. In Proceedings of the 2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Florence, Italy, 4–9 May 2014; pp. 960–964. [Google Scholar]
  49. Ravanelli, M.; Parcollet, T.; Plantinga, P.; Rouhe, A.; Bengio, Y. SpeechBrain: A General-Purpose Speech Toolkit. arXiv 2021, arXiv:2106.04624. [Google Scholar]
  50. Panayotov, V.; Chen, G.; Povey, D.; Khudanpur, S. Librispeech: An asr corpus based on public domain audio books. In Proceedings of the 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), South Brisbane, QLD, Australia, 19–24 April 2015; pp. 5206–5210. [Google Scholar]
  51. Hazarika, D.; Zimmermann, R.; Poria, S. Misa: Modality-invariant and-specific representations for multimodal sentiment analysis. In Proceedings of the 28th ACM International Conference on Multimedia, Seattle, WA, USA, 10–16 October 2020; pp. 1122–1131. [Google Scholar]
  52. Yu, W.; Xu, H.; Yuan, Z.; Wu, J. Learning modality-specific representations with self-supervised multi-task learning for multimodal sentiment analysis. In Proceedings of the AAAI Conference on Artificial Intelligence, Online, 2–9 February 2021; Volume 35, pp. 10790–10797. [Google Scholar]
Figure 1. The architecture of our model. where * represents matrix multiplication; × represents line-by-line softmax function. Green represents textual, orange represents visual, and blue represents audio.
Figure 1. The architecture of our model. where * represents matrix multiplication; × represents line-by-line softmax function. Green represents textual, orange represents visual, and blue represents audio.
Electronics 12 03504 g001
Figure 2. Train-Has0-Acc, Train-Has0-F1, Val-Has0-F1, and Val-Has0-Acc on the MOSI-SpeechBrain dataset.
Figure 2. Train-Has0-Acc, Train-Has0-F1, Val-Has0-F1, and Val-Has0-Acc on the MOSI-SpeechBrain dataset.
Electronics 12 03504 g002
Figure 3. Train-Has0-Acc, Train-Has0-F1, Val-Has0-F1, and Val-Has0-Acc on the MOSI-IBM dataset.
Figure 3. Train-Has0-Acc, Train-Has0-F1, Val-Has0-F1, and Val-Has0-Acc on the MOSI-IBM dataset.
Electronics 12 03504 g003
Figure 4. Train-Has0-Acc, Train-Has0-F1, Val-Has0-F1, and Val-Has0-Acc on the MOSI-iFlytek dataset.
Figure 4. Train-Has0-Acc, Train-Has0-F1, Val-Has0-F1, and Val-Has0-Acc on the MOSI-iFlytek dataset.
Electronics 12 03504 g004
Table 1. The proportion of sentiment-word-replacement errors in three real-world datasets.
Table 1. The proportion of sentiment-word-replacement errors in three real-world datasets.
DatasetsPercentages
MOSI-SpeechBrain26.5%
MOSI-IBM17.6%
MOSI-iFlytek10.6%
Table 2. Results on the MOSI-SpeechBrain, MOSI-IBM, and MOSI-iFlytek datasets. The best results are in bold. ↑ indicates that the larger the numerical value, the better the model performance; ↓ indicates that the smaller the numerical value, the better the model performance.
Table 2. Results on the MOSI-SpeechBrain, MOSI-IBM, and MOSI-iFlytek datasets. The best results are in bold. ↑ indicates that the larger the numerical value, the better the model performance; ↓ indicates that the smaller the numerical value, the better the model performance.
DatasetsModelsEvaluation Metrics
Has0-Acc ↑Has0-F1 ↑Non0-Acc ↑Non0-F1 ↑MAE ↓Corr ↑
MOSI-SpeechBrainTFN68.9868.9569.5169.57115.5548.54
LMF68.8668.8869.3669.48117.4248.66
MulT71.7871.7072.7472.75109.0054.69
MISA73.7973.8574.7574.6698.5265.37
Self-MM73.6773.7274.8574.9890.9567.23
SWRM74.5874.6275.7075.8290.5667.47
Ours76.5376.4378.0578.0388.9868.27
MOSI-IBMTFN71.8171.7872.1373.21109.4258.19
LMF73.0673.0974.3074.41104.7059.07
MulT75.5775.5476.7476.79100.3264.34
MISA76.9776.9978.0878.1791.2371.30
Self-MM77.3277.3778.6078.7285.6573.23
SWRM78.4378.4779.7079.8082.9173.91
Ours80.1780.1581.8681.9079.6175.32
MOSI-iFlytekTFN71.9572.0172.6272.76107.0156.52
LMF71.9872.0372.3572.49106.6359.48
MulT77.3277.0578.7578.5689.8468.14
MISA79.5979.6279.8279.9185.6374.53
Self-MM80.2680.2681.1681.2078.7975.83
SWRM80.4780.4781.2881.3478.3975.97
Ours81.9281.9382.9382.9875.2576.31
Table 3. Experimental results of modal ablation on the MOSI-SpeedBrain dataset. ↑ indicates that the larger the numerical value, the better the model performance; ↓ indicates that the smaller the numerical value, the better the model performance.
Table 3. Experimental results of modal ablation on the MOSI-SpeedBrain dataset. ↑ indicates that the larger the numerical value, the better the model performance; ↓ indicates that the smaller the numerical value, the better the model performance.
TasksHas0-Acc ↑Non0-Acc ↑MAE ↓
T73.3275.0991.22
A55.6956.10137.88
V55.2553.66142.20
V + T73.9175.3089.03
T + A73.8274.7090.41
V + A56.0456.89136.34
T + V + A75.0776.3791.10
TVA + T + V + A76.5378.0588.89
Table 4. Experimental results of the model ablation. ↑ indicates that the larger the numerical value, the better the model performance; ↓ indicates that the smaller the numerical value, the better the model performance.
Table 4. Experimental results of the model ablation. ↑ indicates that the larger the numerical value, the better the model performance; ↓ indicates that the smaller the numerical value, the better the model performance.
DatasetsModelsHas0-Acc ↑Non0-Acc ↑MAE ↓
MOSI-SpeechBrainMWRCMH76.5378.0588.98
w/o Fusion75.6677.2989.76
w/o MWRM75.3776.6289.69
MOSI-IBMMWRCMH80.1781.8679.61
w/o Fusion79.7481.1180.75
w/o MWRM79.6181.0181.02
MOSI-iFlytekMWRCMH81.9282.9375.25
w/o Fusion80.6181.8676.16
w/o MWRM80.6681.2576.81
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Huang, J.; Lu, P.; Sun, S.; Wang, F. Multimodal Sentiment Analysis in Realistic Environments Based on Cross-Modal Hierarchical Fusion Network. Electronics 2023, 12, 3504. https://doi.org/10.3390/electronics12163504

AMA Style

Huang J, Lu P, Sun S, Wang F. Multimodal Sentiment Analysis in Realistic Environments Based on Cross-Modal Hierarchical Fusion Network. Electronics. 2023; 12(16):3504. https://doi.org/10.3390/electronics12163504

Chicago/Turabian Style

Huang, Ju, Pengtao Lu, Shuifa Sun, and Fangyi Wang. 2023. "Multimodal Sentiment Analysis in Realistic Environments Based on Cross-Modal Hierarchical Fusion Network" Electronics 12, no. 16: 3504. https://doi.org/10.3390/electronics12163504

APA Style

Huang, J., Lu, P., Sun, S., & Wang, F. (2023). Multimodal Sentiment Analysis in Realistic Environments Based on Cross-Modal Hierarchical Fusion Network. Electronics, 12(16), 3504. https://doi.org/10.3390/electronics12163504

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop