Next Article in Journal
Graph U-Shaped Network with Mapping-Aware Local Enhancement for Single-Frame 3D Human Pose Estimation
Next Article in Special Issue
A Dynamic Emotional Propagation Model over Time for Competitive Environments
Previous Article in Journal
Blueprint Separable Subsampling and Aggregate Feature Conformer-Based End-to-End Neural Diarization
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Learning Short Text Sentiment Analysis Based on Improved Particle Swarm Optimization

1
School of Computer and Information Engineering, JiangXi Normal University, Nanchang 330224, China
2
School of Digital Industry, JiangXi Normal University, Shangrao 334000, China
*
Authors to whom correspondence should be addressed.
Electronics 2023, 12(19), 4119; https://doi.org/10.3390/electronics12194119
Submission received: 20 August 2023 / Revised: 23 September 2023 / Accepted: 28 September 2023 / Published: 2 October 2023
(This article belongs to the Special Issue New Advances in Affective Computing)

Abstract

:
Manually tuning the hyperparameters of a deep learning model is not only a time-consuming and labor-intensive process, but it can also easily lead to issues like overfitting or underfitting, hindering the model’s full convergence. To address this challenge, we present a BiLSTM-TCSA model (BiLSTM combine TextCNN and Self-Attention) for deep learning-based sentiment analysis of short texts, utilizing an improved particle swarm optimization (IPSO). This approach mimics the global random search behavior observed in bird foraging, allowing for adaptive optimization of model hyperparameters. In this methodology, an initial step involves employing a Generative Adversarial Network (GAN) mechanism to generate a substantial corpus of perturbed text, augmenting the model’s resilience to disturbances. Subsequently, global semantic insights are extracted through Bidirectional Long Short Term Memory networks (BiLSTM) processing. Leveraging Convolutional Neural Networks for Text (TextCNN) with diverse convolution kernel sizes enables the extraction of localized features, which are then concatenated to construct multi-scale feature vectors. Concluding the process, feature vector refinement and the classification task are accomplished through the integration of Self-Attention and Softmax layers. Empirical results underscore the effectiveness of the proposed approach in sentiment analysis tasks involving succinct texts containing limited information. Across four distinct datasets, our method attains impressive accuracy rates of 91.38%, 91.74%, 85.49%, and 94.59%, respectively. This performance constitutes a notable advancement when compared against conventional deep learning models and baseline approaches.

1. Introduction

Text sentiment analysis typically involves the exploration and examination of the emotional inclination present within textual commentaries. This encompasses the extraction of subjectively imbued elements like viewpoints, stances, attitudes, and opinions, subsequently entailing the assessment of the emotional underpinnings of said elements. The proliferation of the internet and the rapid evolution of social media platforms and online comment forums have propelled text sentiment analysis into a realm of scholarly inquiry that commands considerable focus in recent years.
Currently, proposed methods for text sentiment analysis can be broadly categorized into two main groups. The first category entails machine learning approaches, encompassing techniques like support vector machines [1], K-nearest neighbor classification algorithms [2], naive Bayes [3], and Latent Dirchlet allocation [4]. While these algorithms exhibit commendable performance, they predominantly hinge upon manually extracting text features, leading to inefficiencies throughout the entire process. The second approach centers around deep learning, a prevailing paradigm within the realm of natural language processing. Prominent examples encompass the Convolutional Neural Networks (CNN) and the Recurrent Neural Networks (RNN). The former, CNN, adeptly discerns local features from textual matrices through convolutional and pooling operations. Notably, Kim [5] introduced a CNN-based sentence classification model that yields promising classification results across multiple public datasets. However, it struggles to effectively to grasp global context features. Recognizing the shortcomings in obtaining comprehensive textual semantics and contextual insights, sources in the literature [6,7,8] increase the number of layers to increase the receptive field of information obtained by the convolutional kernel, but the effect is not significant. For this reason, many researchers began to change the structure of the convolution layer and the pooling layer from the horizontal. Guo [9] ingeniously constructed an augmented CNN model, harnessing skip convolution and K-max pooling operations, thereby refining the extraction of features from concise texts. Additionally, Wang [10] fortified the TextCNN architecture through two distinct avenues: the creation of N-gram discontinuous sliding windows and K-Max average pooling. Despite these advancements, owing to the sparse nature of textual content and the inherent limitations in capturing temporal dimensions, even these enhanced models grapple with inadequacies in fully acquiring the nuances of text semantics.
RNN possesses the capability to extract global information to a certain extent. Irsoy [11] harnessed RNN for extracting sequence features in sentiment analysis tasks. Similarly, Xin [12] employed a specialized RNN structure to capture text features for subsequent analysis. However, due to its inherent limitations in “memory” capacity, RNN encounters issues such as gradient explosion or vanishing gradients when text length exceeds a certain range, thereby impacting experimental outcomes. In response to this challenge, Hochreiter and Gers [13,14] introduced the LSTM network, and in subsequent advancements, Cho [15] amalgamated input and forget gates into update gates, culminating in the development of Gated Recurrent Units (GRUs). This not only effectively resolves the gradient vanishing or explosion problem but also refines the internal processing units, enabling more adept context information capture. Consequently, GRUs outperform RNNs on numerous tasks. Tai [16] introduced textual topological structure information into LSTM models, constructing LSTM models based on both dependency trees and phrase structure trees. This novel approach modifies the network structure to make gate vectors and internal state updates dependent on states from multiple relevant subunits, effectively amalgamating syntactic features such as dependency relationships and phrase composition in short texts. This results in more accurate semantic representations for short texts. Zhao’s [17] proposed sequential hybrid model initially employs BiLSTM to capture the global features of the text and subsequently utilizes CNN to extract local semantic features and perform classification. This approach mitigates the loss of positional information and contributes to enhanced results.
In recent years, attention mechanisms [18] have been introduced to assign significant attention weights to vectors, highlighting key information. The integration of attention mechanisms enhances feature expression, resulting in models with higher accuracy that intuitively emphasize the importance of different words. Yang [19] combined a hierarchical attention network with Bidirectional Gated Recurrent Units (Bi-GRU) for English short text classification. This approach segments text into three hierarchical levels—words, sentences, and documents—introducing an attention mechanism between each level to assign varying weights to different words and sentences, progressively selecting crucial information. Building upon Yang’s work, Zhou [20] proposed a hybrid attention network that combines both word-level and character-level information for classifying Chinese short texts. By merging the attention mechanism with BiLSTM, this model captures both word and character embeddings, thus extracting the most pivotal word and character information. Cheng Yan’s model [21] integrates a multi-channel CNN with Bi-GRU. This combination exploits the complementarity of local features and contextual information, with the incorporation of an attention mechanism further identifying pivotal information within the text.
While the aforementioned methods have yielded promising results, the configuration of hyperparameters significantly influences the performance of deep learning predictive models. Relying on manual parameter tuning based on empirical knowledge introduces considerable errors, thereby limiting the model’s full potential. To address this issue, Feurer [22] introduced an advanced interface capable of automatically assessing multiple preprocessing and model fitting pipelines. They delved into the concept of achieving automated machine learning (AutoML) through meta-learning, with the aim of streamlining the tedious manual tuning process involved in hyperparameter optimization for deep learning models. Wang [23] proposed an efficient and robust method for automated hyperparameter tuning using Gaussian processes. This approach effectively handles various hyperparameter combinations and employs Gaussian process regression to predict model performance. However, Jeremy [24] breaks away from tradition through a method known as Automatic Gradient Descent (AGD), which allows for training deep learning models without the need for hyperparameters.
This paper presents an innovative approach that integrates an IPSO into a deep learning-based sentiment analysis model for text. Leveraging BiLSTM, TextCNN, and the Attention Mechanism, this model adeptly extracts profound semantic information from text pairs. Furthermore, it introduces a Generative Adversarial Network (GAN) mechanism to generate perturbed texts for training and combines with the IPSO to optimize and match the model’s hyperparameters effectively. Experimental results further underscore the efficacy and feasibility of this approach.

2. Related Works

2.1. IPSO

In general, deep learning models often have numerous hyperparameters, and their search space can be extensive. When employing manual search methods or grid search techniques, it can become impractical due to the presence of continuous hyperparameter variables. This not only makes the search process cumbersome but also incurs high computational time costs. Assuming that a three-dimensional function ‘ Z = g ( x , y ) ’ needs to find its optimal integer solution, when the search range of ‘ x ’ is denoted as ‘ m ’ and the search range of ‘ y ’ is denoted as ‘ n ’, as the search ranges ‘ m ’ and ‘ n ’ continue to expand, the scope of grid search grows quadratically, the horizontal and vertical axes represent the expansion of the search domains, ‘ m ’ and ‘ n ’, while the vertical axis represents the number of parameter tuning iterations, as illustrated in Figure 1. In this, areas with a redder color indicate a higher number of required iterations.
To address the challenge of manually exhaustively searching for optimization solutions, various specialized algorithms have been proposed. These include the Pathfinder Algorithm (PA) [25], Multi-Verse Optimizer (MVO) [26], Moth Flame Optimizer (MFO) [27], and Arithmetic Optimization Algorithm (AOA) [28].
The PA algorithm draws inspiration from the behavior of ants searching for the shortest path. It employs communication based on pheromones to guide the search for the optimal path in the problem space. However, its applicability is limited when it comes to problems unrelated to pathfinding or routing.
The MVO algorithm is based on the concept of multiple universes, where each universe represents a set of solutions. These universes interact, share information, and compete to find the best solutions, making it suitable for multimodal optimization problems. However, when a large number of populations are set, the computational resources required increase significantly, which may not be acceptable.
The MFO algorithm is inspired by the behavior of moths being attracted to flames. Moths represent solutions, and they are attracted to better solutions while avoiding the “flames” or undesirable regions. This algorithm iteratively updates solutions to find the best or near-optimal solution. It is easy to implement and applicable to various optimization problems, but it can easily become stuck in local optima, especially in complex high-dimensional spaces.
The AOA algorithm is a general optimization framework based on arithmetic operations such as addition, subtraction, multiplication, and division. It iteratively improves solutions based on the solutions themselves and has a wide range of applications. However, for complex problems, it may require a substantial amount of computational resources.
In comparison to those algorithms mentioned above, the Particle Swarm Optimization (PSO) algorithm boasts several advantages. It exhibits minimal dependence on initial information, avoids convergence into local minima due to high computational complexity, and is relatively straightforward to comprehend. Originally introduced by Kennedy and Eberhart [29], PSO is a global random search algorithm inspired by the foraging behavior of birds. In complex constrained problems, particles represent optimal solutions under multiple optimization constraints, and a particle swarm constitutes a collection of these optimal solutions. The dimensionality of particles is determined by the number of parameters to be optimized, while the number of particles in the swarm can be adjusted based on factors such as the quantity of optimization problems, required computational resources, and optimization objectives.
The algorithm sets the position state ‘ X i ’ and velocity state ‘ V i ’ to represent the current state of the particle ‘ i ’. Throughout the iterative process, particles continually update their positions and velocities based on the interplay between individual and global best values of the current state. Upon locating the present global and individual optimal solutions, particle information is iteratively updated according to the following Equations (1)–(4) [30,31]:
X i = ( x i 1 , x i 2 , x i 3 , , x i n )
V i = ( v i 1 , v i 2 , v i 3 , , v i n )
V i d t + 1 = w v i d t + c 1 r a n d 1 ( p b e s t i d t x i d t ) + c 2 r a n d 2 ( g b e s t d t x i d t )
x i d t + 1 = x i d t + v i d t + 1
Among them, ‘ X i ’ and ‘ V i ’ represent the collection of positions and velocities of particles in each dimension of the current n -dimensional space. The most crucial step in the particle swarm optimization algorithm lies in Equation (3), which is used to update the velocity information of the current particle in a specific dimension. Here, ‘ w ’ denotes the weight of the current velocity, ‘ c 1 ’ and ‘ c 2 ’ represent learning factors, ‘ r a n d 1 ’ and ‘ r a n d 2 ’ are random numbers between (0, 1), and ‘ p b e s t ’ and ‘ g b e s t ’ stand for the individual best value found by the current particle and the global best value found by all particles, respectively. Therefore, in Equation (3), the velocity increment information ‘ V t + 1 ’ depends not only on the current position ‘ V t ’ but also on the individual best ‘ p b e s t ’ of the particle and the global best solution ‘ g b e s t ’ found by all particles. Finally, the current position information of the particle is updated using Equation (4).
To validate the superior performance of the PSO algorithm, this paper compares the PSO algorithm with other algorithms such as PA, MVO, MFA, and AOA using the CEC benchmark test function set. The mathematical formulas for the CEC standard benchmark functions are presented in Table 1, and the results of various algorithms on the CEC standard benchmark functions are presented in Table 2.
While PSO exhibits notable performance in optimization compared to other algorithms, this paper aims to further enhance the performance of the PSO. Specifically, we propose improvements to the inertia ‘ w ’ and learning factors ‘ c 1 ’ and ‘ c 2 ’ within the PSO.

2.1.1. Improvement of the Inertia Weight ‘w

The inertia weight refers to the extent to which the current velocity is influenced by the previous generation’s velocity. It is directly proportional to the global search capability and inversely proportional to the local search capability. During the early iterations, the particle’s optimal state lies in possessing a strong global search capability, enabling it to explore new regions and discover potential possibilities. However, as the iterations progress, there arises a need to enhance the particle’s local search capability to ensure smooth convergence. Therefore, a novel approach is introduced to dynamically adjust the inertia weight ‘ w ’ using a nonlinear function. The formula is as follows in Equation (5):
w = w max ( w max w min ) × 4 π × arctan t t max
The motivation for using ‘ 4 π × a r c t a n t t m a x ’ is twofold: it exhibits the characteristics of a convex function and can effectively adapt to the iterative process of the weight ‘ w ’ operation. In the early stages of iteration, although individual diversity is required to explore unknown domains, the portion controlled by the inertia weight does not dominate. Therefore, it is essential to apply a certain degree of decay rate to the ‘ w ’ in the early stages, maintaining a relatively high level of acceleration in the ‘ w ’ decay. However, as the iteration progresses, in order to expedite particle convergence to the global optimum, it becomes necessary to smoothly reduce the ‘ w ’ to its minimum value. Consequently, in the later stages, the acceleration of w decay should be maintained at a lower level. Assuming ‘ w m a x = 0.8 ’ and ‘ w m i n = 0.2 ’, the variation of the ‘ w ’ decay acceleration during the iteration process is illustrated in Figure 2.

2.1.2. Improvement of Learning Factors

The learning factors, ‘ c 1 ’ and ‘ c 2 ’, respectively determine the extent to which particles are influenced by group information and individual information. When ‘ c 1 ’ is set to 0, the particle’s state is entirely governed by global information, resulting in reduced diversity within the group and making it difficult for particles to escape local optima. Conversely, when ‘ c 2 ’ is set to 0, the particle’s state solely relies on individual information, displaying a self-centric behavior that disregards the contributions from the group. This can lead to a slower convergence rate of the algorithm. Hence, a nonlinear function is introduced to dynamically adjust the learning factors, following Equations (6) and (7):
c 1 ( t ) = ( c 1 e c 1 s ) ( t / t max ) 3 + c 1 s
c 2 ( t ) = ( c 2 e c 2 s ) ( t / t max ) 3 + c 2 s
where ‘ c 1 s ’ and ‘ c 1 e ’ represent the initial and final values of parameter ‘ c 1 ’, with ‘ c 1 s = 1 ’ being greater than ‘ c 1 e = 0 ’, and ‘ c 2 s ’ and ‘ c 2 e ’ denoting the initial and final values of parameter ‘ c 2 ’ with ‘ c 2 s = 0 ’ being less than ‘ c 2 e = 1 ’. From the perspective of ‘ c 1 ’ during the iteration process, it is necessary to slowly decay ‘ c 1 ’ using the concave function characteristic of ‘ ( t t m a x ) 3 ’. This is because, in the early stages of iterations, particles need to possess sufficient individual skills to explore unknown territories. Therefore, the rate of ‘ c 1 ’ decay in the early stages should be slow, and the gradient should decrease. However, in the later stages, when particles no longer need to explore unknown areas due to their independent characteristics being suppressed or even cleared, the acceleration of ‘ c 1 ’ decay should increase. During iterations, the variation of ‘ c 1 ’ is illustrated in Figure 3a. On the contrary, for ‘ c 2 ’, in the early stages of iteration, it is necessary to suppress the global characteristics of particles, allowing them to move with their “self-will”. Therefore, only a slight increase in ‘ c 2 ’ is required in the early stages. However, as the iterations progress to the later stages, the global optimum requires a significant “attraction” to particles. Thus, ‘ c 2 ’ should undergo a substantial increment. The concave function characteristic of ‘ ( t t m a x ) 3 ’ can meet the requirements for ‘ c 2 ’’s incremental changes during the iteration. Based on the description of ‘ c 2 ’, the variation of ‘ c 2 ’ will be as shown in Figure 3b.
The specific computation process of the IPSO algorithm is illustrated in Figure 4. The algorithm consists of the following six specific steps:
(1)
Randomly generate an initial batch of particles (along with initialized particle velocities and positions) to constitute the current population.
(2)
Calculate the fitness for each particle.
(3)
Update the individual best and global best solutions for each particle based on their fitness function values. The updated equations are listed as Equation (8). Here, ‘ p i ’ represents the fitness value of the particle at the current iteration ‘ i ’, ‘ g i ’ represents the global best solution at iteration ‘ i ’, and ‘ n ’ is the maximum number of iterations.
p b e s t = min ( p 1 , p 2 , p 3 , , p n )
g b e s t = min ( g 1 , g 2 , g 3 , , g n )
(4)
Updating particle velocities and positions using Equations (3) and (4).
(5)
Calculate the fitness for each particle again.
(6)
Check if the iteration is complete. If it is, output the current best solution. Otherwise, proceed to (3).

2.1.3. Performance Analysis of the IPSO

To evaluate the performance of the IPSO, this study utilizes three CEC standard benchmark functions to demonstrate the algorithm’s improvements, namely, ‘ h o l d e r _ t a b l e ’, ‘ e a s o m ’, and ‘ r a s t r i g i n ’, with their expressions shown in Equations (10)–(12). Since it involves finding the global optimum of multivariate functions, the fitness function in the IPSO is the test function itself.
h o l d e r _ t a b l e ( x ) = | i = 1 D sin ( x i ) cos ( x i ) × exp ( 1 i = 1 D x i 2 π ) |
e a s o m ( x ) = i = 1 D [ cos ( x i ) ] × exp [ i = 1 D ( x i π ) 2 ]
r a s t r i g i n ( x ) = 10 n + i = 1 D [ x i 2 10 cos ( 2 π x i ) ]
Figure 5 illustrates the three-dimensional plot of the test function ‘ h o l d e r _ t a b l e ’, ’ e a s o m ’, ’ r a s t r i g i n ’.
During testing, a particle swarm size of 20 and particle dimensionality of 2 were configured, with a maximum iteration count of 200. The parameters were set as w m a x = 0.8 and w m i n = 0.4 . A comparative analysis was conducted on the evolution of fitness curves between the pre-improvement and post-improvement methods, as depicted in Figure 6.
From the comparison in the above figures, it can be observed that the performance of PSO is not optimal, and it exhibits certain shortcomings compared to IPSO. Comparing Figure 6a with Figure 6b, as well as Figure 6c with Figure 6d, it can be seen that in the early iterations, PSO lacks sufficient individual diversity, leading to insufficient capability to escape local optima and often becoming trapped in local optima briefly and frequently. By examining Figure 6e,f, it becomes evident that the global characteristics of particles in PSO do not transition with iterations, resulting in unstable update speeds. In contrast, particles in IPSO exhibit significant individual diversity in the early stages, allowing them to thoroughly explore the unknown space and converge to the global optimum solution quickly in the initial iterations. Furthermore, as the iterations progress, the particles in IPSO quickly stabilize.

2.2. BiLSTM

In the LSTM network, although the model circumvents the issues of vanishing or exploding gradients seen in RNN, its unidirectional nature restricts it to utilizing information from the ‘preamble’, leaving it unaware of the ‘postamble’. In practical scenarios, predictive outcomes often necessitate access to information from the entire input sequence. Consequently, the prevalent approach involves the application of BiLSTM [32]. Building upon the foundation of LSTM, BiLSTM integrates input sequence information from both the forward and backward directions. For the output at time ‘ t ’, the forward LSTM layer incorporates information up to and including time ‘ t ’, while the backward LSTM layer incorporates information from ‘ t ’ onwards in the input sequence. The outputs from both forward and backward layers are then averaged to obtain the BiLSTM layer’s output. The architecture of the BiLSTM network is illustrated in Figure 7.
The gating mechanisms within the hidden layers of the LSTM structure determine the transmission of information and have the capacity to learn critical dependencies for the current information. The forget gate governs the discarding of information that is deemed unimportant for classification, the input gate determines which information needs to be updated, and the output gate determines the information to be output. Let us denote the values of the forget gate, input gate, and output gate at time ‘ t ’ as ‘ f t ’, ‘ i t ’, and ‘ o t ’, respectively. Specifically, the activation function ‘ δ ’ for the forget gate utilizes the sigmoid function. The information update is represented as follows in Equation (13):
f t = δ ( W f × X + b )
Information update for the input gate is computed using Equation (14):
i t = δ ( W i × X + b i )
Computation for updating the output gate’s information is given by Equation (15):
O t = δ ( W o × X + b )
Calculation for updating the cell information state is represented by Equation (16):
c t = f t c t 1 + i t tanh ( W c × X + b c )
The update of the hidden state information at time ‘ t ’ is determined by Equation (17):
h ( t ) = O t tanh ( c t )
where ‘ h ( t ) ’ is composed of the forward ‘ h ( t ) ’ and backward ‘ h ( t ) ’ states, and ‘ W f ’, ‘ W i ’, ‘ W o ’, and ‘ W c ’ represent the weight matrices for the forget gate, input gate, output gate, and updated cell state, respectively. Correspondingly, ‘ b f ’, ‘ b i ’, ‘ b o ’, and ‘ b c ’ are the respective bias terms. The LSTM cell structure is illustrated in Figure 8.

2.3. TextCNN

The TextCNN model proposed by Kim [5] consists of one-dimensional convolutional layers and one-dimensional max-pooling layers. Its input is a text matrix composed of word vectors, and the width of the convolutional kernels is the same as the width of the text matrix. In other words, the width of the convolutional kernels is equal to the dimension of the word vectors. The convolutional kernels only move vertically. The model structure is illustrated in Figure 9.
After convolving a text vector with a kernel, the resulting output is a vector with a shape of (sentence length − kernel height + 1, 1). The calculation of the convolution operation is shown in Equation (18).
c = f ( w i × M + b )
In the equation, ‘ c ’ represents the feature vector obtained from the convolutional kernel output, ‘ f ’ is the activation function of the convolutional layer, ‘ b ’ stands for the bias term, ‘ w i ’ denotes convolutional kernels of different sizes, and ‘ M ’ represents the text matrix.
Due to the varying sizes assigned to the convolutional kernels ‘ w i ’, they can extract text features at different scales. Therefore, the text features ‘ k ’ preserved after passing through the pooling layer represent the multi-scale characteristics of the text. The calculation of the pooling operation is demonstrated in Equation (19).
k = max ( c 1 , c 2 , , c n )
Subsequently, the obtained features ‘k’ from each operation are concatenated to form a text feature vector, which serves as the input for the subsequent stage of operations.

2.4. Self-Attention

The design of the Self-Attention [18] mechanism is inspired by human visual processing, where a preview of the global information is taken to gauge the attention towards local details. This approach allows for a better focus on key information while selectively suppressing irrelevant data. The structure of the mechanism is depicted in Figure 10.
Self-Attention is fundamentally a mapping relationship composed of ‘ Q u e r y ’, ‘ K e y ’, and ‘ V a l u e ’ components for each local information. Its primary computation process can be summarized in the following steps:
Step 1: Compute the similarity between the ‘ Q u e r y ’ of a specific local information and each ‘ K e y ’ to obtain corresponding weights, which can be calculated using Equation (20).
s i m i l a r ( Q , K ) = Q K T
Step 2: Normalize the weight values through ‘ S o f t m a x ’ function to prevent occurrences of extremely large or small values. This normalization process can be computed according to Equation (21).
a i = s o f t m a x ( s i m i l a r ( Q , K ) )
Step 3: Compute the final ‘ A t t e n t i o n ’ by performing a weighted sum of ‘ a i ’ with the corresponding ‘ V a l u e ’ components, which can be calculated using Equation (22).
S e l f A t t e n t i o n ( Q , K , V ) = a i V

3. Methodology

3.1. Noise Injection and Adversarial Training

Given the limited semantic information in short texts coupled with their high dimensionality, overfitting can become a challenge. In the context of deep learning models, various approaches are typically employed to mitigate overfitting and enhance generalization ability:
  • Utilizing dropout: This involves randomly deactivating different weights during training.
  • Leveraging adversarial networks for noise injection: Adversarial networks are used to generate noisy data, introducing necessary perturbations.
These strategies collectively aim to prevent overfitting and enhance the model’s ability to generalize.
The principle of dropout involves randomly deactivating a subset of parameters during each training iteration. In each training cycle, a random selection of parameters is chosen to be hidden, with their values set to zero. Since different parameters are hidden in each training iteration, it is akin to multiple instances of different sub-networks being combined and stacked. This results in diverse computational outcomes being compounded and outputted, leading to a richer representation of information features. In comparison to the original network, the network at this stage is not only simplified but also captures more stochastic factors, consequently reducing the occurrence of overfitting.
The GAN [33,34] is a generative modeling technique based on differentiable generative networks, rooted in the principles of game theory. Miyato [35] introduced adversarial training and virtual training to semi-supervised text classification tasks, effectively mitigating the occurrence of overfitting. Chen [36] combined the attention mechanism with a multitasking-focused GAN network. Under the attention mechanism, a subset of the original text is allocated for adversarial training. Within the GAN architecture, the generator and discriminator compete with each other to achieve a “Nash equilibrium”, as illustrated in Figure 11.
The generator captures the underlying distribution of the training data, aiming to generate synthetic data that conforms to the same distribution. It expects the discriminator to classify the generated data as real training data. The discriminator, guided by the standards set by the training data, encourages the generator to produce more “realistic” synthetic data. Through the adversarial interplay and continuous self-improvement between these two networks, the objective is to generate training data that closely resembles real data. The loss function of the adversarial network is defined by Equation (23):
min G max D V ( D , G ) = E X P d a t a ( x ) [ log D ( x ) ] + E X P g ( x ) [ log ( 1 D ( G ( z ) ) ]
In Equation (21): ‘ x ’ represents the real samples input into the network; ‘ E ’ represents the expectation of inputting real samples; ‘ D ( ) ’ signifies the output values of the discriminator’s judgment; ‘ G ( ) ’ signifies the output values of the generator.
To address the limitations of having a single perturbation parameter, fixed normalization, and an inability to accurately control the noise magnitude within the model, a combination of adversarial training network and dropout is employed. This approach generates a substantial number of noisy samples, enhancing both model performance and robustness, thereby mitigating the risk of overfitting.

3.2. Model Architecture

To emphasize the essential information for sentiment analysis within short texts, this paper introduces a novel text sentiment analysis BiLSTM-TCSA model for binary classification of sentences, which integrates word embedding techniques with BiLSTM, TextCNN, and Self-Attention. The proposed model utilizes Glove Vectors for word Representation (GloVe) [37] pre-trained word embeddings derived from a vast corpus to represent text vectors, and further fine-tunes these embeddings during the entire training process in conjunction with the dataset.
Prior to training, a certain quantity of perturbed texts is generated using a GAN and incorporated into the training dataset. This approach not only enhances the robustness of the model but also effectively addresses overfitting issues. In the learning of contextual dependency, a dual-layer BiLSTM mechanism that incorporates the Resnet concept is employed to fuse initially independent word embeddings. This mechanism establishes connections between contextual semantic information, yielding comprehensive global semantic features. For the learning of local semantic relationships, a TextCNN with convolutional kernels of sizes (3, 4, 5) is employed. The convolutional operations are followed by pooling, resulting in local semantic features that are concatenated. The Self-Attention mechanism is then employed to enhance the sensitivity of sentiment word features while suppressing non-keyword features. This enhances the model’s focus on text information and improves classification efficiency.
Regarding the overall model architecture, during the training process, the IPSO is employed to iteratively search for the optimal combination of model hyperparameters. This ensures that the model achieves maximum convergence. The holistic structure of the model is illustrated in Figure 12.

4. Experiments and Analysis

4.1. Experimental Environment

The experiments were conducted using the open-source deep learning framework Keras (Version-2.4.3), implemented using the Python 3.8 programming language on the PyCharm platform. The experimental setup was as follows: the operating system was Windows 10, the CPU was AMD Ryzen 7 5800H, and the GPU was NVIDIA GeForce GTX 1650. A summary of the experimental environment information is presented in Table 3.

4.2. Datasets

To validate the rationality and effectiveness of the BiLSTM-TCSA model, four widely used public datasets were selected: the IMDB movie review dataset, the MR English movie review dataset, the SST-2 dataset from the Stanford Sentiment Treebank, and the Hotel Reviews dataset containing user reviews from renowned hotels worldwide.
(1)
The IMDB: This movie review dataset comprises user reviews from the IMDB website about movies. For the experiment, all 50,000 records from this dataset were obtained.
(2)
The MR: This dataset is a binary-class English movie review dataset, containing an equal number of positive and negative sentiment samples, totaling 10,662 instances.
(3)
The SST-2: This dataset originates from the Stanford Sentiment Treebank and is designed specifically for binary sentiment classification tasks, comprising a total of 70,044 sentiment-labeled comments.
(4)
The Hotel Reviews: This dataset, hereinafter referred to as the HR dataset, consists of user-generated polarity sentiment reviews about renowned hotels worldwide, collected by the Whale Community. After eliminating irrelevant comments, this dataset comprises a total of 877,640 valuable entries.
However, to enhance the robustness of the model, several sets of perturbed texts were extracted from the dataset and input into the GAN to generate perturbed texts. The number of generated perturbed texts accounted for 1/10 of the dataset’s sample size. Subsequently, these perturbed texts were combined with the original dataset. Following the principle of maintaining a balanced proportion of positive and negative samples, a training set to testing set ratio of 7:3 was established, and this combined dataset was used for model training. The relevant statistical information for each dataset is presented in Table 4.

4.3. Metrics

In the sentiment analysis experiments, this study employs accuracy, recall, and F1 score as evaluation metrics for the models. The specific calculation formulas are as follows in Equations (24)–(26):
A c c u r a c y = T P + T N T P + T N + F P + F N
R e c a l l = T P T P + F P
F 1 _ S c o r e = 2 T P 2 T P + F P + F N
where TP (True Positive) represents the number of samples correctly classified as positive sentiment, FP (False Positive) represents the number of samples incorrectly classified as positive sentiment, FN (False Negative) represents the number of samples incorrectly classified as negative sentiment, and TN (True Negative) represents the number of samples correctly classified as negative sentiment.

4.4. Hyperparameter Optimization Based on IPSO

The parameters for the IPSO were initially set with a particle count of m = 15 . The dimension of the optimization hyperparameters was defined as d = 5 , and the experiment was conducted over a total of k = 100 iterations.
The optimization of hyperparameters for the network model is elaborated upon below. In this study, the network model is constructed within the Keras framework. It employs a two-layer BiLSTM and utilizes the Cross Entropy Loss function as its loss function. The optimizer of choice is Stochastic Gradient Descent (SGD). The hyperparameters selected for optimization encompass the dropout rate, BiLSTM hidden layer dimension (hd), learning rate, batch size, and convolutional kernel stride.
(1)
Dropout rate, denoted as ‘ d p ’, representing the random deactivation rate of the dropout layer in the generators of GAN, with a range of d p ( 0 , 1 ) ;
(2)
BiLSTM hidden layer dimensions, denoted as ‘ h d ’, indicating the dimensionality of the output vectors from the BiLSTM layer, with a range of h d [ 50 , 200 ] ;
(3)
Learning rate, denoted as ‘ l r ’, which controls the magnitude of updates to the internal parameters during model training, with l r [ 0.0001 , 0.01 ] ;
(4)
Batch size, indicating the number of samples used in each training iteration, with b a t c h _ s i z e { 16 , 32 , 64 , 128 } ;
(5)
Convolutional kernel strides, denoted as ‘ s t r i d e s ’, representing the length of each movement of the convolutional kernel in the TextCNN layer, with s t r i d e s { 1 , 2 , 3 , 4 } ;
In the process of fine-tuning deep learning models using IPSO, it is necessary to abstract the model as a ‘ m o d e l ’ function. The fitness function for this deep learning model typically adopts the loss function used by the model. Here, the fitness function is set to the ‘ c r o s s _ e n t e r p o r y ’. The operational procedure is illustrated in Equations (27) and (28):
c r o s s _ e n t e r p o r y = { m o d e l 1 s a m p l e _ l a b e l = 1 m o d e l 0 s a m p l e _ l a b e l = 0
m o d e l = m o d e l ( B i L S T M , G A N , T e x t C N N , L e a r n i n g _ r a t e , b a t c h _ s i z e )
Upon observation, it is apparent that, although the model includes a self-attention mechanism not explicitly listed in the ‘ m o d e l ’ function, the parameters within this mechanism adapt through backpropagation during iteration. In contrast, BiLSTM, GAN, and TextCNN each have their own set of hyperparameters that require optimization. During the iterative process, the designated learning rate and batch size are also chosen as hyperparameters for optimization, as indicated in Equations (29)–(31):
B i L S T M = B i L S T M ( h d )
G A N = G A N ( d p )
T e x t C N N = T e x t C N N ( s t r i d e s )
The optimal hyperparameter search results obtained through IPSO are presented in Table 5.

4.5. Comparative Model Experiments and Results Analysis

To validate the performance of our model and analyze the impact of the GAN mechanism, considering the characteristics of text structure and sentiment analysis tasks, this study conducted comparative experiments on four datasets: IMDB, MR, SST-2, and HR. Our model was compared with classical models such as FastText, TextCNN-MultiChannel, DCNN, RCNN, BiLSTM+EMB_Att, BiLSTM-cnn, MLACNN, and MC-AttCNN-AttBiGRU, as well as our model without the GAN. The details of the comparative models are as follows:
(1)
FastText [38]: an open-source model by Facebook which is a rapid text classification approach. Its fundamental idea involves the summation of N-gram vectors of the text, followed by averaging to obtain the vector representation of the text. This vector is then fed into a softmax classifier to yield the classification outcome.
(2)
TextCNN-MultiChannel [5]: This model, also known as the TextCNN model proposed by Kim, is a multi-channel convolutional neural network. It utilizes convolutional kernels with window sizes of 3, 4, and 5 in the convolutional layers to extract local text features of various scales. These features are then concatenated after undergoing max-pooling, followed by classification using the Softmax operation.
(3)
DCNN [39]: The DCNN employs wide convolutions for feature extraction, allowing it to effectively utilize the edge information present in the text.
(4)
RCNN [40]: The RCNN employs a recurrent structure to learn word representations while simultaneously capturing contextual information as much as possible. Additionally, it incorporates maximum pooling to automatically identify the pivotal role of words in text classification, thereby capturing critical semantics within the text.
(5)
BiLSTM+EMB_Att [41]: The BiLSTM+EMB_Att model employs an attention mechanism to directly learn the weight distribution of each word’s sentiment tendency from the foundational word vectors. It then utilizes a BiLSTM to capture the semantic information within the text, followed by a fusion process for classification.
(6)
BiLSTM-CNN [17]: The BiLSTM-CNN model initially employs BiLSTM to extract contextual semantic information. Subsequently, it utilizes multiple convolutional layers to extract local semantics and then fuses them. The classification is ultimately performed using the Softmax function.
(7)
MLACNN [42]: The MLACNN model employs CNN to extract local semantic information and utilizes LSTM-Attention to obtain global semantic information. The outputs from both components are fused to achieve classification.
(8)
MC-AttCNN-AttBiGRU [21]: This model incorporates the Attention mechanism over word embeddings. It employs both TextCNN and BiGRU to capture local and global features, respectively. The extracted features from both components are then fused and used for classification.
Analysis of the comparative experimental results in Table 6 reveals that, on a macroscopic level, individual deep learning models perform consistently lower across various metrics compared to deep stacked models. This discrepancy can be attributed to the heightened emphasis of deep-layered network models in fully harnessing features, showcasing a more pronounced utilization compared to singular models. At a micro level, the TextCNN-MultiChannel model outperforms FastText across various datasets in terms of accuracy. This is attributed to the fact that the FastText model, relying on text N-gram vectors and simple averaging of word vectors, not only fails to conduct feature extraction but also introduces substantial noise. In contrast, the TextCNN-MultiChannel model employs three different sizes of convolutional kernels for localized feature extraction, resulting in superior and noise-free features. Building upon the foundation of TextCNN-MultiChannel, the DCNN model augments the convolutional kernel width to incorporate edge information within the text, achieving the utilization and comprehension of global textual information. As a result, DCNN and TextCNN-MultiChannel exhibit comparable performances in practical experiments. RCNN focuses not only on temporal continuity but also effectively employs CNN for feature extraction, merging global and local features. This compensates for DCNN’s limitations in temporal feature extraction capability, resulting in slightly improved performance compared to DCNN. Both BiLSTM+EMB_Att and BiLSTM-CNN belong to stacked networks, and they exhibit significant advantages in feature extraction and utilization compared to single-network models. BiLSTM+EMB_Att combines BiLSTM with the attention mechanism to effectively extract keywords relevant to sentiment analysis. BiLSTM-CNN, on the other hand, first extracts contextual semantic features and then employs CNN for local convolutional operations on global features. While their approaches differ, both models achieve similar outcomes, with accuracies of 84.49% and 85.75%, 88.84% and 88.57%, 81.32% and 80.60%, and 89.03% and 89.69% on the four datasets, respectively. MLACNN synthesizes the strengths of both BiLSTM+EMB_Att and BiLSTM-CNN. It incorporates TextCNN for local semantic understanding and BiLSTM for global context comprehension, further utilizing attention to appropriately allocate weights to features. In comparison to BiLSTM-CNN, MLACNN showcases a roughly 1% improvement across three metrics. Differentiating itself from MLACNN, MC-AttCNN-AttBiGRU integrates a dual-channel processing mechanism. After connecting word vectors with attention, it leverages TextCNN and BiGRU for parallel processing. Experimental results indicate that MC-AttCNN-AttBiGRU, with its dual-channel processing approach, outperforms MLACNN in terms of accuracy.
In comparison to the aforementioned models, our proposed model not only incorporates a module for processing both global and local features through Self-Attention but also enhances its robustness by introducing a perturbed text generation mechanism based on GAN and integrating the generated perturbed text into the training process. This augmentation further enhances the model’s fitting capacity and resilience against disturbances. Moreover, by utilizing an IPSO during the training process, the model’s relevant hyperparameters are tuned to their optimal values, fully unleashing the model’s performance potential. In experimental validation, our model achieves accuracy rates of 91.23%, 91.74%, 85.49%, and 94.59% on the four datasets, respectively. Additionally, the Recall and F1 scores of our model outperform those of the other comparative models.

4.6. Impact Analysis of GAN

To assess the impact of GANs on the model’s robustness against perturbations, this section employs several comparative models from the previous section as well as the model presented in this paper. The experimental results are visualized in Table 7 to facilitate analysis.
In analyzing the impact of GAN mechanisms on model performance, we conducted comparative experiments. Through these comparative experiments, it was observed that perturbed text generated during model training with GAN mechanisms can effectively identify ambiguous text in sentiment analysis, resulting in an approximate 1% improvement in accuracy across all models on the dataset. This partially mitigates the interference caused by perturbed text. Additionally, the GAN mechanism in our proposed model also performed impressively, leading to accuracy improvements of 1.14%, 0.86%, 1.13%, and 1.35% on the four datasets, respectively.

4.7. Ablation and Validation Experiments

To validate the effectiveness of the IPSO, we selected BiLSTM, LSTM, TextCNN, and Attention as control groups, combined with the comprehensive algorithm. A comparative experiment was conducted against single models without the IPSO. The optimized hyperparameters chosen for each model are listed in Table 8.
After selecting the hyperparameters that need optimization for the models, we conducted ablation experiments to demonstrate the rationale and effectiveness of the model incorporating the IPSO in hyperparameter selection. The accuracy of the experimental results is shown in Figure 13, Figure 14, Figure 15 and Figure 16.
From the above bar chart, it is evident that the IPSO has effectively identified suitable hyperparameters for each respective model, fully unleashing their performance potential. The accuracy of each model across the four datasets has been enhanced by an average of approximately 0.5% to 1%, with LSTM-TextCNN, BiLSTM-TextCNN, LSTM-Attention, and the proposed model experiencing even greater average improvements of 1.18%, 1.05%, 1.13%, and 1.065%, respectively. For our model, BiLSTM-TCSA, comparative experiments were also conducted. The performance across four datasets reveals that utilizing the IPSO algorithm for hyperparameter tuning outperforms manual tuning in terms of accuracy by 0.89%, 1.58%, 0.61%, and 1.18%, respectively. This is because manual tuning relies on brute-force search, which becomes challenging when dealing with continuous variables, leading to an infinite number of hyperparameter combinations. In such cases, the quality of manually selected hyperparameters largely depends on luck. In contrast, the IPSO algorithm iteratively updates and shares particle information within the swarm, ultimately finding the optimal combination in the search space. This highlights how the IPSO algorithm outperforms manual empirical selection when it comes to choosing model hyperparameters.

5. Conclusions and Future Work

The proposed deep learning model for short-text sentiment analysis, based on the IPSO, effectively leverages the IPSO approach to extract deep textual features. Moreover, the model’s robustness against disturbances is enhanced through the generation of a large amount of perturbed text using a GAN model. Experimental results indicate that utilizing the IPSO yields varied effects based on different combinations of model hyperparameters. The algorithm dynamically and adaptively adjusts inertia weights and learning factors, which not only prevents particles from becoming trapped in local optima to some extent, but also rapidly and effectively facilitates complete model convergence. This feature enables the full unleashing of model performance potential.
In the subsequent stages of experimental research, further efforts will be dedicated to refining the PSO. Additionally, consideration will be given to conducting horizontal comparisons with other parameter optimization algorithms, such as Bayesian optimization, genetic algorithms, and simulated annealing. The aim is to achieve further enhancements in metrics such as the model’s prediction accuracy.

Author Contributions

Y.P. and Y.Y. contributed to the conception of this study; Y.Y. performed the experiment, contributed significantly to analysis and wrote the manuscript; D.W. helped perform the analysis with constructive discussions. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the National Natural Science Foundation of China (No. 61966017) and by the Natural Science Foundation project of JiangXi province (No. 20224BAB202013 and No. 20212BAB202017).

Data Availability Statement

This study is an experimental analysis of a publicly available dataset.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Vries, A.; Mamoulis, N.; Nes, N. Efficient k-NN search on vertically decomposed data. In Proceedings of the 2002 ACM SIGMOD International Conference on Management of Data, Madison, WI, USA, 3–6 June 2002; pp. 322–333. [Google Scholar]
  2. Chen, Z.; Shi, G.; Wang, X. Text Classification Based on Naive Bayes Algorithm with Feature Selection. Int. J. Inf. 2012, 15, 4255–4260. [Google Scholar]
  3. Wang, J.H.; Yu, H.; Chan, W. Automatic text Classification based on KNN+ Hierarchical SVM. Comput. Appl. Softw. 2016, 33, 38–41. [Google Scholar]
  4. Rojas, B.; Maria, L. Deep learning for sentiment analysis. Ling. Linguist. Compass 2016, 10, 205–212. [Google Scholar]
  5. Kim, Y. Convolutional Neural Networks for Sentence Classification. In Proceedings of the Conference on Empirical Methods in Natural Language Processing, Doha, Qatar, 25–29 October 2014; Association for Computational Linguistics: Stroudsburg, PA, USA, 2014; pp. 1746–1751. [Google Scholar]
  6. Conneau, A.; Schwenk, H.; Barrault, L. Very deep convolutional networks for text classification. arXiv 2016, arXiv:1606.01781v1. [Google Scholar]
  7. Le, H.T.; Cerisara, C.; Denis, A. Do convolutional networks need to be deep for text classification? arXiv 2017, arXiv:1707.04108. [Google Scholar]
  8. Johnson, R.; Zhang, T. Deep pyramid convolutional neural networks for text categorization. In Proceedings of the 55th Annual Meetings of the Association for Computational Linguistics, Vancouver, BC, Canada, 30 July–4 August 2017; pp. 562–570. [Google Scholar]
  9. Guo, J.; Yue, B.; Xu, G. An enhanced convolutional neural network model for answer selection. In Proceedings of the 26th International Conference on World Wide Web Companion, Perth, Australia, 3–7 April 2017; pp. 789–790. [Google Scholar]
  10. Wang, H.; He, J.; Zhang, X. A short text classification method based on n-gram and CNN. Chin. J. Electron. 2020, 29, 248–254. [Google Scholar] [CrossRef]
  11. Irsoy, O.; Cardie, C. Opinion Mining with Deep Recurrent Neural Networks. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), Doha, Qatar, 25–29 October 2014; pp. 720–728. [Google Scholar]
  12. Xin, W.; Liu, Y.; Sun, C. Predicting Polarities of Tweets by Composing Word Embeddings with Long Short Term Memory. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing, Beijing, China, 26–31 July 2015; pp. 1343–1353. [Google Scholar]
  13. Hochreiter, S.; Schmidhuber, J. Long short-term memery. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  14. Gers, F.A.; Schermidhber, J.; Cummins, F. Learning to forget: Continual prediction with LSTM. Neual Comput. 2000, 12, 2451–2471. [Google Scholar] [CrossRef]
  15. Cho, K.; Van Merrienboer, B.; Gulcehre, C. Learning phrase representations using RNN encoder-decoder for statistical machine translation. arXiv 2014, arXiv:1406.1078. [Google Scholar]
  16. Tai, K.S.; Socher, R.; Manning, C.D. Improved semantic representations from tree-structured long short-term memory networks. arXiv 2015, arXiv:1503.00075. [Google Scholar]
  17. Zhao, H.; Wang, L.; Wang, W.J. Text sentiment analysis based on serial hybrid model of bi-directional long short-term memory and convolutional neural network. J. Comput. Appl. 2020, 40, 16–22. [Google Scholar]
  18. Vaswani, A.; Shazeer, N.; Parmar, N. Attention is all you need. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 5998–6008. [Google Scholar]
  19. Yang, Z.; Yang, D.; Dyer, C. Hierarchical attention networks for document classification. In Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, San Diego, CA, USA, 12–17 June 2016; pp. 1480–1489. [Google Scholar]
  20. Zhou, Y.; Xu, J.; Cao, J. Hybrid attention networks for Chinese short text classification. Comput. Syst. 2017, 21, 759–769. [Google Scholar] [CrossRef]
  21. Cheng, Y.; Yao, L.B.; Zhang, G.H.; Tang, T.W.; Xiang, G.X.; Chen, H.M.; Feng, Y.; Cai, Z. Text Sentiment Orientation Analysis of Multi-Channels CNN and BiGRU Based on Attention Mechanism. J. Comput. Res. Dev. 2020, 57, 2583–2595. [Google Scholar]
  22. Matthias, F.; Katharina, E.; Stefan, F. Auto-Sklearn 2.0: Hands-free AutoML via Meta-Learning. arXiv 2021, arXiv:2007.04074v2. [Google Scholar]
  23. Wang, Y.; Yan, Y.; Li, Z. Efficient and Robust Auto-tuning of Deep Learning Hyperparameters with Gaussian Processses. JMLR 2023, 24, 1–99. [Google Scholar]
  24. Jeremy, B.; Chirs, M. Automatic Gradient Descent: Deep Learning without Hyperparameters. arXiv 2023, arXiv:2304.05187. [Google Scholar]
  25. Yapici, H.; Cetinkaya, N. A new meta-heuristic optimizer: Pathfinder algorithm. Appl. Soft Comput. 2019, 78, 545–568. [Google Scholar] [CrossRef]
  26. Mirjalili, S.; Hatamlou, A. Multi-verse optimizer: A nature-inspired algorithm for global optimization. Neural Comput. Appl. 2016, 27, 459–513. [Google Scholar] [CrossRef]
  27. Mirijalili, S. Moth-flame optimization algorith: A novel nature-inspired heuristic paradigm. Knowl.-Based Syst. 2015, 89, 228–249. [Google Scholar] [CrossRef]
  28. Abualigah, L.; Diabat, A.; Mirjalili, S. The arithmetic optimization algorithm. Comput. Methods Appl. Mech. Eng. 2021, 376, 133609. [Google Scholar] [CrossRef]
  29. Kennedy, J.; Eberhart, R. Particle Swarm Optimization. In Proceedings of the Icnn95-International Conference on Neural Networks, Perth, WA, Australia, 27 November–1 December 1995; pp. 1942–1948. [Google Scholar]
  30. Wang, J.J.; Tufan, K. Parameter Optimization of Interval Type-2 Fuzzy Neural Network Based on PSO and BBBC Methods. IEEE/CAA J. Autom. Sin. 2019, 1, 247–257. [Google Scholar] [CrossRef]
  31. Shafipour, M.; Rashno, A.; Fadaei, S. Particle distance rank feature selection by particle swarm optimization. Expert Syst. Appl. 2021, 185, 115620. [Google Scholar] [CrossRef]
  32. Graves, A.; Schmidhuber, J. Framewise phoneme classification with bidirectional LSTM networks. In Proceedings of the IEEE International Joint Conference on Neural Networks, Montreal, QC, Canda, 31 July–4 August 2005; pp. 2047–2052. [Google Scholar]
  33. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M. Generative Adversarial Nets. In Proceedings of the International Conference on Neural Information Processing Systems, Montreal, QC, Canada, 8–13 December 2014; pp. 2672–2680. [Google Scholar]
  34. Liang, J.J.; Wei, J.J.; Jiang, Z.F. Generative Adversarial Networks GAN Overview. J. Front. Comput. Sci. Technol. 2020, 14, 1–17. [Google Scholar]
  35. Miyato, T.; Dai, A.M.; Goodfellow, I. Adversarial training methods for semi-supervised text classification. In Proceedings of the International Conference on Learning Representations, Toulon, France, 24–26 April 2017; pp. 1–11. [Google Scholar]
  36. Chen, R.L. Attention-based adversarial multi-task review text classification. Master’s Thesis, Dalian University of Technology, Dalian, China, 10 May 2019. [Google Scholar]
  37. Pennington, J.; Socher, R.; Manning, C. Glove: Global Vectors for Word Representation. In Proceedings of the Conference on Empirical Methods in Natural Language Processing, Doha, Qatar, 25–29 October 2014; pp. 1532–1543. [Google Scholar]
  38. Joulin, A.; Grave, E.; Bojanowski, P. Bag of tricks for efficient text classification. arXiv 2016, arXiv:1607.07159v3. [Google Scholar]
  39. Nal, K.; Edward, G.; Phil, B. A convolutional neural network for modelling sentences. In Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics, Baltimore, MD, USA, 23–25 June 2014; pp. 655–665. [Google Scholar]
  40. Lai, S.; Xu, L.; Liu, K. Recurrent convolutional neural networks for text classification. In Proceedings of the 29th Conference on Artififcal Inteligence, Austin, TX, USA, 25–30 January 2015; pp. 2267–2273. [Google Scholar]
  41. Guan, P.F.; Li, B.A.; Lv, X.Q.; Zhou, J.S. Attention Enhanced Bi-directional LSTM for Sentiment Analysis. JCIP 2019, 33, 105–111. [Google Scholar]
  42. Teng, J.B.; Kong, W.W.; Tian, Q.X.; Wang, Z.Q. Text Classification Method Based on LSTM-Attention and CNN Hybrid Model. Comput. Eng. Appl. 2021, 57, 126–133. [Google Scholar]
Figure 1. Parameter tuning iterations in grid search.
Figure 1. Parameter tuning iterations in grid search.
Electronics 12 04119 g001
Figure 2. The acceleration of ‘w’ decay in iterations.
Figure 2. The acceleration of ‘w’ decay in iterations.
Electronics 12 04119 g002
Figure 3. The variation of ‘ c 1 ’ and ‘ c 2 ’ in iterations. (a) The variation of ‘ c 1 ’. (b) The variation of ‘ c 2 ’.
Figure 3. The variation of ‘ c 1 ’ and ‘ c 2 ’ in iterations. (a) The variation of ‘ c 1 ’. (b) The variation of ‘ c 2 ’.
Electronics 12 04119 g003
Figure 4. IPSO Algorithm Flow.
Figure 4. IPSO Algorithm Flow.
Electronics 12 04119 g004
Figure 5. Three-Dimensional Plot of the (a) ‘ h o l d e r _ t a b l e ’, (b) ‘ e a s o m ’, and (c) ‘ r a s t r i g i n ’.
Figure 5. Three-Dimensional Plot of the (a) ‘ h o l d e r _ t a b l e ’, (b) ‘ e a s o m ’, and (c) ‘ r a s t r i g i n ’.
Electronics 12 04119 g005
Figure 6. Analysis of the PSO and IPSO in ‘ h o l d e r t a b l e ’, ’ e a s o m ’ and ‘ r a s t r i g i n ’. (a) PSO in ‘ h o l d e r t a b l e ’. (b) IPSO in ‘ h o l d e r _ t a b l e ’. (c) PSO in ‘ e a s o m ’. (d) IPSO in ‘ e a s o m ’. (e) PSO in ‘ r a s t r i g i n ’. (f) IPSO in ‘ r a s t r i g i n ’.
Figure 6. Analysis of the PSO and IPSO in ‘ h o l d e r t a b l e ’, ’ e a s o m ’ and ‘ r a s t r i g i n ’. (a) PSO in ‘ h o l d e r t a b l e ’. (b) IPSO in ‘ h o l d e r _ t a b l e ’. (c) PSO in ‘ e a s o m ’. (d) IPSO in ‘ e a s o m ’. (e) PSO in ‘ r a s t r i g i n ’. (f) IPSO in ‘ r a s t r i g i n ’.
Electronics 12 04119 g006
Figure 7. The Structure of BiLSTM.
Figure 7. The Structure of BiLSTM.
Electronics 12 04119 g007
Figure 8. Internal Structure of LSTM Cell.
Figure 8. Internal Structure of LSTM Cell.
Electronics 12 04119 g008
Figure 9. TextCNN Operation.
Figure 9. TextCNN Operation.
Electronics 12 04119 g009
Figure 10. Self-Attention calculation process.
Figure 10. Self-Attention calculation process.
Electronics 12 04119 g010
Figure 11. Architecture of GAN Network Model.
Figure 11. Architecture of GAN Network Model.
Electronics 12 04119 g011
Figure 12. Architecture of BiLSTM-TCSA.
Figure 12. Architecture of BiLSTM-TCSA.
Electronics 12 04119 g012
Figure 13. Analysis of IPSO’s Impact on Model Performance (IMDB).
Figure 13. Analysis of IPSO’s Impact on Model Performance (IMDB).
Electronics 12 04119 g013
Figure 14. Analysis of IPSO’s Impact on Model Performance (SST-2).
Figure 14. Analysis of IPSO’s Impact on Model Performance (SST-2).
Electronics 12 04119 g014
Figure 15. Analysis of IPSO’s Impact on Model Performance (MR).
Figure 15. Analysis of IPSO’s Impact on Model Performance (MR).
Electronics 12 04119 g015
Figure 16. Analysis of IPSO’s Impact on Model Performance (HR).
Figure 16. Analysis of IPSO’s Impact on Model Performance (HR).
Electronics 12 04119 g016
Table 1. CEC standard benchmark functions.
Table 1. CEC standard benchmark functions.
Function NameMathematical Expression
axis_parallel_hyper_ellipsoid f ( x ) = i = 1 D x i 2 × 10 6
rosenbrocks_valley f ( x ) = i = 1 D 1 100 × ( x i + 1 x i 2 ) 2 + ( 1 x i ) 2
schwefel f ( x ) = i = 1 D x i × sin ( | x i | )
ackley f ( x ) = 20 × exp [ 1 5 1 n i = 1 D x i 2 ] exp [ 1 n i = 1 D cos ( 2 π x i ) ] + 20 + e
Griewank f ( x ) = 1 4000 × i = 1 D x i 2 i = 1 D ( cos ( x i s q r t ( i ) ) ) + 1
Table 2. The final convergence values of various algorithms on the CEC standard benchmark functions.
Table 2. The final convergence values of various algorithms on the CEC standard benchmark functions.
Axis
Parallel
Hyper
Ellipsoid
Rosenbrocks
Valley
SchwefelAckleyGriewank
PA 1.28 × 10 7 1.19 × 10 7 830.983 3.62 × 10 4 7.09 × 10 5
MVO 1.43 × 10 7 5.82 × 10 8 830.983 1.03 × 10 4 1.29 × 10 6
MFA 1.99 × 10 8 6.22 × 10 8 830.075 1.13 × 10 4 9.86 × 10 7
AOA 2.21 × 10 3 6.13 × 10 3 830.075 1.26 × 10 1 5.61 × 10 3
PSO 1.75 × 10 8 1.07 × 10 9 830.069 5.51 × 10 5 9.52 × 10 8
Table 3. Experimental environment.
Table 3. Experimental environment.
DeviceVersion
Operating SystemWin10 64-bit
CPUAMD Ryzen 7 5800H
RAM16G
GPUGTX1650
Programming LanguagePython 3.8
Integrated Development EnvironmentPycharm Community Edition
Deep Learning FrameworkKreas 2.4.3
Table 4. Information for each dataset.
Table 4. Information for each dataset.
DatasetsNumber of SamplesAverage
Length
Number of
Train Data
Number of
Test Data
IMDB50,000 + 500028238,50016,500
MR10,662 + 10662082103518
SST-270,042 + 70041353,93223,114
HR877,640 + 87,76424668,083286,321
Table 5. Results of Hyperparameter Optimization (IMDB/SST-2/MR/HR).
Table 5. Results of Hyperparameter Optimization (IMDB/SST-2/MR/HR).
HyperparameterValue
d p 0.38/0.42/0.37/0.44
h d 141/97/94/101
l r 0.0079/0.0072/0.0068/0.0078
b a t c h _ s i z e 32/32/32/64
s t r i d e s 1/1/1/1
Table 6. Results of Comparative Experiments for four datasets (Accuracy/Recall/F1).
Table 6. Results of Comparative Experiments for four datasets (Accuracy/Recall/F1).
ModelsIMDBSST-2MRHR
FastText81.25/80.95/82.7185.03/84.82/84.9176.81/77.03/76.7584.32/84.26/85.71
TextCNN-MultiChannel82.33/81.27/80.5687.12/86.91/87.2477.94/76.02/78.1384.79/83.09/83.26
DCNN81.99/83.26/82.0386.83/86.03/86.8077.52/76.59/77.4686.76/84.23/83.79
RCNN82.41/81.71/80.2687.82/88.41/87.7480.28/81.99/80.2587.91/85.99/85.41
BiLSTM+EMB_Att84.49/83.02/85.1388.84/87.56/88.6581.32/82.31/81.3789.03/87.23/88.41
BiLSTM-CNN85.75/84.26/84.5188.57/89.21/88.4180.60/79.21/80.4789.69/87.29/87.03
MLACNN86.47/86.74/85.0289.83/89.95/89.9182.51/81.03/81.8990.24/89.46/88.73
MC-AttCNN-AttBiGRU88.39/87.09/87.3890.35/90.04/90.3283.38/82.71/83.3092.35/91.23/91.45
BiLSTM-TCSA(Ours)91.38/92.15/91.5291.74/91.04/92.2085.49/84.76/84.2194.59/92.71/92.17
Table 7. The optimized hyperparameters for each model (without GAN/with GAN).
Table 7. The optimized hyperparameters for each model (without GAN/with GAN).
ModelsIMDBSST-2MRHR
FastText81.25/82.0385.03/85.9976.81/77.7984.32/85.72
TextCNN-MultiChannel82.33/83.1687.12/87.8177.94/79.0384.79/86.23
DCNN81.99/83.2686.83/87.7177.52/78.4986.76/88.01
RCNN82.41/83.4487.82/88.5280.28/81.3187.91/89.02
BiLSTM-TCSA(Ours)90.24/91.3890.88/91.7484.36/85.4993.24/94.59
Table 8. The optimized hyperparameters for each model.
Table 8. The optimized hyperparameters for each model.
ModelHidden Layer DimensionLearning RateBatch_SizeStridesDropout
LSTM
BiLSTM
LSTM-TextCNN
BiLSTM-TextCNN
LSTM-Attention
BiLSTM-Attention
TextCNN-Attention
BiLSTM-TCSA (Ours)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yue, Y.; Peng, Y.; Wang, D. Deep Learning Short Text Sentiment Analysis Based on Improved Particle Swarm Optimization. Electronics 2023, 12, 4119. https://doi.org/10.3390/electronics12194119

AMA Style

Yue Y, Peng Y, Wang D. Deep Learning Short Text Sentiment Analysis Based on Improved Particle Swarm Optimization. Electronics. 2023; 12(19):4119. https://doi.org/10.3390/electronics12194119

Chicago/Turabian Style

Yue, Yaowei, Yun Peng, and Duancheng Wang. 2023. "Deep Learning Short Text Sentiment Analysis Based on Improved Particle Swarm Optimization" Electronics 12, no. 19: 4119. https://doi.org/10.3390/electronics12194119

APA Style

Yue, Y., Peng, Y., & Wang, D. (2023). Deep Learning Short Text Sentiment Analysis Based on Improved Particle Swarm Optimization. Electronics, 12(19), 4119. https://doi.org/10.3390/electronics12194119

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop