1. Introduction
Sentiment analysis involves extracting viewpoints and judging emotional tendencies in subjective texts. Aspect-level sentiment analysis is more refined, particularly when judging the emotional polarity of texts in terms of given aspect words (e.g., positive, negative, and neutral) [
1,
2,
3]. As shown in
Figure 1, the sentence has two sides: “price” and “screen”. The emotional polarity of the sentence is negative with regard to “price” and positive in terms of “screen”. The main task of aspect-level sentiment analysis is to accurately find the opinions and emotional tendencies of sentences according to the identified aspects. In this way, the statement is partially understood. This is currently an important tool for online opinion gathering for governments and businesses concerning policies or business strategies [
4]. For example, it can be used to analyze a company’s blockchain statements and obtain a short-term value evaluation of the company [
5]. It can also be employed to analyze the behavior of users in the blockchain [
6].
Attention mechanisms can improve the structure of deep learning by concentrating on the important points. Wang et al. [
7] presented the application of attention mechanisms to long short-term memory (LSTM) [
8]. They used an attention mechanism to acquire the semantic information between words in a text and aspect words, obtaining crucial parts of the sentences to obtain a more accurate aspect sentiment analysis. Ma et al. [
9] introduced interactive attention networks (IANs) based on LSTM and attention mechanisms to realize the interactive learning of attention based on goals and contexts. However, many studies have used attention mechanisms to obtain the relation between words and aspect words in sentences from a semantic perspective [
10,
11,
12]. When a sentence has several aspect words and the emotional polarity of these words is different, the attention mechanism can easily focus on the wrong part or a part that is unrelated to the aspect words. As shown in
Figure 1, if the attention mechanism mistakenly concentrates on the opinion word “good” for the aspect of “price”, it is likely to cause errors in the emotional polarity.
In order to enable the model to accurately concentrate on opinion words that correspond to the aspect words, it is important to introduce the grammatical relations between words in sentiment analysis. Graph convolutional networks (GCNs) [
12] combined with syntactic dependency trees can better obtain grammatical features. Zhang C. et al. [
10] employed graph convolutional networks in aspect-level sentiment analysis, and their results showed excellent model performance. In the following study, Zhang Z. et al. [
11] joined the grammatical structure and semantic correlation of aspect words. They presented an aspect perception attention mechanism combined with a self-attention mechanism, and their work shows promising results.
It is evident that a sentence not only contains aspect words but also other types of words. If sentiment analysis was only performed with the combination of grammatical and semantic features of aspect words, some important information in the sentence would be lost. Therefore, we introduced sentiment support words (SSWs; proposed in our previous study [
13]) to supplement a sentiment analysis in order to obtain more complete grammatical and semantic features of sentences. SSWs are words that influence an aspect word in a sentence, often supporting the emotional polarity of the whole sentence in one aspect. Through a grammatical dependency tree, we can obtain the aspect of sentiment support words.
Figure 2 shows a concrete example of this. From the figure, we can see that the aspect words are attributes of the object entities in the sentence. For example, an aspect can be the price of computers or computer screens, while sentiment support words directly evaluate part of the target information. For example, when the aspect is the price of computers, sentiment support words evaluate the price of computers as “expensive”. The emotional polarity here is clearly negative. Therefore, combining aspect words with SSWs is beneficial for identifying the emotional polarity corresponding to certain aspect words. We can identify the SSWs corresponding to each aspect of a sentence by analyzing the parts of speech and applying fixed rules. This approach allows us to consider the grammatical and semantic information of emotive support words in the sentence.
Based on this idea, this paper presents a sentiment-supported graph conventional network (SSGCN). First, we used an aspect perception attention mechanism and a self-attention mechanism to obtain attention scores; then, we input these scores into the GCN through a syntactic mask matrix to obtain sentence features that initially combined semantic and grammatical information. Then, the semantic and grammatical information of the SSW combination was obtained. Finally, this was combined with the semantic and grammatical information of the aspect words to obtain the enhanced information of the combination of syntactic and semantic information. In this way, a more complete combination of semantic and grammatical information could be obtained. When evaluated on three publicly available datasets, the model presented in this paper demonstrated a greater efficiency than the eight comparison models.
The main contributions of this paper are as follows:
- (1)
A method of combining semantics and grammar using sentiment support words is presented. This method effectively captures the evaluation information of aspect words in sentences and assists the model in extracting the correct feature information.
- (2)
A further combination method is presented. This method integrates syntactic and semantic information based on aspect and sentiment support words. It enhances the comprehensiveness of the constructed combined features by incorporating semantic and grammatical information.
- (3)
Experiments on three public datasets demonstrated that the SSGCN was more effective than the benchmark models.
The remainder of this paper is organized as follows:
Section 2 provides a review of the research associated with aspect-level sentiment analysis. Then, the proposed SSGCN model is introduced in detail in
Section 3. The experiments and discussion are presented in
Section 4. Finally,
Section 5 provides our conclusions.
2. Related Work
Aspect-level sentiment analysis is a subtask of sentiment analysis. Compared with ordinary sentiment analysis tasks, it is a more refined emotional classification task aimed at discovering the emotional polarity of a sentence in a specific aspect.
Early sentiment analysis tasks mostly adopted machine learning methods [
14]. Kiritchenko et al. [
15] obtained a large number of analytical attributions for classification by manually constructing features. However, this method is labor-intensive, demands a lot of human resources and time, and loses the dependency between the aspect words and the context. Kim et al. [
16] used neural networks to model text, concluding that convolutional neural networks (CNNs) perform well when extracting the local features of sentences in a text. Therefore, CNNs have begun to be widely used in sentiment analysis [
17,
18,
19]. The variant model of the recurrent neural network (RNN) is also a common method used in sentiment analysis [
20]. For instance, in the case of using aspect word information, Tang et al. [
21] introduced target-dependent long short-term memory (TD-LSTM). This approach employs two LSTMs to model the context before and after aspect words, utilizing the aspect words’ semantic information to judge the emotional polarity. This method achieved good results in their experiments. In aspect-level sentiment analysis, an attention-mechanism-based approach [
22,
23] directs the model’s increased attention toward the feature information of aspect words. For instance, Wang et al. [
7] input the word vector of aspect words into the structure of an attention mechanism. The obtained attention score represents the weight of each word in this aspect of the sentence. The higher the score, the more important it becomes. Finally, the sentence features are obtained by weighted summation. Their experiments showed that this method is superior to its baseline methods. Similarly, Huang et al. [
24] used BiLSTM-SNP to obtain the correlation between aspect words and content words, applying attention mechanisms to determine their weights. The experiments on multiple Chinese and English datasets showed that this method has significant advantages in terms of accuracy.
To a certain extent, the attention mechanism can find the semantic correlation between aspect words and other words in a sentence, but when the sentence composition is complex, the attention mechanism may make the model concentrate on the wrong part of the sentence [
25]. Unlike traditional deep learning models, graph convolutional networks capture the relationships and feature information between nodes in a graph by defining the convolution operations on the graph structure [
26]. Specifically, they first represent nodes in the graph as low-dimensional vectors and aggregate the information of adjacent nodes through matrix multiplication to obtain new feature representations for each node in the graph. Since GCNs address this deficiency better [
27], it was used for the first time by Zhang et al. [
10] for aspect-level sentiment analysis. Their proposed method was named the aspect-specific graph convolutional network (ASGCN). The model established the GCN on the syntactic dependency tree and acquired grammar-dependent information from a sentence. The dual graph convolutional network (DualGCN) [
28] was proposed by Li et al. They also argued for the complementarity of grammatical structure and semantic relations, and combined an attention mechanism with a grammatical dependency tree. In this model, a GCN is employed to obtain the abundant syntactic information on the syntactic dependency tree, and another GCN is applied to obtain the semantic relations using a self-attention mechanism. Yang et al. [
29] introduced tight connections into GCNs, enabling GCNs to obtain rich local and global information. A syntactic- and semantic-enhanced graph convolutional network (SSEGCN) was presented by Zhang et al. [
11]. It initially integrates the syntactic and semantic information of sentences. In order to model the relevance of specific semantics of terms from different aspects, Zhang et al. also introduced the aspect-aware attention mechanism and combined it with the self-attention mechanism. In order to highlight the aspect-related semantic information, Zhang et al. also took the attention score as the initial adjacency matrix of the graph convolutional network, constructed a syntactic mask matrix calculated from the different distances between words with the syntactic information, and used the aspect word mask matrix corresponding to the position of the aspect word to highlight the features after the graph convolutional network. Current research focuses on the simultaneous use of attention mechanisms and GCNs, beginning attempts to combine semantic information obtained using attention mechanisms with grammatical information obtained using GCNs. Therefore, the current research is devoted to finding a more efficient way to combine semantic and grammatical information.