Next Article in Journal
Evaluation of Specific Absorption Rate in Three-Layered Tissue Model at 13.56 MHz and 40.68 MHz for Inductively Powered Biomedical Implants
Next Article in Special Issue
Gender Classification Using Sentiment Analysis and Deep Learning in a Health Web Forum
Previous Article in Journal
Multi-Resonant-Based Sliding Mode Control of DFIG-Based Wind System under Unbalanced and Harmonic Network Conditions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Deep Learning-Based Approach for Multi-Label Emotion Classification in Tweets

by
Mohammed Jabreel
1,2,* and
Antonio Moreno
1
1
ITAKA Research Group, Universitat Rovira i Virgili, 43007 Tarragona, Spain
2
Department of Computer Science, Hodeidah University, Hodeidah 1821, Yemen
*
Author to whom correspondence should be addressed.
Appl. Sci. 2019, 9(6), 1123; https://doi.org/10.3390/app9061123
Submission received: 15 February 2019 / Revised: 11 March 2019 / Accepted: 12 March 2019 / Published: 17 March 2019
(This article belongs to the Special Issue Sentiment Analysis for Social Media)

Abstract

:
Currently, people use online social media such as Twitter or Facebook to share their emotions and thoughts. Detecting and analyzing the emotions expressed in social media content benefits many applications in commerce, public health, social welfare, etc. Most previous work on sentiment and emotion analysis has only focused on single-label classification and ignored the co-existence of multiple emotion labels in one instance. This paper describes the development of a novel deep learning-based system that addresses the multiple emotion classification problem in Twitter. We propose a novel method to transform it to a binary classification problem and exploit a deep learning approach to solve the transformed problem. Our system outperforms the state-of-the-art systems, achieving an accuracy score of 0.59 on the challenging SemEval2018 Task 1:E-cmulti-label emotion classification problem.

1. Introduction

Emotions are the key to people’s feelings and thoughts. Online social media, such as Twitter and Facebook, have changed the language of communication. Currently, people can communicate facts, opinions, emotions, and emotion intensities on different kinds of topics in short texts. Analyzing the emotions expressed in social media content has attracted researchers in the natural language processing research field. It has a wide range of applications in commerce, public health, social welfare, etc. For instance, it can be used in public health [1,2], public opinion detection about political tendencies [3,4], brand management [5], and stock market monitoring [6]. Emotion analysis is the task of determining the attitude towards a target or topic. The attitude can be the polarity (positive or negative) or an emotional state such as joy, anger, or sadness [7].
Recently, the multi-label classification problem has attracted considerable interest due to its applicability to a wide range of domains, including text classification, scene and video classification, and bioinformatics [8]. Unlike the traditional single-label classification problem (i.e., multi-class or binary), where an instance is associated with only one label from a finite set of labels, in the multi-label classification problem, an instance is associated with a subset of labels.
Most previous work on sentiment and emotion analysis has only focused on single-label classification. Hence, in this article, we focus on the multi-label emotion classification task, which aims to develop an automatic system to determine the existence in a text of none, one, or more out of eleven emotions: the eight Plutchik [9] categories (joy, sadness, anger, fear, trust, disgust, surprise, and anticipation) that are shown in Figure 1, plus optimism, pessimism, and love.
One of the most common approaches to addressing the problem of multi-label classification is the problem transformation. With this approach, a multi-label problem is transformed into one or more single-label (i.e., binary or multi-class) problems. Specifically, single-label classifiers are learned and employed; after that, the classifiers’ predictions are transformed into multi-label predictions.
Different transformation methods have been proposed in the multi-label literature. The most common method is called binary relevance [10,11]. The idea of the binary relevance method is simple and intuitive. A multi-label problem is transformed into multiple binary problems, one problem for each label. Then, an independent binary classifier is trained to predict the relevance of one of the labels. Although binary relevance is popular in the literature, due to its simplicity, it suffers from directly modeling correlations that may exist between labels. However, it is highly resistant to overfitting label combinations, since it does not expect examples to be associated with previously-observed combinations of labels.
In this article, we propose a novel transformation method, called x y -pair-set, for the multi-label classification problem. Unlike binary relevance methods, our method transforms the problem into only one binary classification problem as described in Section 3. Additionally, we exploit the successes of deep learning models, especially the word2vecmethods’ family [12] and the recurrent neural networks [13,14] and attention models [15,16], to develop a system that solves the transformed binary classification problem. The critical component of our system is the embedding module, which uses three embedding models and an attention function to model the relationship between the input and the label.
To summarize, the contribution of this work is four-fold.
  • We propose a novel transformation mechanism for the multi-label classification problem.
  • We propose a novel, attentive deep learning system, which we call Binary Neural Network (BNet), which works on the new transformation method. Our system is a data-driven, end-to-end neural-based model, and it does not rely on external resources such as parts of speech taggers and sentiment or emotion lexicons.
  • We evaluate the proposed system on the challenging multi-label emotion classification dataset of SemEval-2018 Task1: Affect in Tweets.
  • The experimental results show that our system outperforms the state-of-the-art systems.
The rest of the article is structured as follows. In Section 2, we overview the related work on multi-label problem transformation methods and Twitter sentiment and emotion analysis. In Section 3, we explain in detail the methodology. In Section 4, we report the experimental results. In Section 5, the conclusions and the future work are presented.

2. Related Works

In this section, we overview the most popular research studies related to this work. In Section 2.1, we summarize the most common multi-label problem transformation methods. Section 2.2 gives an overview of the state-of-the-art works on the problem of multi-label emotion classification on Twitter.

2.1. Problem Transformation Methods

Let X = { x 1 , x 2 , x n } be the set of all instances and Y = { y 1 , y 2 , , y m } be the set of all labels. We can define the set of data:
D = { ( x i , Y ^ i ) | x i X and Y i ^ Y is the set of labels associated with x i }
In this expression, D is called a supervised multi-label dataset.
The task of multi-label classification is challenging because the number of label sets grows exponentially as the number of class labels increases. One common strategy to address this issue is to transform the problem into a traditional classification problem. The idea is to simplify the learning process by exploiting label correlations. Based on the order of the correlations, we can group the existing transformation methods into three approaches [17,18], namely first-order approaches, second-order approaches, and high-order approaches.
First-order approaches decompose the problem into some independent binary classification problems. In this case, one binary classifier is learned for each possible class, ignoring the co-existence of other labels. Thus, the number of independent binary classifiers needed is equal to the number of labels. For each multi-label training example ( x i , Y ^ i ) D , y k Y , we construct a binary classification training set, D k as the following: x i will be regarded as one positive example if y k Y ^ i and one negative example otherwise. In the first case, we will get a training example in the form ( x i , 1 ) D k , which will be ( x i , 0 ) D k in the second case. Thus, for all labels { y 1 , y 2 , , y m } Y , m training sets { D 1 , D 2 , , D m } are constructed. Based on that, for each training set D k , one binary classifier can be learned with popular learning techniques such as AdaBoost [19], k-nearest neighbor [20], decision trees, random forests [21,22], etc. The main advantage of first-order approaches is their conceptual simplicity and high efficiency. However, these approaches can be less effective due to their ignorance of label correlations.
Second-order approaches try to address the lack of modeling label correlations by exploiting pairwise relationships between the labels. One way to consider pairwise relationships is to train one binary classifier for each pair of labels [23]. Although second-order approaches perform well in several domains, they are more complicated than the first-order approaches in terms of the number of classifiers. Their complexity is quadratic, as the number of classifiers needed is m 2 . Moreover, in real-world applications, label correlations could be more complex and go beyond second-order.
High-order approaches tackle the multi-label learning problem by exploring high-order relationships among the labels. This can be fulfilled by assuming linear combinations [24], a nonlinear mapping [25,26], or a shared subspace over the whole label space [27]. Although high-order approaches have stronger correlation-modeling capabilities than their first-order and second-order counterparts, these approaches are computationally demanding and less scalable.
Our transformation mechanism, shown in Section 3.1, is a simple as the first-order approaches and can model, implicitly, high-order relationships among the labels if some requirements, detailed in Section 3.1, are fulfilled. It requires only one binary classifier, and the number of training examples grows polynomially in terms of the number of instances and the number of labels. If the number of training examples in the multi-label training dataset is n and the number of the labels is m, then the number of the training examples in the transformed binary training set is n × m .

2.2. Emotion Classification in Tweets

Various machine learning approaches have been proposed for traditional emotion classification and multi-label emotion classification. Most of the existing systems solve the problem as a text classification problem. Supervised classifiers are trained on a set of annotated corpora using a different set of hand-engineered features. The success of such models is based on two main factors: a large amount of labeled data and the intelligent design of a set of features that can distinguish between the samples. With this approach, most studies have focused on engineering a set of efficient features to obtain a good classification performance [28,29,30]. The idea is to find a set of informative features to reflect the sentiments or the emotions expressed in the text. Bag-of-Words (BoW) and its variation, n-grams, is the representation method used in most text classification problems and emotion analysis. Different studies have combined the BoW features with other features such as the parts of speech tags, the sentiment and the emotion information extracted from lexicons, statistical information, and word shapes to enrich the text representation.
Although BoW is a popular method in most text classification systems, it has some drawbacks. Firstly, it ignores the word order. That means that two documents may have the same or a very close representation as far as they have the same words, even though they carry a different meaning. The n-gram method resolves this disadvantage of BoW by considering the word order in a context of length n. However, it suffers from sparsity and high dimensionality. Secondly, BoW is scarcely able to model the semantics of words. For example, the words beautiful, wonderful and view have an equal distance in BoW, where the word beautiful is closer to the word wonderful than the word view in the semantic space.
Sentiment and emotion lexicons play an essential role in developing efficient sentiment and emotion analysis systems. However, it is difficult to create such lexicons. Moreover, finding the best combination of lexicons in addition to the best set of statistical features is a time-consuming task.
Recently, deep learning models have been utilized to develop end-to-end systems in many tasks including speech recognition, text classification, and image classification. It has been shown that such systems automatically extract high-level features from raw data [31,32].
Baziotis et al. [33], the winner of the multi-label emotion classification task of SemEval-2018 Task1: Affect in Tweets, developed a bidirectional Long Short-Term Memory (LSTM) with a deep attention mechanism. They trained a word2vec model with 800,000 words derived from a dataset of 550 million tweets. The second place winner of the SemEval leaderboard trained a word-level bidirectional LSTM with attention, and it also included non-deep learning features in its ensemble [34]. Ji Ho Park et al. [35] trained two models to solve this problem: regularized linear regression and logistic regression classifier chain [11]. They tried to exploit labels’ correlation to perform multi-label classification. With the first model, the authors formulated the multi-label classification problem as a linear regression with label distance as the regularization term. In their work, the logistic regression classifier chain method was used to capture the correlation of emotion labels. The idea is to treat the multi-label problem as a sequence of binary classification problems by taking the prediction of the previous classifier as an extra input to the next classifier.
In this work, we exploited the deep learning-based approach to develop a system that can extract a high-level representation of the tweets and model an implicit high-order relationship among the labels. We used the proposed system alongside the proposed transformation method to train a function that can solve the problem of multi-label emotion classification in tweets. The next section explains the details of our proposed system.

3. Methodology

This section shows the methodology of this work. First, we explain in Section 3.1 the proposed transformation method, x y -pair-set. Afterwards, we describe the proposed system in Section 3.2.

3.1. x y -Pair-Set: Problem Transformation

The proposed transformation method x y -pair-set transforms a multi-label classification dataset D into a supervised binary dataset D ^ as follows:
x i X , y Y and ( x i , Y ^ i ) D , ! ( ( x i , y ) , ϕ ) D ^ where ϕ = 1 if y Y ^ i 0 o t h e r w i s e
Algorithm 1 explains the implementation of the proposed transformation method. It takes as inputs a multi-label dataset D (Equation (1)) and a set of labels Y, and it returns a transformed binary dataset. We show next an illustrative example.
Algorithm 1: x y -pair-set algorithm.
Applsci 09 01123 i001
Let X = { x 1 , x 2 } , Y = { a , b , c } and D = { ( x 1 , { a , c } ) , ( x 2 , { b } ) } . The output of the binary relevance transformation method is a set of three independent binary datasets, one for each label. That is, D a = { ( x 1 , 1 ) , ( x 2 , 0 ) } , D b = { ( x 1 , 0 ) , ( x 2 , 1 ) } , and D c = { ( x 1 , 1 ) , ( x 2 , 0 ) } . In contrast, the output of our transformation method is a single binary dataset D ^ = { ( ( x 1 , a ) , 1 ) , ( ( x 1 , b ) , 0 ) , ( ( x 1 , c ) , 1 ) , ( ( x 2 , a ) , 0 ) , ( ( x 2 , b ) , 1 ) , ( ( x 2 , c ) , 0 ) } .
The task in this case, unlike the traditional supervised binary classification algorithms, is to develop a learning algorithm to learn a function g : X × Y { 0 , 1 } . The success of such an algorithm is based on three requirements: (1) an encoding method to represent an instance x X as a high-dimensional vector V x , (2) a method to encode a label y Y as a vector V y , and (3) a method to represent the relation between the instance x and the label y. These three conditions make g able to capture the relationships inputs-to-labels and labels-to-labels. In this work, we take advantage of the successes of deep learning models to fulfill the three requirements listed above. We empirically show the success of our system with respect to these conditions as reported in Section 4.6 and Section 4.7.

3.2. BNet: System Description

This subsection explains the proposed system to solve the transformed binary problem mentioned above. Figure 2 shows the graphical depiction of the system’s architecture. It is composed of three parts: the embedding module, the encoding module, and the classification module. We explain in detail each of them below.

3.2.1. Embedding Module

Let ( W , y ) be the pair of inputs to our system, where W = { w 1 , w 2 , , w l } is the set of the words in a tweet and y is the label corresponding to an emotion. The goal of the embedding module is to represent each word w i by a vector v w i and the label by a vector v y .
Our embedding module can be seen as a function that maps a query and a set of key-value pairs to an output, where the query, keys, values, and output are all vectors. The output is computed as a weighted sum of the values, where the weight assigned to each value is computed by a compatibility function of the query with the corresponding key. The query is the trainable label embedding, E Q ( y ) ; the keys are the pretrained words embeddings, E K ( w i ) w i W ; and the values are the trainable words embeddings, E V ( w i ) w i W .
As shown in Figure 2, we used the output of E Q and E K as inputs to the attention model to find the alignments, i.e., the weights α , between the label y and the words W of the input tweet. This step models the relation between the input and the label. As soon as the weights were obtained, we then multiplied each word’s vector that came from the embedding E V by its corresponding weight. Given that, the final representation of a word w i W is given as the following:
v w i = E V ( w i ) · α ( w i , y )
The function α is an attention-based model, which finds the strength of the relationship between the word w i and the label y based on their semantic similarity. That is, α ( w i , y ) is a value based on the distance Δ ( w i , y ) between w i and y as:
α ( w i , y ) = e Δ ( w i , y ) w j W e Δ ( w j , y )
Here, Δ ( w i , y ) is a scalar score that represents the similarity between the word w i and the label y:
Δ ( w i , y ) = E k ( w i ) · v y T
v y = E Q ( y )
It is worth noting that α ( w i , y ) [ 0 , 1 ] and:
w i W α ( w i , y ) = 1

3.2.2. Encoding Module

The goal of the encoding module is to map the sequence of word representations { v w 1 , v w 2 , , v w l } that is obtained from the embedding module to a single real-valued dense vector. In this work, we used a Recurrent Neural Network (RNN) to design our encoder. RNN reads the input sequence of vectors in a forward direction (left-to-right) starting from the first symbol v w 1 to the last one v w l . Thus, it processes sequences in temporal order, ignoring the future context. However, for many tasks on sequences, it is beneficial to have access to future, as well as to past information. For example, in text processing, decisions are usually made after the whole sentence is known. The Bidirectional Recurrent Neural Network (BiRNN) variant [13] proposed a solution for making predictions based on both past and future information.
A BiRNN consists of forward ϕ and backward ϕ RNNs. The first one reads the input sequence in a forward direction and produces a sequence of forward hidden states ( h 1 , , h l ) , whereas the former reads the sequence in the reverse order ( v w l , , v w 1 ) , resulting in a sequence of backward hidden states ( h l , , h 1 ) .
We obtained a representation for each word v w t by concatenating the corresponding forward hidden state h t and the backward one h t . The following equations illustrate the main ideas:
h t = ϕ ( v w t , h t 1 )
h t = ϕ ( v w t , h t + 1 )
h t = [ h t ; h t ]
The final input representation of the sequence is:
c = F ( { h 1 , h 2 , , h l } )
We simply chose F to be the last hidden state (i.e., F ( { h 1 , h 2 , , h n } ) = h n ).
In this work, we used two Gated Recurrent Units (GRUs) [14], one as ϕ and the other as ϕ . This kind of RNN was designed to have more persistent memory, making them very useful to capture long-term dependencies between the elements of a sequence. Figure 3 shows a graphical depiction of a gated recurrent unit.
A GRU has reset ( r t ) and update ( z t ) gates. The former can completely reduce the past hidden state h t 1 if it finds that it is irrelevant to the computation of the new state, whereas the later is responsible for determining how much of h t 1 should be carried forward to the next state h t .
The output h t of a GRU depends on the input x t and the previous state h t 1 , and it is computed as follows:
r t = σ W r · [ h t 1 ; x t ] + b r
z t = σ W z · [ h t 1 ; x t ] + b z
h ˜ t = t a n h W h · [ ( r t h t 1 ) ; x t ] + b h
h t = ( 1 z t ) h t 1 + z t h ˜ t
In these expressions, r t and z t denote the reset and update gates, h ˜ t is the candidate output state, and h t is the actual output state at time t. The symbol ⊙ stands for element-wise multiplication; σ is a sigmoid function; and; stands for the vector-concatenation operation. W r , W z , W h R d h × ( d + d h ) and b r , b z , b h R d h are the parameters of the reset and update gates, where d h is the dimension of the hidden state and d is the dimension of the input vector.

3.2.3. Classification Module

Our classifier was composed of two feed-forward layers with the R e L U activation function followed by a Sigmoid unit.

4. Experiments and Results

In this section, we first describe the experimental details, and then, we describe the dataset and the pre-processing we used. Afterwards, we introduce the state-of-the-art systems we compared our system with, and finally, we report the empirical validation proving the effectiveness of our system.

4.1. Experimental Details

Table 1 shows the hyperparameters of our system, which was trained using Adam [36], with a learning rate of 0.0001, β 1 = 0.5, and a mini-batch size of 32 to minimize the binary cross-entropy loss function:
L ( θ , D ^ ) = E ( ( x i , y i ) , ϕ i ) D ^ ϕ i · g ( x i , y i ) + ( 1 ϕ i ) · ( 1 g ( x i , y i ) )
where, g ( x i , y i ) is the predicted value, ϕ i is the real value, and θ is the model’s parameters.
The hyperparameters of our system were obtained by applying Bayesian optimization [37]. We used the development set as a validation set to fine-tune those parameters.

4.2. Dataset

In our experiments, we used the multi-label emotion classification dataset of SemEval-2018 Task1: Affect in Tweets [30]. It contains 10,983 samples divided into three splits: training set (6838 samples), validation set (886 samples), and testing set (3259 samples). For more details about the dataset, we refer the reader to [38]. We trained our system on the training set and used the validation set to fine-tune the parameters of the proposed system. We pre-processed each tweet in the dataset as follows:
  • Tokenization: We used an extensive list of regular expressions to recognize the following meta information included in tweets: Twitter markup, emoticons, emojis, dates, times, currencies, acronyms, hashtags, user mentions, URLs, and words with emphasis.
  • As soon as the tokenization was done, we lowercased words and normalized the recognized tokens. For example, URLs were replaced by the token “<URL>”, and user mentions were replaced by the token “<USER>”. This step helped to reduce the size of the vocabulary without losing information.

4.3. Comparison with Other Systems

We compared the proposed system with the state-of-the-art systems used in the task of multi-label emotion classification, including:
  • SVM-unigrams: a baseline support vector machine system trained using just word unigrams as features [30].
  • NTUA-SLP: the system submitted by the winner team of the SemEval-2018 Task1:E-cchallenge [33].
  • TCS: the system submitted by the second place winner [34].
  • PlusEmo2Vec: the system submitted by the third place winner [35].
  • Transformer: a deep learning system based on large pre-trained language models developed by the NVIDIA AI lab [39].

4.4. Evaluation Metrics

We used multi-label accuracy (or Jaccard index), the official competition metric used by the organizers of SemEval-2018 Task 1: Affect in Tweets, for the E-c sub task, which can be defined as the size of the intersection of the predicted and gold label sets divided by the size of their union.
J a c c a r d = 1 | T | t T G t P t G t P t
In this expression, G t is the set of the gold labels for tweet t, P t is the set of the predicted labels for tweet t, and T is the set of tweets. Additionally, we also used the micro-averaged F-score and the macro-averaged F-score.
Let # c ( l ) denote the number of samples correctly assigned to the label l, # p ( l ) the number of samples assigned to l, and # ( l ) the number of actual samples in l. The micro-averaged F1-score is calculated as follows:
P m i c r o = l L # c ( l ) l L # p ( l )
R m i c r o = l L # c ( l ) l L # ( l )
F 1 m i c r o = 2 × P m i c r o × R m i c r o P m i c r o + R m i c r o
Thus, P m i c r o is the micro-averaged precision score, and R m i c r o is the micro-averaged recall score.
Let P l , R l , and F l denote the precision score, recall score, and the F1-score of the label l. The macro-averaged F1-score is calculated as follows:
P l = # c ( l ) # p ( l )
R l = # c ( l ) # ( l )
F l = 2 × P l × R l P l + R l
F 1 m a c r o = 1 | L | l L F l

4.5. Results

We submitted our system’s predictions to the SemEval Task1:E-C challenge. The results were computed by the organizers on a golden test set, for which we did not have access to the golden labels.
Table 2 shows the results of our system and the results of the compared models (obtained from their associated papers). As can be observed from the reported results, our system achieved the top Jaccard index accuracy and macro-averaged F1 scores among all the state-of-the-art systems, with a competitive, but slightly lower score for the micro-average F1.
To get more insight about the performance of our system, we calculated the precision score, the recall score, and the F1 score of each label. The results of this analysis are shown in Figure 4. We found that our system gave the best performance on the “joy” label followed by the “anger”, “fear”, “disgust”, and “optimism” labels. The obtained F1-score of these labels was above 70%. The worst performance was obtained on the “trust”, “surprise”, “anticipation”, and “pessimism” labels. In most cases, our system gave a recall score higher than the precision score. It seems that the system was aggressive against the emotions “trust”, “surprise”, “anticipation”, and “pessimism” (i.e., the system associated a low number of samples to these labels). This can be attributed to the low number of training examples for these emotions and to the Out-Of-Vocabulary (OOV) problem.

4.6. Attention Visualizations

We visualized the attention weights to get a better understanding of the performance of our system. The results are described in Figure 5, Figure 6, Figure 7 and Figure 8, which show heat-maps of the attention weights on the top four example tweets from the validation set. The color intensity refers to the weight given to each word by the attention model. It represents the strength of the relationship between the word and the emotion, which reflects the importance of this word in the final prediction. We can see that the attention model gave the important weights to the common words, such as the stop words, in case the tweet was not assigned to the emotion; for example, the word “for” in Figure 5 and the word “this” in Figure 7 and the token “<user>” in Figure 8. Moreover, it also gives a high weight for the words and the emojis related to emotions (e.g., “cheering” and “awesome” for joy, “birthday” for love, etc.). An interesting observation is that when emojis were present, they were almost always selected as important if they were related to the emotion. For instance, we can see in Figure 7 that the sadness emotion relied heavily on the emoji. We also found that considering only one word to model the relation between the tweet and the emotions was not enough. In some cases, the emotion of a word may be flipped based on the context. For instance, consider the following tweet as an example: “When being #productive (doing the things that NEED to be done), #anxiety level decreases and #love level increases. #personalsexuality”, the word “anxiety” is highly related to the emotion fear, but in this context, it shows optimism and trust emotions. However, our system misassociated this example with the fear emotion.

4.7. Correlation Analysis

Figure 9 shows the correlation analysis of emotion labels in the validation set. Each cell in the figure represents the correlation score of each pair of emotion labels. The reported values show exciting findings. Our system captured the relations among the emotion labels. The correlation scores of the predicted labels were almost identical to the ground-truth. There was an exception in the surprise and trust emotions. Our system was unsuccessful in capturing the relationships between these two emotions and the inputs or the other emotions. We attribute this apparent lack of correlation to the low number of training examples of these two emotions.
Moreover, there was always a positive correlation between related emotions such as “joy” and “optimism” (the score from the ground truth labels and from the predicted labels was 0.74). On the other side, we can see that there was a negative correlation between unlinked emotions like “anger” and “love”. The scores were −0.27 and −0.3, respectively.
This result further strengthened our hypothesis that the proposed system was able to, implicitly, model the relationships between the emotion labels.

5. Conclusions

In this work, we presented a new approach to the multi-label emotion classification task. First, we proposed a transformation method to transform the problem into a single binary classification problem. Afterwards, we developed a deep learning-based system to solve the transformed problem. The key component of our system was the embedding module, which used three embedding models and an attention function. Our system outperformed the state-of-the-art systems, achieving a Jaccard (i.e., multi-label accuracy) score of 0.59 on the challenging SemEval2018 Task 1:E-c multi-label emotion classification problem.
We found that the attention function can model the relationships between the input words and the labels, which helps to improve the system’s performance. Moreover, we showed that our system is interpretable by visualizing the attention weights and analyzing them. However, some limitations have been identified. Our system does not model the relationships between the phrases and the labels. Phrases play a key role in determining the most appropriate set of emotions that must be assigned to a tweet. For instance, an emotion word that reflects “sadness” can be flipped in a negated phrase or context. Thus, in our future work, we plan to work on solving this drawback. One possible solution is to adapt the attention function to model the relationships between different n-gram tokens and labels. Structured attention networks [40] can also be adapted and used to address this issue.
Moreover, we plan to work on developing a non-aggressive system that performs robustly and equally on all the emotion labels by experimenting with different ideas like using data augmentation to enrich the training data or using transfer learning.

Author Contributions

Conceptualization, M.J. and A.M.; methodology, M.J.; software, M.J.; validation, M.J. and A.M.; formal analysis, M.J. and A.M.; investigation, M.J.; resources, M.J.; data curation, M.J.; writing, original draft preparation, M.J.; writing, review and editing, A.M.; visualization, M.J.; supervision, A.M.; funding acquisition, A.M.

Acknowledgments

The authors acknowledge the research support of URV(2018PFR-URV-B2 and Martí i Franqués PhD grant).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Chen, Y.; Zhou, Y.; Zhu, S.; Xu, H. Detecting Offensive Language in Social Media to Protect Adolescent Online Safety. In Proceedings of the 2012 International Conference on Privacy, Security, Risk and Trust (PASSAT), and 2012 International Conference on Social Computing (SocialCom), Amsterdam, The Netherlands, 3–5 September 2012; pp. 71–80. [Google Scholar]
  2. Cherry, C.; Mohammad, S.M.; Bruijn, B. Binary Classifiers and Latent Sequence Models for Emotion Detection in Suicide Notes. Biomed. Inform. Insights 2012, 5, BII-S8933. [Google Scholar] [CrossRef] [PubMed]
  3. Mohammad, S.M.; Zhu, X.; Kiritchenko, S.; Martin, J. Sentiment, Emotion, Purpose, and Style in Electoral Tweets. Inf. Process. Manag. 2015, 51, 480–499. [Google Scholar] [CrossRef]
  4. Cambria, E. Affective computing and sentiment analysis. IEEE Intell. Syst. 2016, 31, 102–107. [Google Scholar] [CrossRef]
  5. Jabreel, M.; Moreno, A.; Huertas, A. Do Local Residents and Visitors Express the Same Sentiments on Destinations Through Social Media? In Information and Communication Technologies in Tourism; Springer: New York, NY, USA, 2017; pp. 655–668. [Google Scholar]
  6. Yun, H.Y.; Lin, P.H.; Lin, R. Emotional Product Design and Perceived Brand Emotion. Int. J. Adv. Psychol. IJAP 2014, 3, 59–66. [Google Scholar] [CrossRef]
  7. Mohammad, S.M. Sentiment Analysis: Detecting Valence, Emotions, and Other Affectual States from Text. In Emotion Measurement; Meiselman, H.L., Ed.; Woodhead Publishing: Cambridge, UK, 2016; pp. 201–237. ISBN 9780081005088. [Google Scholar]
  8. Read, J.; Pfahringer, B.; Holmes, G.; Frank, E. Classifier Chains for Multi-label Classification. Mach. Learn. 2011, 85, 333. [Google Scholar] [CrossRef]
  9. Robert, P. Emotions: A general Psychoevolutionary Theory. In Approaches to Emotion; Scherer, K.R., Ekman, P., Eds.; Psychology Press: London, UK, 2014; pp. 197–219. [Google Scholar]
  10. Tsoumakas, G.; Katakis, I. Multi-label Classification: An Overview. Int. J. Data Warehous. Min. IJDWM 2007, 3, 1–13. [Google Scholar] [CrossRef]
  11. Read, J. Scalable Multi-label Classification. Ph.D. Thesis, University of Waikato, Hamilton, New Zealand, 2010. [Google Scholar]
  12. Mikolov, T.; Chen, K.; Corrado, G.; Dean, J. Efficient Estimation of Word Representations in Vector Space. arXiv, 2013; arXiv:1301.3781. [Google Scholar]
  13. Schuster, M.; Paliwal, K.K. Bidirectional Recurrent Neural Networks. IEEE Trans. Signal Process. 1997, 45, 2673–2681. [Google Scholar] [CrossRef]
  14. Chung, J.; Gulcehre, C.; Cho, K.; Bengio, Y. Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling. arXiv, 2014; arXiv:1412.3555. [Google Scholar]
  15. Al-Molegi, A.; Jabreel, M.; Martínez-Ballesté, A. Move, Attend and Predict: An attention-based neural model for people’s movement prediction. Pattern Recognit. Lett. 2018, 112, 34–40. [Google Scholar] [CrossRef]
  16. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is All You Need. In Proceedings of the 31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017; pp. 5998–6008. [Google Scholar]
  17. Zhang, M.L.; Zhou, Z.H. A review on Multi-label Learning Algorithms. IEEE Trans. Knowl. Data Eng. 2014, 26, 1819–1837. [Google Scholar] [CrossRef]
  18. Zhang, M.L.; Zhang, K. Multi-label Learning by Exploiting Label Dependency. In Proceedings of the 16th ACM International Conference on Knowledge Discovery and Data Mining (SIGKDD), Washington, DC, USA, 25–28 July 2010; pp. 999–1008. [Google Scholar]
  19. Schapire, R.E.; Singer, Y. BoosTexter: A boosting-Based System for Text Categorization. Mach. Learn. 2000, 39, 135–168. [Google Scholar] [CrossRef]
  20. Zhang, M.L.; Zhou, Z.H. ML-KNN: A Lazy Learning Approach to Multi-label Learning. Pattern Recognit. 2007, 40, 2038–2048. [Google Scholar] [CrossRef]
  21. Clare, A.; King, R.D. Knowledge Discovery in Multi-label Phenotype Data. In Proceedings of the European Conference on Principles of Data Mining and Knowledge Discovery, Freiburg, Germany, 3–5 September 2001; pp. 42–53. [Google Scholar]
  22. De Comite, F.; Gilleron, R.; Tommasi, M. Learning Multi-label Alternating Decision Trees from Texts and Data. In Proceedings of the 3rd International Conference on Machine Learning and Data Mining in Pattern Recognition, Leipzig, Germany, 5–7 July 2003; Volume 2734, pp. 35–49. [Google Scholar]
  23. Mencia, E.L.; Fürnkranz, J. Efficient Pairwise Multilabel Classification for Large-Scale Problems in the Legal Domain. In Proceedings of the Joint European Conference on Machine Learning and Knowledge Discovery in Databases, Antwerp, Belgium, 15–19 September 2008; Springer: Berlin/Heidelberg, Germany, 2008; pp. 50–65. [Google Scholar]
  24. Cheng, W.; Hüllermeier, E. Combining Instance-Based Learning and Logistic Regression for Multilabel Classification. Mach. Learn. 2009, 76, 211–225. [Google Scholar] [CrossRef]
  25. Godbole, S.; Sarawagi, S. Discriminative Methods for Multi-labeled Classification. In Proceedings of the Pacific-Asia Conference on Knowledge Discovery and Data Mining, Osaka, Japan, 20–23 May 2008; Springer: Berlin/Heidelberg, Germany, 2008; pp. 22–30. [Google Scholar]
  26. Younes, Z.; Abdallah, F.; Denoeux, T.; Snoussi, H. A Dependent Multilabel Classification Method Derived From the k-Nearest Neighbor Rule. J. Adv. Signal Process. 2011, 1, 645964. [Google Scholar] [CrossRef]
  27. Yan, R.; Tesic, J.; Smith, J.R. Model-Shared Subspace Boosting for Multi-label Classification. In Proceedings of the 13th ACM SIGKDD, San Jose, CA, USA, 12–15 August 2007; pp. 834–843. [Google Scholar]
  28. Jabreel, M.; Moreno, A. SentiRich: Sentiment Analysis of Tweets Based on a Rich Set of Features. In Artificial Intelligence Research and Development; Nebot, Á., Binefa, X., López de Mántaras, R., Eds.; IOS Press: Amsterdam, The Netherlands, 2016; Volume 288, pp. 137–146. ISBN 978-1-61499-695-8. [Google Scholar]
  29. Jabreel, M.; Moreno, A. SiTAKA at SemEval-2017 Task 4: Sentiment Analysis in Twitter Based on a Rich Set of Features. In Proceedings of the 11th International Workshop on Semantic Evaluation (SemEval-2017), Vancouver, BC, Canada, 3–4 August 2017; pp. 694–699. [Google Scholar]
  30. Mohammed, S.M.; Bravo-Marquez, F.; Salameh, M.; Kiritchenko, S. Semeval-2018 task 1: Affect in Tweets. In Proceedings of the 12th International Workshop on Semantic Evaluation, New Orleans, LA, USA, 5–6 June 2018; pp. 1–17. [Google Scholar]
  31. LeCun, Y.; Bengio, Y.; Hinton, G. Deep Learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  32. Tang, D.; Qin, B.; Liu, T. Deep Learning for Sentiment Analysis: Successful Approaches and Future Challenges. Wiley Interdiscip. Rev. Data Min. Knowl. Discov. 2015, 5, 292–303. [Google Scholar] [CrossRef]
  33. Baziotis, C.; Athanasiou, N.; Chronopoulou, A.; Kolovou, A.; Paraskevopoulos, G.; Ellinas, N.; Narayanan, S.; Potamianos, A. NTUA-SLP at SemEval-2018 Task 1: Predicting Affective Content in Tweets with Deep Attentive RNNs and Transfer Learning. In Proceedings of the 12th International Workshop on Semantic Evaluation, New Orleans, LA, USA, 5–6 June 2018; pp. 245–255. [Google Scholar]
  34. Meisheri, H.; Dey, L. TCS Research at Semeval2018 Task 1: Learning Robust Representations using Multi-Attention Architecture. In Proceedings of the 12th International Workshop on Semantic Evaluation, New Orleans, LA, USA, 5–6 June 2018; pp. 291–299. [Google Scholar]
  35. Park, J.H.; Xu, P.; Fung, P. PlusEmo2Vec at SemEval-2018 Task 1: Exploiting Emotion Knowledge from Emoji and #hashtags. In Proceedings of the 12th International Workshop on Semantic Evaluation, New Orleans, LA, USA, 5–6 June 2018; pp. 264–272. [Google Scholar]
  36. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. arXiv, 2014; arXiv:1412.6980. [Google Scholar]
  37. James, B.; Yamins, D.; Cox, D.D. Hyperopt: A python library for optimizing the hyperparameters of machine learning algorithms. In Proceedings of the 12th Python in Science Conference, Austin, TX, USA, 24–29 June 2013. [Google Scholar]
  38. Mohammad, S.; Kiritchenko, S. Understanding emotions: A dataset of tweets to study interactions between affect categories. In Proceedings of the Eleventh International Conference on Language Resources and Evaluation, Miyazaki, Japan, 7–12 May 2018. [Google Scholar]
  39. Kant, N.; Puri, R.; Yakovenko, N.; Catanzaro, B. Practical Text Classification With Large Pre-Trained Language Models. arXiv, 2018; arXiv:1812.01207. [Google Scholar]
  40. Kim, Y.; Denton, C.; Hoang, L.; Rush, A.M. Structured Attention Networks. arXiv, 2017; arXiv:1702.00887. [Google Scholar]
Figure 1. The set of the eight basic emotions proposed by Plutchik [9].
Figure 1. The set of the eight basic emotions proposed by Plutchik [9].
Applsci 09 01123 g001
Figure 2. An illustration of the proposed system. Shaded parts are trainable. Fw and Bw refer to the Forward and Backward cells respectively, and FF means Feed-Forward layer.
Figure 2. An illustration of the proposed system. Shaded parts are trainable. Fw and Bw refer to the Forward and Backward cells respectively, and FF means Feed-Forward layer.
Applsci 09 01123 g002
Figure 3. Gated Recurrent Unit (GRU).
Figure 3. Gated Recurrent Unit (GRU).
Applsci 09 01123 g003
Figure 4. Performance analysis.
Figure 4. Performance analysis.
Applsci 09 01123 g004
Figure 5. Attention visualization example. Golden labels are {joy, optimism} and predicted labels are {joy (0.91), optimism (0.51)}.
Figure 5. Attention visualization example. Golden labels are {joy, optimism} and predicted labels are {joy (0.91), optimism (0.51)}.
Applsci 09 01123 g005
Figure 6. Attention visualization example. Golden labels are {joy, surprise} and predicted labels are {joy (0.97), optimism (0.87)}.
Figure 6. Attention visualization example. Golden labels are {joy, surprise} and predicted labels are {joy (0.97), optimism (0.87)}.
Applsci 09 01123 g006
Figure 7. Attention visualization example. Golden labels are {sadness, surprise} and predicted labels are {love (0.74), sadness (0.98)}.
Figure 7. Attention visualization example. Golden labels are {sadness, surprise} and predicted labels are {love (0.74), sadness (0.98)}.
Applsci 09 01123 g007
Figure 8. Attention visualization example. Golden labels are {joy, love, optimism} and predicted labels are {joy (0.98), love (0.91) optimism (0.95)}.
Figure 8. Attention visualization example. Golden labels are {joy, love, optimism} and predicted labels are {joy (0.98), love (0.91) optimism (0.95)}.
Applsci 09 01123 g008
Figure 9. Correlation matrices of emotion labels of the development set.
Figure 9. Correlation matrices of emotion labels of the development set.
Applsci 09 01123 g009
Table 1. Hyperparameters of our system.
Table 1. Hyperparameters of our system.
ParameterValue
Embedding Module E Q :
Dimensions: 11 × 310
Initialization: Uniform (−0.02, 0.02)
Trainable: Yes
E K :
Dimensions: 13,249 × 310
Initialization: Pretrained model (we used the pretrained embeddings provided in [33]).
Trainable: No
E V :
Dimensions: 13,249 × 310
Initialization: Uniform (−0.02, 0.02)
Trainable: Yes
Encoding ModuleRNN Cell: GRU
Hidden size: 200
Layers: 2
Encoding: last hidden state
RNN dropout: 0.3
Classification ModuleFF1: 1024 units
FF2: 512 units
Sigmoid: 1 unit
Activation: ReLU
Dropout: 0.3
Table 2. Results of our system and state-of-the-art systems. The best values are in bold.
Table 2. Results of our system and state-of-the-art systems. The best values are in bold.
ModelAccuracy (Jaccard)Micro F1Macro F1
BNet(Our System)0.5900.6920.564
SVM-Unigrams0.4420.570.443
Transformer0.5770.6900.561
NTUA-SLP0.5880.7010.528
TCS0.5820.6930.530
PlusEmo2Vec0.5760.6920.497

Share and Cite

MDPI and ACS Style

Jabreel, M.; Moreno, A. A Deep Learning-Based Approach for Multi-Label Emotion Classification in Tweets. Appl. Sci. 2019, 9, 1123. https://doi.org/10.3390/app9061123

AMA Style

Jabreel M, Moreno A. A Deep Learning-Based Approach for Multi-Label Emotion Classification in Tweets. Applied Sciences. 2019; 9(6):1123. https://doi.org/10.3390/app9061123

Chicago/Turabian Style

Jabreel, Mohammed, and Antonio Moreno. 2019. "A Deep Learning-Based Approach for Multi-Label Emotion Classification in Tweets" Applied Sciences 9, no. 6: 1123. https://doi.org/10.3390/app9061123

APA Style

Jabreel, M., & Moreno, A. (2019). A Deep Learning-Based Approach for Multi-Label Emotion Classification in Tweets. Applied Sciences, 9(6), 1123. https://doi.org/10.3390/app9061123

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop