Next Article in Journal
Wearable Devices in Health Monitoring from the Environmental towards Multiple Domains: A Survey
Next Article in Special Issue
Evolutionary Mahalanobis Distance-Based Oversampling for Multi-Class Imbalanced Data Classification
Previous Article in Journal
The Wildland Fire Heat Budget—Using Bi-Directional Probes to Measure Sensible Heat Flux and Energy in Surface Fires
Previous Article in Special Issue
Unsupervised Trademark Retrieval Method Based on Attention Mechanism
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Level Context Pyramid Network for Visual Sentiment Analysis

School of Electronic and Information Engineering, South China University of Technology, Guangzhou 510640, China
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(6), 2136; https://doi.org/10.3390/s21062136
Submission received: 26 January 2021 / Revised: 6 March 2021 / Accepted: 15 March 2021 / Published: 18 March 2021

Abstract

:
Sharing our feelings through content with images and short videos is one main way of expression on social networks. Visual content can affect people’s emotions, which makes the task of analyzing the sentimental information of visual content more and more concerned. Most of the current methods focus on how to improve the local emotional representations to get better performance of sentiment analysis and ignore the problem of how to perceive objects of different scales and different emotional intensity in complex scenes. In this paper, based on the alterable scale and multi-level local regional emotional affinity analysis under the global perspective, we propose a multi-level context pyramid network (MCPNet) for visual sentiment analysis by combining local and global representations to improve the classification performance. Firstly, Resnet101 is employed as backbone to obtain multi-level emotional representation representing different degrees of semantic information and detailed information. Next, the multi-scale adaptive context modules (MACM) are proposed to learn the sentiment correlation degree of different regions for different scale in the image, and to extract the multi-scale context features for each level deep representation. Finally, different levels of context features are combined to obtain the multi-cue sentimental feature for image sentiment classification. Extensive experimental results on seven commonly used visual sentiment datasets illustrate that our method outperforms the state-of-the-art methods, especially the accuracy on the FI dataset exceeds 90%.

1. Introduction

Studies have shown that image sentiment affect visual perception [1]. Compared with the non-emotional stimulus content in the image, the affective content attracts the attention of the viewer more strongly, and the viewer has a more detailed understanding of the affective stimulus content [2]. Therefore, the purpose of visual sentiment analysis is to understand the emotional impact of visual materials on viewers [3], which plays an important role in opinion mining, user behavior prediction, emotional image retrieval, game scene modeling and other aspects.
Initially inspired by psychological and artistic principles, researchers studied color, texture and other hand-crafted features at the image level for visual sentiment analysis. Like many visual tasks, gradually, Convolutional Neural Networks (CNN) replaces the hand-crafted features because it can automatically learn the deeper representation of images, and the actual research also proves that the CNN-based method is obviously superior to the hand-crafted features.
However, visual sentiment analysis is inherently more challenging than traditional visual tasks (such as object classification, scene recognition, etc.) [4], mainly because it involves more complex semantic abstraction and subjectivity [5]. Existing works illustrate that the viewer’s emotional changes are often caused by certain areas of the image [6] because of the selective attention [7], so many of recent studies have proposed different methods on how to use local area representation to improve classification accuracy [4,8]. However, most of the methods extract local features from the final feature map of backbone, so that the obtained features are single-level and can perceive objects of extremely limited scale. Based on the co-existence of sentiment evoked and the objects in local region [9,10], visual sentiment analysis needs to solve the following two basic problems more than many traditional visual tasks in image content understanding [11,12]:
  • Perception of different scale objects. The size and location of objects in images from social networks are diverse, which means that we need to have multi-scale analysis capabilities in the model. As shown in Figure 1, the scale of people in the images (a) to (c) is from small to large. Thus, a single-scale method object perception can only capture objects at a limited scale, while losing object information at other scales.
  • Different levels of emotional representation. Different objects can evoke dissimilar degrees of sentiment. As shown in Figure 1e–h. Some simple objects contain less semantic information, such as the flower (e), and the street lamp (f). The emotional stimulation they express is weak, and their emotional information can be described by the low-level features. However, the complex semantic information will give us stronger emotional stimulation such as humans and human-related objects. Human’s non-verbal communication such as facial expression (g), body language and posture (h), have a strong ability to express emotions [13]. These complex objects need more abstract high-level semantic features to describe their emotional information.
In addition, the viewer’s attention is affected by different levels of features, such as low-level attributes (for example, intensity, color) and high-level semantic information [14]. Figure 1d is also a person holding an umbrella, but due to the different colors of the umbrella, our focus is different from Figure 1b,c. Therefore, the methods only use the final high-level semantic features of the model are insufficient.
In this paper, a multi-level context pyramid network (MCPNet) composed of multi-scale adaptive context modules (MACM) is proposed to deal with the above two problems at the same time. Multi-scale, adaptive context modules combined with different levels of features are presented to capture more different cues to improve model performance. Moreover, compared with the current local area analysis method that does not consider the relationship between different areas, our method can adaptively learn the relationship between different regions in the image to obtain the areas related to sentiment.
In summary, the contributions of this paper can be highlighted as follows:
  • Adaptive context framework is introduced for the first time in the image sentiment analysis task. This method can learn the correlation degree of different regions in the image by combining different scale representations, which is helpful to improve the ability of the model to understand complex scenarios.
  • The multi-scale attributes in the proposed MACM module are alterable. Compared with many existing multi-scale methods that can only capture objects of fixed limited scales, our method can combine different scales to capture objects of different positions and sizes in the image.
  • The proposed MPCNet adopts cross-layer and multi-layer feature fusion strategies to enhance the ability of the model to perceive semantic objects at different levels.
  • The experiment proves the advancement of our method, and the visualization results show that our method can effectively identify the small semantic objects related to emotional expression in complex scenarios.

2. Related Work

In this section, we will review the CNN method with additional information and the region-based CNN method. The difference between the two is that the CNN model of the former is based on the existing excellent image classification model, while the model of the latter is designed for visual sentiment analysis. Then we will introduce the context-based approaches that are closely related to our method.

2.1. CNN with Additional Information

The hand-crafted features have limitations, while the features extracted by excellent CNN show better results than the hand-crafted features on different datasets [9,15]. On this basis, researchers began to introduce more additional information to help CNN model improve performance. Borth et al. [16] proposed a form of multiple adjective noun pairs to describe image content, called ANPs; Chen et al. [17] further proposed a classification CNN called DeepSentiBank based on visual sentiment concepts. Li et al. [18] further combined the text information in ANPs and calculated the emotional information of the text value in ANPs in the form of weighted sum. Yuan et al. [19] proposed an algorithm called Sentribute with 102 mid-level attributes that is readily comprehensible and can be used for higher-level sentiment analysis. Yang et al. [20] introduced the probability distribution constraint of sentiment in the loss calculation of single label classification task. Kim et al. [21], Ali et al. [22] used the excellent model of target recognition and scene classification competition to generate objects and scene category labels as high-level semantic features for image sentiment classification.
The additional information of the above methods mainly comes from the annotation information. The cost of introducing additional annotation information is expensive, and thus this approach is not suitable for large-scale datasets. Because visual sentiment analysis is more difficult than general visual tasks, it is necessary for us to design corresponding models or modules specifically for visual sentiment analysis tasks.

2.2. Region-Based CNN

Studies have shown that the expression of human emotion is related to the region of local concern in the image, and different regions may have different effects on the expected expression [1,23]. Recently, local region-based image sentiment analysis methods have achieved encouraging performance improvements on many image sentiment datasets. Zhou et al. [24] showed the potential of CNN in local learning. Peng et al. [23] proposed that sentiment can be induced by specific regions, and proposed EmotionROI dataset. Fan et al. [1] proposed an affective priority effect through comparative experiments. Yang et al. [25] proposed WSCNet, which weights the final output features of the network by cross spatial pooling module and integrates local image features into classification. Song et al. [8] integrated the concept of visual attention into visual sentiment classification, using multi-layer CNN to model the attention distribution of the whole image which constraining by saliency detection, and located the region with the largest amount of information to improve the classification performance. On the basis of NASNet, Yadav et al. applied the residual attention module to learn the important areas related to emotion in the image [26]. Wu et al. suggested to use the object detection module to determine if use a local module [27]. Yang et al. [28] proposed a method to finding related regions using a ready-made method to get object proposals as local emotional information and used VGG to learn global information. According to Rao (c) [4], most of the current methods only consider the use of the final level features, ignoring the contribution of different levels of features to visual sentiment expression, which limits their model performance. Therefore, based on Faster R-CNN [29], they further improved the model performance by combining different levels features and the global information. Their experimental results show that they are currently the best results.

2.3. Context-Based CNN

Most of current methods try to obtain the emotional local area of the image, and treat these local areas independently, without considering whether these local areas related to emotion are connected. This is not conducive to the analysis of complex scenarios. Contextual information can help the model understand complex scenarios [12] and are widely used in scene parsing and semantic segmentation tasks. On the basis of existing work, APCNet [12] further summarizes the role of the multi-scale, adaptive and global information guidance context features in understanding complex scenes. A single scale can only capture objects of single scale, and there is information loss on other scales. Therefore, it is necessary to use multi-scale in most of visual tasks. The attribute of adaptive makes a pixel in the context feature not only associate with the pixel nearby, but also infer the relationship between the pixel and other regions in the global context, which can capture the long-distance dependences between the pixel and the regions. Therefore, the adaptive attributes can help us to obtain the relationship between different regions.
Inspired by APCNet, we introduce multi-scale adaptive context features into visual sentiment analysis tasks for the first time. The multi-level features can combine the details and the semantic information from different level features. The low-level features can describe the simple-looking objects, and the high-level features are more suitable for describing complex-looking objects.

3. Methodology

In this section, we will introduce the proposed multi-level context pyramid network (MCPNet) in detail. The framework is illustrated in Figure 2, which includes multi-scale adaptive context modules (MACM) from different levels. The proposed MACM can learn the degree of association of different regions in the image at different scales for corresponding level. In order to make better use of different levels of features, we use the cross-layer and multi-layer feature fusion strategies to combine different levels of features from MACM to obtain multi-cue sentimental feature for classification.

3.1. Proposed Multi-Level Context Pyramid Network

The architecture of proposed MCPNet is illustrated in Figure 2. The ResNet101 is utilized as backbone, because it is a commonly used and effective model in visual tasks, and ResNet101 has enough convolutional layers to help us extract multi-level features. As shown in Figure 2, the ResNet101 network can be divided into four parts: c 2 , c 3 , c 4 , and c 5 , representing different levels of features. Considering the amount of computation, we only take the features of the output of c 3 , c 4 , and c 5 . Among features at different levels, low-level features have more detailed information, but weaker semantics representation and more noise. High-level features can represent more complex semantic objects but lacking in detail. For c 3 , c 4 , and c 5 , different level features are sent to the corresponding MACM to obtain the corresponding cross-layer multi-scale context features called O 3 , O 4 , O 5 as the local emotional representation. For different levels of contextual features, a cross-layer feature fusion strategy is adopted to increase the relationship between adjacent levels of features. Then multi-level contextual features are combined with global emotional representation to form the multi-cue emotional feature E for sentiment classification. In Section 4.6, we will further verify the effectiveness of the multi-cue emotional feature E through visualization.

3.2. Multi-Scale Adaptive Context Module

MACM is the core module of the context pyramid network, which aims to calculate the contextual features of each location by using the local region sentiment affinity coefficient of different scales under global guidance. The process of MACM is illustrated in Figure 3. For the input I , X l is the feature map from the layer l of backbone, and X i l is the representation at the position of feature i in X l . The value of l is c 3 , c 4 , or c 5 . In order to get the context features adaptively with different scales on X l , Z i l is introduced to represent the multi-scale context feature of X i l :
z i l = F c o n t e x t l ( X l , i )
MACM consists of two branches. The following shows how MACM works through one layer of feature X l .
(1) Sub-regions Branch: This branch is to learn the local sub-region representations of the input feature map X l under different scale divisions. For each scale S k , the feature map X l is divided into S k × S k sub-regions Y s k :
Y s k   =   [ Y 1 s k , Y 2 s k , , Y s k × s k s k ] , k = 1 , 2 . . n
For each sub region Y j s k , the feature will be extracted by averaging pooling and 1 × 1 convolution. In implementation, the feature of Y s k R s k × s k × 512 under the scale S k is extracted through an adaptive average pooling with a 1 × 1   convolution. Then, Y s k will be reshaped to Y s k R s k 2 × 512 to match the shape of the affinity coefficient α s k of the corresponding scale calculated by the other branch.
(2) Region Sentiment Affinity Coefficient Branch: The purpose of this branch is to learn the affinity coefficient weights between sub-regions at the same scales under the guidance of global information. The region sentiment affinity coefficient α i , j s k is introduced to represent the degree of association of the sub-region Y j s k with the sentiment of the estimated X i l . In order to realize α i , j s k local feature association property, X l firstly will be sent to the resize operation to get M l R 14 × 14 × 512 . The resize operation is realized by controlling the parameters of convolution kernel, as showing in Figure 4. There are 1 × 1 and 3 × 3 convolution kernels in this operation. After 1 × 1 convolution, the length and width of the feature map remain unchanged, and the number of channels is 512; after the convolution of 3 × 3 , the channel becomes 512, and the length and width become half of the original.
After that, by using global average pooling for M l , the global information characterization g ( M l ) is obtained. Then, M l and g ( M l ) are multiplied and calculate the local affinity vector for each local location i under the global perspective:
α i , j s k   =   f s k ( M i l , g ( M l ) , j )
In implementation, f s k is achieved by a 1 × 1 convolution and sigmoid activation function. For each location i , the affinity vector corresponding to each scale is s k × s k , which corresponds to the number of sub-regions in this scale. So, it has a total of h × w affinity vectors, each of which has a length of s k 2 and reshaped it to the size of h w × s k 2 .
(3) Computing multi-scale context feature Z: For each scale s k , the region sentiment affinity coefficient α i , j s k indicates the degree of relevance of the sub-region Y j s k to the sentiment of the estimated X i l . The single-scale adaptive context vector W i s k of position i for scale s k can be calculated as
W i s k =   j = 1 s k × s k α i , j s k Y j s k
For the context feature W s 1 , W s 2 , , W s n obtained by different scale calculation sequence, context features of different scales will be combined, and then do batch normalization (BN) after a 1 × 1 convolution. Finally, the multi-scale context feature Z i l for X i l will be obtained. The details are as follows:
Z i l =   f ( [ W i s 1 , W i s 2 , , W i s n ] )
Z l represents the multi-scale adaptive context feature of the l layer, and through cross-layer feature fusion, the multi-scale emotional representation of l layer is obtained.

3.3. Cross-Layer and Multi-Layer Feature Fusion Strategies

In order to enhance the connection between features of different depths, here two fusion strategies are adopted to balance semantic information and detailed information.
(1) Cross-layer features fusion: Among the features of different levels, the semantic and detailed information of the features of two adjacent levels is the closest. The idea of fusion of features of two adjacent layers has been reflected in Feature Pyramid Networks (FPN). However, unlike FPN, the shape of the context feature O l output by each layer of MACM is 14 × 14 × 512, it can’t use an up-sampling method for feature fusion between adjacent layers. To reduce the noise from low-level features, we have adopted the same shape of M l + 1 and O l for feature fusion. There are different numbers of 3 × 3 convolutions using to transform the feature maps of different layers to obtain M R 14 × 14 × 512 . In each layer, multiple scales are used to capture semantic information of different sizes and positions. In the network, we adopt three scales of 1, 2, and 4. In particular, when s   =1, α i , j s k represents the global weight of each location, which is the feature learning under the global perspective. Then, the context features of all scales are fused to obtain the multi-scale context feature Z l of this layer. Then for different layers, adding context information Z l obtained by the MACM module of this layer and the M l + 1 obtained from the previous layer to get the output feature O l of this layer. Details as follow:
{ O i l = Z i l × M i l   ,   l = c 5       O i l = Z i l × M i l + 1   ,   l c 5
After cross-layer feature fusion, the shapes of the obtained O are the same, which is conducive to further multi-level feature fusion.
(2) Multi-layer features fusion: To increase the global and local connections, we also combine the global information of the input image and the context features O 3 , O 4 , O 5 obtained from layers c 3 , c 4 , and c 5 . Specifically, the last output X 5 R 14 × 14 × 2048 of backbone and O 3 , O 4 , O 5 are concatenated together to get the multi-cue sentimental feature E :
E = [ X g l o b a l , O ] = [ X 5 , O 3 , O 4 , O 5 ]
Then, after global average pooling, it enters the full connection layer for classification.

4. Experiment

4.1. Dataset

The proposed method is evaluated on seven widely used datasets of different sizes, including IAPSsubset [30], ArtPhoto [31], Abstract Paintings [31], Twitter I [32], Twitter II [16], EmotionROI [23] and FI [15]. Some examples as shown in Figure 5 and the details are illustrated in Table 1.
  • Small-scale dataset: IAPSsubset comes from International Affective Picture System [33] and has 395 images in eight emotional categories. Different from other datasets, ArtPhoto is a dataset composed of 806 art photos, and Abstract Paintings consists of 228 abstract pictures of colors and textures. Twitter I and Twitter II were collected from social network Twitter and labeled by AMT workers with 1,269 and 603 images. We set up experiments similar to [28,32] on all three subsets of Twitter I. EmotionROI is developed from Emotion6 [34], and the image comes from the Flickr website. Compared to Emotion6, EmotionROI adds 15 emotion-related annotation boxes to each image marked by participants and believes that the more repeated annotation boxes on a pixel point, the greater the contribution of the point to emotional expression.
  • Large-scale dataset: FI is currently the most commonly used large-scale visual sentiment dataset, which is collected through social network using emotional categories as search keywords. 225 participants from AMT were employed to label resulting in 23,308 images.
In visual sentiment analysis tasks, different labeling methods are used, and the number of categories in the dataset is different. At present, there is no uniformity in the number of dataset categories in visual sentiment analysis tasks. Due to the influence of subjectivity, the dataset of visual sentiment analysis task requires expensive manual annotation [35]. The six used affective datasets contain less than two thousand images, except the FI dataset (see Table 1), which are far from the required number for training robust deep networks. Therefore, in this paper, we focus on binary emotion prediction (positive and negative) and convert the emotional labels of Mikel [30] and Ekman [36] into the original binary affective tags, which are compared with existing advanced methods in the above datasets.

4.2. Implementation Details

The proposed MCPNet uses ResNet-101 pre-trained on ImageNet [37] as the backbone. Before training, random horizontal flipping and clipping random 448 × 448 patches are used as data augmentation to reduce over fitting. We use the SGD optimizer and the momentum is 0.9. The learning rate is set to 0.001 and decreased 10 times every 7 epochs of 100 epochs. FI dataset was randomly divided into 80% for training, 5% for validation and 15% for testing. Other datasets were randomly divided into 80% training and 20% testing [16], except for datasets with specifying training/test partition [23,31]. Further proves the effectiveness of our method through 5-fold cross validation experiment in the above 7 datasets. All our experiments were carried out on one NVIDIA GPU and completed all codes and experiments through Pytorch.

4.3. Baseline

In the following, we will evaluate our methods with the state-of-the-art algorithms of image sentiment classification, including the hand-crafted features and deep learning methods.

4.3.1. Hand-Crafted Features

GCH [38]: the global view of image with 64-bin color histogram features.
LCH [38]: the local view of image with 64-bin color histogram features.
PAEF [39]: This research is one of the early works in visual sentiment analysis that focuses on more complex features than the low features. It contains low-level and middle-level features inspired by artistic principles.
Rao(a) [40]: It is an early exploration of analyzing local areas related to sentiment. Dividing a picture into different blocks through image segmentation, called multi-scale blocks, and SIFT-based bag-of-visual features contain local and global information extracted from the image blocks.
SentiBank [16]: Proposing a 1200-dimensional middle-level feature called adjective noun pairs (ANPs) to describe the relationship between image content and sentiment. This work is an important work that explored the correspondence between semantic information and emotion in the early stage.

4.3.2. Features Based on CNN

AlexNet [41]/ VGG-16 [42]/ ResNet101 [37]: pre-trained on ImageNet, and then fine-tuned the FC layer on the sentiment dataset.
DeepSentiBank [17]: This is a work to improve SentiBank and propose a CNN called DeepSentiBank. Compared with SentiBank, this work proposes 2089-dimensional ANPs, which significantly improves labeling accuracy and retrieval performance.
PCNN [32]: a progressive training framework based on VGGNet. They use a large amount of weakly supervised data to let the model learn some common visual features to reduce the difficulty of training the visual emotion dataset.
Rao(b) [9]: multi-level deep features from the framework based on AlexNet with side branch. This is an important exploration of the emotional analysis method using CNN for multi-level features.
Zhu [43]: a CNN framework containing a bidirectional RNN module with multi-task losses for visual emotion recognition.
AR [28]: This work puts forward a new concept called Affective Regions in the research of exploring local regions and sentiment evoked and use the ready-made object detection technology as local information combined with VGG model for analysis.
RA-DLNet [26]: The residual attention module is applied in NASNET to focus on the local areas of the image which are related to emotion.
GMEI &LRMSI [27]: The work found through research that not all images in the dataset contain salient objects. Therefore, this work believes that visual sentiment analysis should not only focus on local features. The work respectively proposed global modules and local modules, and determine whether to use a local module through the object detection module.
Rao(c) [4]: This work argues that most of the current researches on visual emotion analysis only focus on high level features and do not pay attention to the effects of different level features on visual emotion tasks. They use the multi-level framework of object detection algorithms to obtain the local representations that trigger emotions, are combined with the output features of the last layer of backbone to classify image emotions as global features.

4.4. Experimental Validation

4.4.1. Choice of Scale s

In the proposed MCPNet, as shown in Figure 3, different s represents different scales used by the model, and we give the corresponding classification performance on the FI dataset. For feature maps at different levels, we unify them into one size. Higher-level feature has larger perception fields. It is not that the larger the scale, the better the performance. The performance of different s by taking it from 1 to 6 is shown in Figure 6. When s = 1, 2, 4, the model achieves the top three performance respectively. Therefore, our experimental setup adopts s = 1, 2, 4. The results of subsequent experiments with different datasets also show that our selection of s is equally effective on other datasets. Of course, for different datasets, we can better adapt to different task by choosing different s combinations. However, here we want to use a more general combination of s .

4.4.2. Effective of Different Level Features

Low-level and high-level features have different effects on sentiment classification How to combine the two is a consideration for many visual tasks. To provide more cues to the task of visual sentiment analysis, previous studies have shown the effectiveness of using multi-level features in experimental results [4,9]. Here, we further explore the role of different levels of features. As can be seen from Table 2, O 5 represents the context of high-level features, with the largest impact on the model. When O 5 is removed and the model performance drops by almost 2%. O 3 denotes the context of the underlying features and has the smallest impact. One interesting finding, however, is that removing O 5 or O 4 has little difference. We think this is due to more than half of the backbone convolution being performed at the c 4 level. Because we just remove O 5 or O 4 , not c 4 or c 5 . Therefore, the model accuracy of using only O 5 or O 4 is still 89.011 and 88.668, which illustrates that both of them have a significant contribution to model performance. The accuracy was declined almost 10 percentage points by only using O 3 . This indicates that the model performance can be improved in varying degrees by using different level of features.
Here it also showed the effect of removing global information and the performance of the model decreased by 0.52% on FI. This indicates that it is necessary to combine global information with the context features. We also performed comparative experiments on different levels of context feature fusion strategies. When no cross-layer fusion strategy was used, the model decreased by 0.3%. Further performance changes are shown in Figure 7 and Figure 8. They show that the contributions of O 4 and O 5 to the model performance are close, but the effect of O 3 is much lower than that of O 4 and O 5 . In this regard, it can be seen that on the large dataset such as FI, it is not enough to only use the low-level features.

4.5. Comparisons with State-of-the-Art Methods

This section shows the comparison between the proposed MCPNet and the state-of-the-art method on the seven datasets. The experimental settings are mainly referred to by Rao(c). To more fully reflect the performance of our methods, all of methods are used five-fold cross-validation to get the final classification score.
As shown in Table 3, the results illustrate that the performance of the CNN methods far exceeds the hand-crafted methods. Hand-crafted features only perform well in some small datasets such as Abstract, ArtPhoto, and EmotionROI. Mainly because the Abstract is composed of abstract paintings and ArtPhoto is composed of art photos. They are greatly affected by the low-level features like color and texture. EmotionROI actively eliminates images that evoke sentiment with some high-level semantic such as facial expressions and text during the collection process. Therefore, some hand-crafted methods perform better than some deep learning methods on these three datasets. For example, the Rao(a) method is nearly 1.2% higher than ResNet01 on Abstract. However, the hand-crafted method is far inferior to the CNN method on datasets with more high-level semantic information such as IAPSsubset and the large dataset FI. With fewer layers, AlexNet is still nearly 6% higher than Rao(a) in both FI and IAPSsubset. Region-based methods such as AR and Rao(c) have significantly higher performance than global image-based methods. Our method mines sub-region relations and their long-distance dependencies in images, which have multi-scale characteristics.
Rao(c)’s method [3] obtains local areas of different scales through faster R-CNN based on FPN which requires a tedious pre-training process that the entire model needs to be pre-trained on the COCO [44] and EmotionROI datasets to obtain object perception capabilities. However, our model does not require other additional cumbersome pre-training processes. In our method, the feature map of the input MACM module is divided into different sub-regions, and the relationship between different regions is learned to obtain different scale perception abilities. Compared with methods such as Rao(c), our method is 2.8% higher on FI and 2.16% higher on EmotionROI. Our performance is also better than Rao (c) on other small datasets as shown in Table 3. To our knowledge, it is the first time that the accuracy of binary classification task over 90% on FI dataset. FI dataset is a large dataset, the others are small datasets and they rely on the pre-trained model on FI to continue training. Therefore, the proposed MCPNet has the better improvement effect on the FI dataset then other dataset. Figure 9 shows the accuracy of our method on the 5-fold cross-validation strategy, which is trained five times on the FI and EMotionROI. The dashed line represents the accuracy of Rao(c). Thus, our method is better than the existing methods in binary classification performance, which demonstrates the effectiveness of adaptive context features in visual sentiment analysis tasks.
Table 4 shows the classification results on Twitter I and Twitter II. It can be seen that our method also advances. Although the result of our method is lower than GMEI & LRMSI 0.4% on Twitter I_4, and lower than RA-DLNet 0.01% on Twitter II dataset. However, from the perspective of these two data sets as a whole, the proposed MACPNet is still more advanced than these two methods.

4.6. Visualization

Same as other visual classification problems, for visual sentiment analysis tasks, an important question is whether the proposed model can recognize the affective areas or objects in the image. This is important for us to evaluate and understand the model. In this section, we use the Class Activation Mapping method [45] to further evaluate our model by visualizing the multi-cue sentiment feature E R ( 2048 + 512 × 3 ) × 14 × 14 . Further evaluations were made on the EmtionROI and EMod datasets. EMod is a dataset specifically designed for the study of visual saliency and image sentiment analysis. It contains 1019 images coming from IAPS and Google’s search engine, and each image has eye tracking data annotated by 16 subjects. In order to evaluate these two datasets uniformly and to further test the robustness of our method, we abandon the conventional practice of training the model on the corresponding dataset before visualizing it.
But do visualization experiments on EmtionROI and EMod separately after training the models on large datasets FI. The visualization results are given in the form of heat map. A comparison of the visualization results using a single scale with the scale combinations used in our model is given.
As shown in Figure 10 and Figure 11, there are obvious differences in the objects or regions captured with different scale. However, when the objects’ size of the same class varies greatly within the same dataset like person in Figure 10 and Figure 11, the single scale is difficult to fully handle. Even different scales can correctly perceive the same object, the areas of interest at different scales are biased. As shown in Figure 10 e,f, both scales can correctly focus on squirrels and humans. However, different scales focus on different positions of the squirrel’s head and human body. It may even be that in the case of Figure 11f, the correct object is not well attended by a single scale, requiring a comprehensive multi-scale analysis. Therefore, multi-scale attributes are necessary in visual affective analysis. More results on the EmotioROI and EMod are given in Figure 12. From Figure 12, our visualization results are surprising. From some difficult situations such as people from far away (d1, h4), small flies on paper cups (c1), to faces with different sizes and appearances (a4, c4, d4, i4, j1, j4), our model all illustrates excellent object perception. However, it is worth noting that our model is not trained on these two datasets.
In Figure 13, we also give some representative examples of failure. For simple natural scenes such as (a) and buildings in (b), the eye tracking collection of many taggers is not uniform and does not focus on a certain area well. This is because there is no obvious object to attract the viewer in the image. The affective signal is from the overall content of the image, while the eye tracking data of the viewer come from their own habits. It is also very difficult to the model to locate similar objects with the same type and small differences, such as (c). Human-related or human-specific semantic information in images can attract us more effectively [1], such as plain text, as shown in (d) in Figure 13. However, for sentiment analysis of images, the understanding of plain text in Figure 13d is extremely difficult.

5. Conclusions

In this paper, we propose two attributes for visual sentiment analysis model: multi-scale perception and different levels of emotional representation. Furthermore, a novel multi-level context pyramid network composed of multi-scale adaptive context modules is proposed to learn the sentiment correlation degree of different regions for different scale in the image. The adaptive context framework is introduced for the first time in the image sentiment analysis task, which is helpful to improve the ability of the model to understand complex scenarios. The multi-scale attribute of the proposed MACM module is independent of the model structure and can combine different scales according to different data sources to mine the semantic information of images, which has good selectivity and scalability and can capture objects of different positions and sizes in the image. Two different level features fusion strategies are analyzed in the model to make better use of different levels of emotional representation to enhance the ability of the model to perceive semantic objects. The experimental results illustrate the performance advantage of our method on different datasets. Furthermore, the visualization results also show that the proposed MCPNet can effectively perceive some extreme situations, such as extremely small objects in the image.

Author Contributions

Methodology, C.Q., H.O., X.X., J.J.; software, H.O.; formal analysis, J.J.; supervision, C.Q.; writing, H.O., C.Q.; data curation, X.X., J.J.; funding acquisition, X.X. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Natural Science Foundation of China, grant number (61972163, U1801262, 61702192, 61802131), Natural Science Foundation of Guangdong Province, grant number(2019A1515012146), Fundamental Research Funds for the Central Universities, grant number (2019PY21, 2019MS028), Guangzhou Key Laboratory of Body Data Science, grant number (201605030011), Science and Technology Project of Zhongshan, grant number (2019AG024), Technology Program of Guangzhou, grant number (202002030354), Research and Development Project in Key Areas of Guangdong Province (2018B010109004) and UBTECH-SCUT Joint Research Lab.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Fan, S.; Shen, Z.; Jiang, M.; Koenig, B.L.; Xu, J.; Kankanhalli, M.S.; Zhao, Q. Emotional attention: A study of image sentiment and visual attention. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018. [Google Scholar]
  2. Brosch, T.; Pourtois, G.; Sander, D. The perception and categorisation of emotional stimuli: A review. Cogn. Emot. 2010, 24, 377–400. [Google Scholar] [CrossRef]
  3. Ortis, A.; Farinella, G.M.; Battiato, S. Survey on visual sentiment analysis. IET Image Process. 2020, 14, 1440–1456. [Google Scholar] [CrossRef]
  4. Rao, T.; Li, X.; Zhang, H.; Xu, M. Multi-level region-based convolutional neural network for image emotion classification. Neurocomputing 2019, 333, 429–439. [Google Scholar] [CrossRef]
  5. Joshi, D.; Datta, R.; Fedorovskaya, E.; Luong, Q.T.; Wang, J.Z.; Li, J.; Luo, J. Aesthetics and emotions in images. IEEE Signal Process. Mag. 2011, 28, 94–115. [Google Scholar] [CrossRef]
  6. You, Q.; Jin, H.; Luo, J. Visual sentiment analysis by attending on local image regions. In Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017; Volume 31. [Google Scholar]
  7. Fan, S.; Jiang, M.; Shen, Z.; Koenig, B.L.; Kankanhalli, M.S.; Zhao, Q. The role of visual attention in sentiment prediction. In Proceedings of the 25th ACM International Conference on Multimedia, Mountain View, CA, USA, 23–27 October 2017; pp. 217–225. [Google Scholar]
  8. Song, K.; Yao, T.; Ling, Q.; Mei, T. Boosting image sentiment analysis with visual attention. Neurocomputing 2018, 312, 218–228. [Google Scholar] [CrossRef]
  9. Rao, T.; Li, X.; Xu, M. Learning multi-level deep representations for image emotion classification. Neural Process. Lett. 2019, 1–19. [Google Scholar] [CrossRef] [Green Version]
  10. Chen, T.; Yu, F.X.; Chen, J.; Cui, Y.; Chen, Y.Y.; Chang, S.F. Object-based visual sentiment concept analysis and application. In Proceedings of the 22nd ACM international conference on Multimedia, Orlando, FL, USA, 3–7 November 2014. [Google Scholar]
  11. Zhao, Q.; Sheng, T.; Wang, Y.; Tang, Z.; Chen, Y.; Cai, L.; Ling, H. M2det: A single-shot object detector based on multi-level feature pyramid network. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 9259–9266. [Google Scholar]
  12. He, J.; Deng, Z.; Zhou, L.; Wang, Y.; Qiao, Y. Adaptive pyramid context network for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019. [Google Scholar]
  13. Surekha, S. Deep Neural Network-based human emotion recognition by computer vision, Advances in Electrical and Computer Technologies. Springer LNEE 2020, 672, 453–463. [Google Scholar]
  14. Cerf, M.; Frady, E.P.; Koch, C. Faces and text attract gaze independent of the task: Experimental data and computer model. J. Vis. 2009, 9, 10–11. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. You, Q.; Luo, J.; Jin, H.; Yang, J. Building a large scale dataset for image emotion recognition: The fine print and the benchmark. In Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence, Phoenix, AZ, USA, 12–17 February 2016. [Google Scholar]
  16. Borth, J.R.; Chen, T.; Breuel, T.; Chang, S.F. Large-scale visual sentiment ontology and detectors using adjective noun pairs. In Proceedings of the 21st ACM International Conference on Multimedia, Barcelona, Spain, 21–25 October 2013. [Google Scholar]
  17. Chen, T.; Borth, D.; Darrell, T.; Chang, S.F. DeepSentiBank: Visual Sentiment Concept Classification with Deep Convolutional Neural Networks. arXiv 2014, arXiv:1410.8586. [Google Scholar]
  18. Li, Z.; Fan, Y.; Liu, W.; Wang, F. Image sentiment prediction based on textual descriptions with adjective noun pairs. Multimed. Tools Appl. 2018, 77, 1115–1132. [Google Scholar] [CrossRef]
  19. Yuan, J.; Mcdonough, S.; You, Q.; Luo, J. Sentribute: Image sentiment analysis from a mid-level perspective. In Proceedings of the Second International Workshop on Issues of Sentiment Discovery and Opinion Mining, Chicago, IL, USA, 11–14 August 2013. [Google Scholar]
  20. Yang, J.; She, D.; Ming, S. Joint Image Emotion Classification and Distribution Learning via Deep Convolutional Neural Network. In Proceedings of the Twenty-sixth International Joint Conference on Artificial Intelligence, Melbourne, Australia, 19–25 August 2017. [Google Scholar]
  21. Kim, H.R.; Kim, S.J.; Lee, I.K. Building emotional machines: Recognizing image emotions through deep neural networks. IEEE Trans. Multimed. 2018, 20, 2980–2992. [Google Scholar] [CrossRef] [Green Version]
  22. Ali, A.R.; Shahid, U.; Ali, M.; Ho, J. High-level concepts for affective understanding of images. In Proceedings of the 2017 IEEE Winter Conference on Applications of Computer Vision (WACV), Santa Rosa, CA, USA, 27–29 March 2017. [Google Scholar]
  23. Peng, K.C.; Sadovnik, A.; Gallagher, A.; Chen, T. Where do emotions come from? predicting the emotion stimuli map. In Proceedings of the 2016 IEEE International Conference on Image Processing (ICIP), Phoenix, AZ, USA, 25–28 September 2016. [Google Scholar]
  24. Zhou, B.; Khosla, A.; Lapedriza, A.; Oliva, A.; Torralba, A. Object detectors emerge in deep scene cnns. International Conference on Learning Representations; San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  25. Yang, J.; She, D.; Lai, Y.K.; Rosin, P.L.; Yang, M.H. Weakly supervised coupled networks for visual sentiment analysis. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018. [Google Scholar]
  26. Yadav, A.; Vishwakarma, D.K. A deep learning architecture of RA-DLNet for visual sentiment analysis. Multimed. Syst. 2020, 26, 431–451. [Google Scholar] [CrossRef]
  27. Wu, L.; Qi, M.; Jian, M.; Zhang, H. Visual Sentiment Analysis by Combining Global and Local Information. Neural Process. Lett. 2020, 51, 2063–2075. [Google Scholar] [CrossRef]
  28. Yang, J.; She, D.; Sun, M.; Cheng, M.M.; Rosin, P.L.; Wang, L. Visual sentiment prediction based on automatic discovery of affective regions. IEEE Trans. Multimed. 2018, 20, 2513–2525. [Google Scholar] [CrossRef] [Green Version]
  29. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef] [Green Version]
  30. Mikels, J.A.; Fredrickson, B.L.; Larkin, G.R.; Lindberg, C.M.; Maglio, S.J.; Reuter-Lorenz, P.A. Emotional category data on images from the International Affective Picture System. Behav. Res. Methods 2005, 37, 626–630. [Google Scholar] [CrossRef] [PubMed]
  31. Machajdik, J.; Hanbury, A. Affective image classification using features inspired by psychology and art theory. In Proceedings of the 18th ACM International Conference on Multimedia, Firenze, Italy, 25–29 October 2010. [Google Scholar]
  32. You, Q.; Luo, J.; Jin, H.; Yang, J. Robust image sentiment analysis using progressively trained and domain transferred deep networks. In Proceedings of the Twenty-Ninth AAAI Conference on Artificial Intelligence, Austin, TX, USA, 25–30 January 2015. [Google Scholar]
  33. Lang, P.J.; Bradley, M.M.; Cuthbert, B.N. International affective picture system (IAPS): Technical manual and affective ratings. NIMH Cent. Study Emot. Atten. 1997, 1, 39–58. [Google Scholar]
  34. Peng, K.C.; Chen, T.; Sadovnik, A.; Gallagher, A. A mixed bag of emotions: Model, predict, and transfer emotion distributions. In Proceedings of the Computer Vision & Pattern Recognition IEEE, Boston, MA, USA, 8–12 June 2015. [Google Scholar]
  35. Panda, R.; Zhang, J.; Li, H.; Lee, J.Y.; Lu, X.; Roy-Chowdhury, A.K. Contemplating visual emotions: Understanding and overcoming dataset bias. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018. [Google Scholar]
  36. Ekman, P.; Friesen, W.V.; Ellsworth, P. What emotion categories or dimensions can observers judge from facial behavior? Emot. Hum. Face 1982, 39–55. [Google Scholar] [CrossRef]
  37. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  38. Siersdorfer, S.; Minack, E.; Deng, F.; Hare, J. Analyzing and predicting sentiment of images on the social web. In Proceedings of the 18th ACM international conference on Multimedia, Firenze, Italy, 25–29 October 2010. [Google Scholar]
  39. Zhao, S.; Gao, Y.; Jiang, X.; Yao, H.; Chua, T.S.; Sun, X. Exploring principles-of-art features for image emotion recognition. In Proceedings of the 22nd ACM International Conference on Multimedia, Orlando, FL, USA, 3–7 November 2014. [Google Scholar]
  40. Rao, T.; Xu, M.; Liu, H.; Wang, J.; Burnett, I. Multi-scale blocks based image emotion classification using multiple instance learning. In Proceedings of the 2016 IEEE International Conference on Image Processing (ICIP). IEEE, Phoenix, AZ, USA, 25–28 September 2016. [Google Scholar]
  41. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  42. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  43. Zhu, X.; Li, L.; Zhang, W.; Rao, T.; Xu, M.; Huang, Q.; Xu, D. Dependency exploitation: A unified CNN-RNN approach for visual emotion recognition. In Proceedings of the 26th International Joint Conference on Artificial Intelligence, Melbourne, Australia, 19–25 August 2017. [Google Scholar]
  44. COCO: Common Objects in Context. 2016. Available online: http://mscoco.org/dataset/#detections-leaderboard (accessed on 25 July 2016).
  45. Zhou, B.; Khosla, A.; Lapedriza, A.; Oliva, A.; Torralba, A. Learning deep features for discriminative localization. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
Figure 1. Images are from the emotional attention dataset EMOd [1]. In the eight columns of images from (ah), the above is the original image, and the below is the eye-tracking focus of the image. Images (ad) represent the backs of people, but their scales are significantly different, representing the different size of objects in the images. Images (eh) represent different levels of semantic content from simple to complex.
Figure 1. Images are from the emotional attention dataset EMOd [1]. In the eight columns of images from (ah), the above is the original image, and the below is the eye-tracking focus of the image. Images (ad) represent the backs of people, but their scales are significantly different, representing the different size of objects in the images. Images (eh) represent different levels of semantic content from simple to complex.
Sensors 21 02136 g001
Figure 2. The framework of multi-level context pyramid network (MCPNet). Firstly, the input image is sent to backbone ResNet101 to get the different level features. The features of c 3 , c 4 , and c 5 are put into MACM to obtain multi-scale context features, and then O 3 , O 4 , O 5 and c 5 layer features will be combined to obtain multi-cue emotional feature E for visual emotion analysis. In this figure, concat is a channel-wise concat, GAP means global average pooling, FC means fully connected layer.
Figure 2. The framework of multi-level context pyramid network (MCPNet). Firstly, the input image is sent to backbone ResNet101 to get the different level features. The features of c 3 , c 4 , and c 5 are put into MACM to obtain multi-scale context features, and then O 3 , O 4 , O 5 and c 5 layer features will be combined to obtain multi-cue emotional feature E for visual emotion analysis. In this figure, concat is a channel-wise concat, GAP means global average pooling, FC means fully connected layer.
Sensors 21 02136 g002
Figure 3. The process of MACM module. For the input feature graph X l , there are two branches. One is sub-regions branch. Another one is region sentiment affinity coefficient branch. Finally, the multi-scale context features are fused with the M l + 1 of the upper layer. In this figure, concat is a channel-wise concat.
Figure 3. The process of MACM module. For the input feature graph X l , there are two branches. One is sub-regions branch. Another one is region sentiment affinity coefficient branch. Finally, the multi-scale context features are fused with the M l + 1 of the upper layer. In this figure, concat is a channel-wise concat.
Sensors 21 02136 g003
Figure 4. The resize operation for different level layer.
Figure 4. The resize operation for different level layer.
Sensors 21 02136 g004
Figure 5. Some examples from the datasets we used.
Figure 5. Some examples from the datasets we used.
Sensors 21 02136 g005
Figure 6. The verification of the impact of different scale s on FI dataset.
Figure 6. The verification of the impact of different scale s on FI dataset.
Sensors 21 02136 g006
Figure 7. The performance with different levels of context features and global features removed on FI dataset.
Figure 7. The performance with different levels of context features and global features removed on FI dataset.
Sensors 21 02136 g007
Figure 8. The performance without different-layer context features and cross-layer feature fusion on FI dataset.
Figure 8. The performance without different-layer context features and cross-layer feature fusion on FI dataset.
Sensors 21 02136 g008
Figure 9. The results of our method on 5-fold cross validation on FI and EMotionROI. The final results are compared with the Rao (c)’s method, which has the highest accuracy presently.
Figure 9. The results of our method on 5-fold cross validation on FI and EMotionROI. The final results are compared with the Rao (c)’s method, which has the highest accuracy presently.
Sensors 21 02136 g009
Figure 10. Visualization of the proposed model on EmotionROI. Here we compare with different scale of S = 1, S = 2, S = 4. Our method uses S = [1,2,4]. GT is the ground truth. As shown in (bd), S = 2 and S = 1 are better than S = 4. However, in figure (a) the case of S = 1 is better than the case of S = 2 and S = 4. (e,f) reflect the different effects of different scales on the same object. These examples show that multi-scale models have more stable object perception capabilities than single-scale models.
Figure 10. Visualization of the proposed model on EmotionROI. Here we compare with different scale of S = 1, S = 2, S = 4. Our method uses S = [1,2,4]. GT is the ground truth. As shown in (bd), S = 2 and S = 1 are better than S = 4. However, in figure (a) the case of S = 1 is better than the case of S = 2 and S = 4. (e,f) reflect the different effects of different scales on the same object. These examples show that multi-scale models have more stable object perception capabilities than single-scale models.
Sensors 21 02136 g010
Figure 11. Visualization of the proposed model on EMod. Here we compare with different scale of S = 1, S = 2, S = 4. Our method uses S = [1,2,4]. GT is the ground truth. As shown in (ae), sometimes single-scale or multi-scale can effectively perceive objects related to sentiment, but in complex situations such as (f), multi-scale is more effective than single-scale.
Figure 11. Visualization of the proposed model on EMod. Here we compare with different scale of S = 1, S = 2, S = 4. Our method uses S = [1,2,4]. GT is the ground truth. As shown in (ae), sometimes single-scale or multi-scale can effectively perceive objects related to sentiment, but in complex situations such as (f), multi-scale is more effective than single-scale.
Sensors 21 02136 g011
Figure 12. More visualization comparisons on EmotionROI and EMod. Here the complexity of appearance change varies with different objects. For the small-sized objects, such as (f1) birds in grass, (g1) masks on chairs, (b1, i1, h4) people, their appearance and surroundings are quite different. Since our model is not trained on these two datasets, the comparisons on these two datasets are fairly. GT is the ground truth.
Figure 12. More visualization comparisons on EmotionROI and EMod. Here the complexity of appearance change varies with different objects. For the small-sized objects, such as (f1) birds in grass, (g1) masks on chairs, (b1, i1, h4) people, their appearance and surroundings are quite different. Since our model is not trained on these two datasets, the comparisons on these two datasets are fairly. GT is the ground truth.
Sensors 21 02136 g012
Figure 13. Some typical examples of failures on the EMod dataset. GT is the ground truth. For example, there are no prominent semantic objects in (a,b), and (c) exemplifies that when there are a large number of similar objects in the scene (people in this example), the visualization of the model in these examples is terrible. Human-specific factors are also difficult for the model. For example, the viewer can understand the text information in (d,e), but the model does not work.
Figure 13. Some typical examples of failures on the EMod dataset. GT is the ground truth. For example, there are no prominent semantic objects in (a,b), and (c) exemplifies that when there are a large number of similar objects in the scene (people in this example), the visualization of the model in these examples is terrible. Human-specific factors are also difficult for the model. For example, the viewer can understand the text information in (d,e), but the model does not work.
Sensors 21 02136 g013
Table 1. Seven commonly used visual sentiment analysis datasets. Most datasets are from social networks. Except for FI dataset, the others are relatively small datasets, because of the subjective and labor-intensive labeling process.
Table 1. Seven commonly used visual sentiment analysis datasets. Most datasets are from social networks. Except for FI dataset, the others are relatively small datasets, because of the subjective and labor-intensive labeling process.
DatasetPositiveNegativeSumType
IAPSsubset [30]209186395natural
Abstract [31]13989228abstract
ArtPhoto [31]378428806artistic
Twitter I [32]7695001269social
Twitter II [16]463133596social
EmotionROI [23]66013201980social
FI [15]16430687823308social
Table 2. Comparison of different configurations of our model on FI dataset, including “only” and ”without”. For example, “only” of O 5 refers to the final sentimental feature E   =   O 5 , “without” means that E is the concatenation of O 4 , O 3 , X 5 ,without O 5 . The bold score is the highest score of each column.
Table 2. Comparison of different configurations of our model on FI dataset, including “only” and ”without”. For example, “only” of O 5 refers to the final sentimental feature E   =   O 5 , “without” means that E is the concatenation of O 4 , O 3 , X 5 ,without O 5 . The bold score is the highest score of each column.
FeatureOnlyWithout
O l = Z l × M l 90.004-
global-89.789
O 5 89.01188.393
O 4 88.66888.484
O 3 81.25089.675
ours90.3190.31
Table 3. Classification results of different state-of-the-art methods on five different datasets: FI, ArtPhoto, Abstract, IAPSsubset, EmotionROI. The bold score is the highest score of each column.
Table 3. Classification results of different state-of-the-art methods on five different datasets: FI, ArtPhoto, Abstract, IAPSsubset, EmotionROI. The bold score is the highest score of each column.
MethodFIArtPhotoAbstractIAPSsubsetEmotionROI
LCH [38]45.3764.3370.9352.8463.79
GCH [38]47.9566.5367.3369.9666.85
PAEF [39]58.4268.4266.2365.7773.45
Rao(a) [40]62.7971.5367.8278.3474.51
SentiBank [16]56.4767.3364.3080.5765.73
AlexNet [41]68.6369.2765.4984.5871.60
VGG-16 [42]73.9570.4865.8887.2072.49
ResNet101 [37]75.7671.0866.6488.1573.92
DeepSentiBank [17]64.3970.2669.0786.3170.38
PCNN [32]73.5971.4770.2688.6574.06
Rao(b) [9]79.5474.8371.9690.5378.99
Zhu [43]84.2675.5073.8891.3880.52
AR [28]86.3574.8076.0392.3981.26
Rao(c) [4] 87.5178.3677.2893.6682.94
ours90.3179.2478.1494.8285.10
Table 4. Classification results of different state-of-the-art methods on twitter I and twitter II. The bold score is the highest score of each column.
Table 4. Classification results of different state-of-the-art methods on twitter I and twitter II. The bold score is the highest score of each column.
MethodTwitter ITwitter II
Twitter I_5Twitter I_4Twitter I_3
GCH [38]67.9167.2065.4177.68
LCH [38]70.1868.5465.9375.98
PAEF [39]72.9069.6167.9277.51
SentiBank [16]71.3268.2866.6365.93
DeepSentiBank [17]76.3570.1571.2570.23
PCNN [32]82.5476.5276.3677.68
VGG-16 [42]83.4478.6775.4971.79
AR [28]88.6585.1081.0680.48
RA-DLNet [26]89.1083.2081.3081.20
GMEI &LRMSI [27]89.5086.9781.6580.97
ours89.7786.5783.8881.19
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Ou, H.; Qing, C.; Xu, X.; Jin, J. Multi-Level Context Pyramid Network for Visual Sentiment Analysis. Sensors 2021, 21, 2136. https://doi.org/10.3390/s21062136

AMA Style

Ou H, Qing C, Xu X, Jin J. Multi-Level Context Pyramid Network for Visual Sentiment Analysis. Sensors. 2021; 21(6):2136. https://doi.org/10.3390/s21062136

Chicago/Turabian Style

Ou, Haochun, Chunmei Qing, Xiangmin Xu, and Jianxiu Jin. 2021. "Multi-Level Context Pyramid Network for Visual Sentiment Analysis" Sensors 21, no. 6: 2136. https://doi.org/10.3390/s21062136

APA Style

Ou, H., Qing, C., Xu, X., & Jin, J. (2021). Multi-Level Context Pyramid Network for Visual Sentiment Analysis. Sensors, 21(6), 2136. https://doi.org/10.3390/s21062136

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop