1. Introduction
The last few years have been characterized by a marked growth in the use of social media, leading to the production of huge amounts of digital data effectively exploitable to investigate human dynamics and behaviors. Such data, commonly referred as Social Big Data, contain valuable information about people that makes them intrinsically suited to a very large set of application fields [
1], such as regions-of-interest and user trajectory extraction [
2], influence maximization [
3], sentiment analysis and emotional profiling [
4,
5], topic detection and opinion mining [
6,
7]. However, the quality of data extracted from social media can be lowered by the presence of fake news that can hinder this type of analysis, leading to misleading results.
This paper focuses on the use of social media data, in particular those coming from Twitter, to estimate the polarization of public opinion concerning a political event characterized by the rivalry of different factions or parties. In particular, we propose a new methodology, called TIMBRE (Time-aware opInion Mining via Bot REmoval) that exploits a keyword-based classification to determine the political polarization of social media posts. The proposed methodology is temporally aware, as it takes into account time-related aspects in computing the importance weight of each classified post. This weight represents the relevance of that post on the voting intentions of the user who published it. Finally, the political orientation of a user is obtained starting from his/her published posts, according to their polarization and weight.
Depending on the political event, social media users can be classified towards a particular faction, candidate, or choice. However, in this kind of application, the results could be biased and distorted by many factors, including data artificially produced by social media bots. They consist of software applications used to automatically generate messages on social media so as to influence public opinion, spam messages or amplify propaganda. Bots can act as fake accounts (e.g., for posting messages and gaining followers itself) or as followers of other social media users. It is estimated that 9–15% of Twitter accounts may be social bots. Due to this, a key aspect of TIMBRE is the bot removal step, aimed at avoiding the distortion effect introduced by the presence of bot-generated data. In this way the methodology is able to grasp the real voting intentions on social media platforms, capturing only the polarization of legitimate users who belong to the voting eligible population.
To test the proposed methodology we applied it to a real-world case study that analyzes the polarization of a large number of Twitter users during the 2016 US presidential elections, which was characterized by the rivalry between Hillary Clinton and Donald Trump. This use case is particularly interesting, since it was characterized by a marked use of Twitter to foster political debate along with a significant activity by social bots, which would have strongly influenced voter decisions [
8,
9,
10]. In particular, we focused on the analysis of the main US Swing States, characterized by a great political uncertainty, finding out that both the temporal weighting of posts and bot removal are crucial in order to get a correct estimate of users’ voting intentions. The achieved results have been compared with opinion polls collected before voting and with the actual results obtained after the vote, revealing a high accuracy of TIMBRE in estimating the polarization of social media users. In particular, our methodology was able to correctly identify the winner in 8 out of 10 Swing States, outperforming the opinion polls, which identified the winning candidate in 6 out of 10 cases.
As a last step, we studied how the presence of social bots may have affected political discussion around the 2016 US presidential election, focusing on two main aspects. On one hand we analyzed the publishing behavior of both real users and social bots, along with the differences between human and artificial political support. On the other hand, we exploited a competitive diffusion model to estimate the degree of influence of social bots on legitimate users.
The remainder of the paper is organized as follows.
Section 1.1 reviews the main social bot detection techniques present in literature.
Section 2 discusses related work.
Section 3 describes the proposed methodology.
Section 4 presents the case study and obtained results. Finally,
Section 5 concludes the paper.
1.1. Problem Statement
The last few years have been characterized by a marked growth of social media legitimate use and manipulation, fostering democratic conversation about socio-political issues [
9] and, at the same time, a large spread of misinformation. This phenomenon has made social platforms one of the most used sources of information, exposing users to risks caused by the lack of veracity of news. Moreover, political online discussion is often strongly polarized, leading to the formation of echo chambers that provide selective exposure to news sources biasing the opinion of users. This effect sometimes is amplified by the priority policies of the main social media platforms, which tend to favor engaging rather than trustworthy posts [
11]. In such a scenario, getting reliable and impartial news, discerning them from rumor, constructed reports and fake news, could be a hard task. Social bots, also known as a sybil account, are among the factors that most undermine the reliability of online news. They can be defined as algorithmically-driven entities that automatically produce content and interact with humans on social media, trying to emulate and alter their behavior. In a political scenario, bots can be used illicitly to artificially increase the support for a candidate, influencing the outcome of the election. Campaigns of this type are usually called astroturf or Twitter bombs. Many efforts were made by the research community towards developing social bot detection and classification systems, especially on Twitter, one of the most used microblogging platforms. According to [
12], state-of-the-art techniques can be categorized in three main classes: graph-based detection, crowdsourcing and machine learning.
Graph-based detection. Methods in this category exploit a graph-based representation of a social network to understand the relationships between edges or links across accounts, using this information for detecting bot activity. As described in [
13], there are three main graph based approaches aimed at detecting social bots and malicious accounts: (i) trust propagation that quantifies the strength of the relationship among users; (ii) graph clustering groups similar users according to their characteristics. (iii) graph analysis that relies on several metrics and properties of the social graph, like degree distribution and centrality measures. SybilWalk [
14] is a sybil detection method that exploits a random walk-based method on an undirected social graph. It proceeds by assigning a score to users in the social graph, which is then used to classify them as legitimate users or sybils. Mehrotra et al. [
15] proposed a supervised method for fake followers detection based on several centrality metrics which exploits a Random Forest classifier.
Crowdsourcing. This class of methods leverages human detection to identify social bot behaviors, seeking patterns across profile information or shared content. As an example, DARPA held a Twitter bot challenge competition [
16] in which teams were asked to identify influential bots that supported pro-vaccination discussions on Twitter. A common use of human annotation in bot detection involves the generation of annotated datasets, which can be then used by supervised techniques. In [
17] four annotators were employed for the classification of Twitter profiles as bot or human, starting from a wide range of features such as the number of tweets or favorites. Similarly, in [
18] ten volunteers were tasked with labeling 2000 random accounts, in order to build a ground truth dataset.
Machine learning. These methods are based on machine learning algorithms and statistical techniques for social bot detection. Kantepe et al. [
19] proposed a supervised approach which relies on an extensive process of feature extraction. In particular, they used Apache Spark for data collection, categorizing features in three types, i.e., user, tweet and periodic features. Afterwards, a gradient boosting classifier is used to label users as human or bots. Devis et al. [
20] proposed Botometer (formerly BotOrNot), a classification system that leverages more than one thousand features to evaluate the extent to which a Twitter account exhibits similarity to the known characteristics of social bots. Specifically, such features are extracted from available meta-data, shared content, and interaction patterns. Ersahin et al. [
21] presented a supervised method for fake account detection on Twitter which leverages a naïve bayes classifier and an entropy minimization discretization technique. Cai et al. [
22] proposed a behavior-enhanced deep learning model (BeDM) for social bot detection. In particular, they jointly exploited a convolutional neural network and a long short-term memory network to capture temporal patterns in user behavior.
2. Related Work
With the rapid growth in their use, social media platforms have become a valuable source of information, effectively exploitable in many application fields. In particular social media data can be leveraged for investigating the patterns of information diffusion, the interactions between users and their opinion about a specific topic [
7]. Several opinion mining techniques have been proposed in literature for understanding the opinion of social media users regarding political events. These techniques belong to a research area called computational politics, that includes a wide range of methods aimed at analyzing the behavior of social media users during a political event of interest, modeling and influencing their perception and opinion about facts, events and public decisions.
Belcastro et al. [
7] proposed an opinion mining technique, namely IOM-NN, aimed at discovering the political polarization of social media users during election campaigns characterized by the competition of political factions. The methodology relies on an iterative and incremental procedure based on feed-forward neural networks, aimed at determining the political orientation of posts used for discovering the political polarization of social media users. Marozzo and Bessi [
23] proposed a methodology that exploits the keywords contained in tweets for calculating the polarization of social media users and news sites during political campaigns. Diamantini et al. [
24] proposed a lexicon-based sentiment analysis algorithm, which uses a combination of word sense disambiguation and negation handling techniques for extracting user opinion from social media data. Burnap et al. [
25] proposed a model for using Twitter as an election forecasting tool, applying it to the UK 2015 General Election. Oikonomou et al. [
26] used a naïve bayes classifier with text mining techniques given by TextBlob, a Python library which provides an API for Natural language processing (NLP), to predict the outcome of USA presidential elections in three states of interest (i.e., Florida, Ohio and North Carolina). Jaidka et al. [
27] compared three different methods (i.e., volumetric, sentiment and social media analysis) in order to predict the outcome of the elections from Twitter posts in three Asian countries: Malaysia, India, and Pakistan. Olorunnimbe et al. [
28] presented an incremental learning method based on multiple naïve bayes independent models for predicting the political orientation of users over time. Wong et al. [
29] modeled the political behaviour of users by analyzing their publishing activity using SentiStrength, a lexicon-based sentiment analysis tool. Alashri et al. [
30] leveraged CoreNLP, one of the most popular tools for natural language processing, for the analysis of Facebook posts related to the 2016 US presidential election. Specifically, authors examined the dynamics between candidate posts and comments they received on Facebook for calculating a score for each political candidate aimed at measuring his/her credibility. Finally, Singh et al. [
31] carried out a comparison among four machine and deep learning algorithms (i.e., textblob, naïve bayes, SVM, and BERT [
32]) for sentiment analysis, taking the 2020 US presidential election as a case study. Authors found that the use of BERT leads to the best results, which shows the effectiveness of transformer-based language representation models.
The aforementioned techniques are often heavily dependent on the representativeness of social media data. As a consequence, the bias introduced by content artificially produced by social media bots can compromise the final results. There are several studies that show how the presence of social bots has altered the political discussion on social media platforms. As regards the 2016 US presidential election, Bessi and Ferrara [
9] analyzed the pervasive presence and activity of social bots involved in social media conversation. They found out that about 400,000 bots were engaged in the political discussion about the Presidential election, responsible for roughly 3.8 million tweets (i.e., about one-fifth of the entire conversation). For this reason, the methodology we propose in this work filters out the data produced by social bots, identifying them through the use of the Botometer [
20] framework. Thus, by jointly exploiting a bot detection system and a temporally-aware polarization technique, TIMBRE is able to accurately detect the real voting intentions on social media platforms, capturing only the polarization of legitimate users.
Our manuscript is one of the few research works that focuses on the study of bots and their effect on the specific task of analyzing election results. We show how the estimation of election results from social data can be biased by the presence of bots, measuring this effect in terms of voting percentages estimates and incorrectly classified states. We also show how bots have influenced social discussions by analyzing information production patterns and the spread of influence within the social network.
3. Materials and Methods
As mentioned above, TIMBRE (Time-aware opInion Mining via Bot REmoval) exploits a keyword-based classification for determining the political polarization of social media users and the Botometer framework to distinguish legitimate users (i.e., voters) from social bots. In addition, it analyzes how the presence of social media bots may have negatively affected online discussion during the political event under analysis, potentially altering public opinion.
Given a political event , a set of the factions , and a set the keywords associated to , the proposed methodology consists of four main steps:
Post collection: posts are collected by using the set of keywords related to the political event .
Post classification and weighting: for each post we determine its political orientation, neutral or in favor of a specific faction , and a weight indicating the importance of the post p in estimating the voting intentions of the user u who published it.
User polarization and classification: starting from classified posts and related weights, we determine the political partisanship of each user in our dataset, classifying it as a real user or a social bot. This information is then used to forecast the outcome of the event .
Bot influence analysis: during this step we analyze information production patterns, estimating also the degree of influence of social bots on real users.
For each step, a formal description is provided in the following sections.
3.1. Post Collection
A political event
is characterized by the rivalry of different parties or factions
. Examples of political events and relative factions are: (i) municipal election, in which a faction supports a mayor candidate; (ii) parliament election, in which a faction supports a party; (iii) presidential election, in which a faction supports a presidential candidate. Following the approach proposed in [
7], posts are collected by using the keywords that people commonly use to refer to a given political event
on social media. Such keywords
K can be divided in two classes:
- -
, which contains generic keywords that can be associated with without referring to any specific faction in .
- -
, where contains the keywords used for supporting .
The keywords in
K are given as input to public APIs provided by social media platforms, which permit collecting posts containing one or more keywords. Since data collection is usually a continuous process, new keywords can be discovered and integrated in
K during the collection procedure. As the author of [
7] highlighted, obtaining a representative collection of posts depends on two main factors: (i) the quality and the number of keywords used; (ii) the amount of data that can be downloaded from social media. Regarding the latter factor, it is worth mentioning that it is increasingly difficult to obtain complete data from social media platforms due to the restrictions introduced for protecting the privacy of users. The collected posts are pre-processed before the analysis as follows:
Hashtags are normalized removing non-alphanumerical character and transforming them to lowercase. This way we can avoid differences between different versions of the same hashtag, e.g., , or becomes .
Data representativeness is further improved by filtering out all the posts having a language different from the one spoken in the nation hosting the considered political event.
As the proposed method relies on a hashtag-based analysis without exploiting other textual information, no further preprocessing like stopwords removal or lemmatization is needed. The output of this step is a collection of posts P related to the event .
3.2. Post Classification and Weighting
In this phase we assign each post included in P to a specific faction in by analyzing the keywords it contains, defined as the set . In particular, if a post p contains only keywords that are in favor of a specific faction f, then p is classified as in favor of f; otherwise, p is classified as neutral. This is a very strict and conservative partisanship assignment, which leads to a small but high-confidence annotated dataset, likely less prone to misclassification than automatic machine-learning techniques.
Besides classifying posts in favor of a specific faction, we determine a weight indicating the importance of the post p in estimating the voting intentions of the user u who published it. The intuition behind this is that more recent posts are more suited for deriving useful information about voting intentions of a user. In fact, users’ polarization can vary over time as they can influence each other or be influenced by external events, such as political debates or scandals. The importance weight is computed as follows. Given a user and the set of his/her posts , we determine as the day the user u published his/her last post before the end of . Given a post p published by user u the day d, and , we define the importance weight as:
This weight undergoes exponential decay according to a constant
(
decay rate): larger values of this constant make the quantity vanish much more rapidly. Algorithm 1 shows the pseudo-code of the classification procedure, whose output
S consists of a set of triple containing the post
p, the associated faction
and the importance weight
.
Algorithm 1: Post classification and weighting |
|
3.3. User Polarization and Classification
Starting from the set S containing classified and weighted posts, we use a one-vs-all strategy for determining the political partisanship of each user in our dataset. Specifically, given the set of opposing factions , we compute user polarization as follows. Given a user , let be the set containing all of his/her posts, and its subset containing only post published by u classified as in favor of f in the previous step. For each faction f we determine the support of u towards f as:
As the above formula is normalized in the interval , positive values of means that user u tends to be polarized towards the faction f, and the polarization become stronger as approaches the value of 1. Negative values, instead, suggest a polarization towards the set of all the remaining factions. Therefore, given a threshold used for assign a faction only to users who show a strong polarization, political partisanship of u is determined as follows:
- -
, if
- -
otherwise
Besides determining user partisanship, we also exploited the Botometer framework for the automatic classification of social media users into real or fake accounts, related to potential electors and automatic entities respectively. Given a user
u Botometer determines a real-valued score
which measures the likelihood that user
u is a social bot. According to prior studies ([
9,
20]), we selected a threshold value for
l equal to
, for the classification process. At the end of the entire procedure two dictionaries
B and
R are obtained, related to bots and real users respectively, composed by
key-value pairs. The pseudo-code of the user polarization and classification procedure is shown in Algorithm 2.
Algorithm 2: User polarization and classification |
|
Once the user polarization and classification step is completed, the outcome of the political event
can be determined starting from the
R set, containing the polarity of legitimate users. Let
be the subset of
R containing all users polarized in favor of
f; the final consensus
for each faction
is determined as follows:
3.4. Bot Influence Analysis
During this step we analyze how the presence of social media bots may affect political discussion around the event under analysis. After having built the set P of classified posts and the sets R and B, indicating bots and real users partisanship, the proposed methodology analyzes them exploiting different algorithms and techniques, focusing on the following aspects.
First of all, we built the repost graph , a directed graph where is the set of bots and real users involved in repost relationships and E is the set of edges where v reposted u, with . For each edge we assigned a unique real-valued weight corresponding to the impact of node u on v, computed as follows. Let be the number of times node v reposted u and the number of total reposts made by v; the weight of the edge is defined as: , with . Therefore, a node u has a high influence on v if v shows a high tendency in reposting u’s posts more then the others.
Once the network is built, given the set of factions involved in the political event , and the set of polarized bots , we partitioned this set in n disjoint subsets , such as contains only social bots polarized towards the faction f. For remaining users (i.e., neutral bots and real users ), a threshold values for each faction is selected, picked uniformly at random in the interval , representing the resistance of user u to be influenced in favor of the faction f. At the step t, for each faction , let be the set of nodes influenced by faction f. During this step, a neutral node v becomes polarized towards f if , which means that the influence exercised on v in favor of f is higher than its resistance to that faction. If for the node v more than one threshold is exceeded during the step t, then this node will be polarized in favor of the faction that exercises the highest influence. This process ends when all neutral nodes become influenced, returning n disjoint sets, containing the users (both real and bot) polarized towards one of the factions and an additional set containing unpolarized nodes.
4. Results and Discussion
In the following we discuss a case study related to the 2016 US presidential election characterized by the rivalry between Hillary Clinton and Donald Trump. Our analysis focused on 10 US Swing States: Colorado, Florida, Iowa, Michigan, Ohio, New Hampshire, North Carolina, Pennsylvania, Virginia, and Wisconsin. These states are given high strategic importance as they are characterized by a great political uncertainty. Therefore, information manipulation in those states, carried out by influencing the political orientation of social media users, can have significant effects on the election outcome.
As explained in
Section 3.1, posts were collected using a set of neutral keywords and two sets of faction keywords, one for each candidate. An extract of these sets is shown in the following:
= {election2016, elections2016, uselections, uselection, earlyvote, ivoted}
= {clintokaine16, democrats, hillary16, imwithher, nevertrump, strongertogether}
= {wakeupamerica, votetrump, maga, trump16, americafirst, neverhillary, podestaemails}
We analyzed about 4.7 million posts posted by 1.5 million users, finding a non-negligible impact of social bots on political discussion. As shown in
Table 1, states like Colorado, Iowa and Ohio, are characterized by a high rate of bot posts, from 20.6% to 24.6%. Furthermore, 7% of total user accounts have been identified as social bots, which produced about 15% of the total posts related to the 2016 US presidential election coming from the analyzed swing states. This last result is in agreement with [
9], which found a percentage of posts published by bots equal to 20%, albeit using a different sample of tweets and analysis methodology.
Collected data are representative of the analyzed event as:
All the posts under analysis have the lang field set to en (i.e., English).
About 94% of the social media users in the USA are adults and almost equally divided by gender (42.7% females and 57.3% males).
For each state, we measured the correlation between collected users and voting eligible population (VEP). We observed a strong linear correlation, with a Pearson coefficient
, which improved after removing bots reaching
. Both results are significant at
, therefore collected users can be considered voters in the related swing state.
Figure 1 summarizes these results by showing a linear interpolation, along with the goodness-of-fit measured through the determination coefficient (
).
In the next two subsections, we analyze the polarization of users during the 2016 US presidential election campaign and how the presence of bots may have affected the political discussion on Twitter.
4.1. Polarization Analysis
In this step we exploited Algorithms 1 and 2, described in
Section 3.2 and
Section 3.3, for determining the political orientation of the collected posts and the corresponding users. Furthermore, posts are assigned an importance weight and users are classified as real accounts or social bots. The decay rate
and the threshold
have been set to
and
respectively.
Table 2, shows how the support detected for the different factions is distributed among real users and bots. We would like to clarify that with pro-
X bots we indicate Twitter accounts classified as bots, which have mainly published tweets in favor of candidate
X.
We found a greater presence of pro-Trump bots, which have a more marked impact on the online discussion, producing almost 18% of the contents classified as in favor of Trump. This suggests a greater use of social bots that published contents supporting the Trump political positioning compared to the other faction, which however shows a quite high volume of bot-generated content, in line with work [
9].
Once posts and users were classified according to their polarity and social bot were detected using Botometer, we determined the outcome of the 2016 US election as explained in
Section 3.3. The achieved results are summarized in
Table 3, which shows a comparison among the real voting percentages, the average values of the latest opinion polls before the election, and the results obtained by using TIMBRE. The winning candidate is written in bold when it is correctly identified.
Compared to the latest opinion polls, which gave a correct forecast for only 6 out of 10 swing states, the proposed methodology was able to correctly identify the winning candidate in 8 out of 10 states, confirming its ability to accurately determine the polarization of social media users. TIMBRE outperformed the latest opinion polls even in terms of average absolute error, improving it from
to
. We computed this metric only focusing on wrong predictions by using the following formula:
where
and
are the set of considered factions and states,
and
are the real and predicted voting percentages related to the faction
f in the state
s, and
is a binary function which outputs 1 if the predicted polarity is wrong, 0 otherwise (i.e., the winning candidate is correctly identified). Using this metric we both penalized the absolute error in terms of percentage points and the inversions predicted polarity, which can be a crucial issue while analyzing these states, characterized by a high degree of uncertainty. Another noteworthy advantage is related to the number of polarized users, which is much larger than that of the people interviewed. Consequently, this approach can be thought as a valid alternative to traditional opinion polls, since it is able to capture the opinion of a larger number of people more quickly and at a lower cost.
We further extended our experimental evaluation by analyzing the benefits brought by each of the two key steps introduced by the proposed methodology: temporal weighting and bot removal.
The achieved results, reported in
Table 4, show that both the temporal weighting of posts and bot removal steps are crucial in order to get a correct estimate of users’ voting intentions. In particular, the
base version of the proposed methodology, that does not leverage neither the removal of bots nor the temporal weighing of posts, achieved the same accuracy of the latest polls, correctly identifying the winning candidate in 6 out of 10 states. By adding the bot removal step to the base version, the resulting methodology was able to correctly predict the final outcome in Colorado, increasing its accuracy from 6 to 7 out of 10 states correctly classified. Similarly, by only adding the time-base weighting mechanism, we observed an increase in the forecasting ability of our methodology, which corrected its prediction for the state of Florida. Finally, TIMBRE was able to maintain the benefits coming from both of the aforementioned steps, combining them and correctly determining the winning candidate in 8 out of 10 states. Finally, it is worth noting that the results for Pennsylvania and Wisconsin that were not correctly predicted by TIMBRE were not correctly predicted even by opinion polls.
4.2. Bot Influence Analysis
In this section we analyze how the presence of social bots may have affected the political online discussion around the 2016 US presidential election. Specifically, we firstly analyzed the publishing behavior of both real users and social bots focusing on the patterns of information production. Then, we studied the main differences in supporting the two candidates between human-driven and artificial accounts. Finally, we estimated the degree of influence of social bots on legitimate users using a competitive information diffusion model.
4.2.1. Information Production Patterns
In order to extract the publishing behavior of social media users involved in the political discussion, we used the information about their political orientation coming from the user polarization step, computing a publishing model for each candidate. In particular, such models are represented by the complementary cumulative distribution function (CCDF) of the number of posts posted by users supporting Clinton and Trump respectively. Obtained results considering all accounts and excluding Bot accounts from them are shown in
Figure 2. Specifically, for a given number of posts
x the scatter plots show, in a log-log scale, the frequency of users publishing a number of posts greater than
x (i.e.,
).
Analyzing the publishing behavior of all polarized users (both real and fake accounts), shown in
Figure 2a, we observed a greater publication tendency of pro-Trump accounts, which result much more prolific than pro-Clinton ones. However, the role of polarized bots behind this phenomenon should be investigated: for this purpose
Figure 2b shows the publishing behavior of legitimate users only. By excluding the bots from the CCDF of both candidates, we observed a narrowing of the distance between the two curves relating to pro-Trump and pro-Clinton users. Therefore the polarity does not seem to be a deciding factor affecting the volume of posts published by legitimate users. As a consequence, it can be deduced that the differences emerging in
Figure 2a are due to an amplifying effect caused by social bots. Moreover, this agrees with the higher activity of pro-Trump bots with respect to pro-Clinton ones, detected in the previous sections. For completeness, in
Table 5 we provide the description of the most prolific real accounts in our dataset, according to the detected polarity. In particular, for each candidate we selected the user labeled as real by Botometer that published the highest number of posts, i.e., the rightmost point of the scatter plot in
Figure 2b.
Despite the high number of published posts, Botometer gave for the two accounts a BotScore score far below 0.5, which suggests that they are truly managed by prominent users or news sites, but not by automatic entities.
4.2.2. Influence Spread
This last step is aimed at estimating the degree of influence of social bots on legitimate users, following the approach described in
Section 3.4. For this purpose, we built a graph
G based on repost relationships characterized by 437,854 nodes and almost 1.5 million edges. From that graph have been removed self-loops, duplicated edges and isolated nodes. Afterwards we analyzed the spread of influence by adapting the Separated-Threshold Model for Competing Technologies (see
Section 3.4) to our case study, characterized by the rivalry of two candidates. Due to this, the diffusion process starts from two distinct seed-sets containing respectively the bots polarized for the Democratic and the Republican party. When convergence is reached, we end up with a list of influenced nodes labeled with the related polarity. We conducted 20 simulations varying the initial assignment of the random thresholds that represent the resistance of the users in the network to be influenced by social bots. Starting from the achieved results we computed two quantities:
The expected spread for each candidate, determined as the average number of influenced nodes across the 20 simulations by pro-Trump and pro-Clinton nodes.
The set of influenceable nodes, obtained through the voting technique. In particular, all the nodes activated at least once during the different simulations were assigned to the faction that influenced them the greatest number of times.
The final results obtained after the different simulations of the diffusion process are shown in
Table 6. Both the expected number of influenced nodes and the total number of influenceable nodes confirmed the greatest activity of pro-Trump bots, which had a more marked impact on social media conversation compared to pro-Clinton ones. In particular, the expected number of nodes influenced by the seed-set of pro-Trump bots was 12.4 times greater than compared to the opposite seed-set of pro-Clinton ones. Similarly, the number of influenceable nodes was 7.8 times greater.
Figure 3 graphically summarizes the results obtained in this step. In particular, the entire
G graph is plotted (left graph), coloring the different nodes according to their polarity and characteristics. In particular, the polarized bots belonging to pro-Trump and pro-Clinton seed-sets are colored in dark red and dark blue respectively, influenceable nodes assigned to Trump are represented in light red, those assigned to Clinton in light blue and neutral nodes in gray. Finally, in order to obtain a clearer view of the influenced nodes in the network, we reduced the initial graph by 90% while keeping the top-
k nodes with highest degree (right graph). In this way we maintained almost unchanged the polarity-based clustering structure emerged in the total graph, achieving a neater representation of the results of the diffusion process.
5. Conclusions and Final Remarks
This paper proposes a new methodology, namely TIMBRE (Time-aware opInion Mining via Bot REmoval), aimed at discovering the polarization of social media users during election campaigns characterized by the rivalry of political factions or parties. This methodology exploits a keyword-based classification to determine the political polarization of social media posts and users. It is temporally-aware, as it considers time-related aspects in deciding how much a post can be helpful to determine the voting intentions of the user who published it. Moreover, it recognizes and filters out data produced by social media bots, algorithmically-driven entities that participate in online discussion with the aim of altering the public opinion about political candidates.
In order to assess the effectiveness of TIMBRE, it was applied to a real-world case study related to the 2016 US presidential election. By leveraging Twitter metadata, we focused only on posts coming from 10 US Swing States, in particular: Colorado, Florida, Iowa, Michigan, Ohio, New Hampshire, North Carolina, Pennsylvania, Virginia, and Wisconsin. The achieved results showed the high accuracy of the proposed approach, along with the benefits brought on forecasting accuracy by its two key steps, i.e., temporal weighting and bot removal. Specifically, our methodology was able to correctly identify the winning candidate in 8 states out of 10, with an average absolute error of 0.9 percentage points, outperforming the latest opinion polls, which identified the winner in 6 out of 10 cases, with an average error of points.
As a final step, we investigated how the presence of social bots may have affected political discussion around the 2016 US presidential election. In particular, we firstly analyzed the publishing behavior of both real users and social bots focusing on the patterns of information production. Then, we studied the main differences in supporting the two main candidates between human-driven and artificial accounts. Finally, we estimated the degree of influence of social bots on legitimate users finding out that in the analyzed scenario bots had a marked impact on social media conversation, showing a significant activity and influence on legitimate users. The obtained results are based on a politically neutral research analysis that produces accurate estimates, which are in accordance with related work. In addition, it is worth noticing that, although our analysis discovered a high presence of social media bots that may have affected online political discussion, it is impossible to know who was running those bots, as they can also be exploited for provocative campaigns or as part of an information war.