Next Article in Journal
Extremal Matching Energy of Random Polyomino Chains
Next Article in Special Issue
Energy and Entropy Measures of Fuzzy Relations for Data Analysis
Previous Article in Journal
Testing the Beta-Lognormal Model in Amazonian Rainfall Fields Using the Generalized Space q-Entropy
Previous Article in Special Issue
Capturing Causality for Fault Diagnosis Based on Multi-Valued Alarm Series Using Transfer Entropy
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Do We Really Need to Catch Them All? A New User-Guided Social Media Crawling Method

1
Department of Computer Science and Engineering, Blekinge Institute of Technology, 371 79 Karlskrona, Sweden
2
Department of Computational Intelligence, Wrocław University of Science and Technology, 50-370 Wrocław, Poland
*
Author to whom correspondence should be addressed.
Entropy 2017, 19(12), 686; https://doi.org/10.3390/e19120686
Submission received: 18 October 2017 / Revised: 28 November 2017 / Accepted: 11 December 2017 / Published: 13 December 2017
(This article belongs to the Special Issue Entropy and Complexity of Data)

Abstract

:
With the growing use of popular social media services like Facebook and Twitter it is challenging to collect all content from the networks without access to the core infrastructure or paying for it. Thus, if all content cannot be collected one must consider which data are of most importance. In this work we present a novel User-guided Social Media Crawling method (USMC) that is able to collect data from social media, utilizing the wisdom of the crowd to decide the order in which user generated content should be collected to cover as many user interactions as possible. USMC is validated by crawling 160 public Facebook pages, containing content from 368 million users including 1.3 billion interactions, and it is compared with two other crawling methods. The results show that it is possible to cover approximately 75% of the interactions on a Facebook page by sampling just 20% of its posts, and at the same time reduce the crawling time by 53%. In addition, the social network constructed from the 20% sample contains more than 75% of the users and edges compared to the social network created from all posts, and it has similar degree distribution.

1. Introduction

In the age of big data, billion of people are using social media, such as Facebook, Snapchat, Twitter, and Instagram, to socialize, interact and create new content at a remarkable rate [1,2]. Facebook alone increased its number of users by 13% between 2015 and 2016, and in May 2017, the count had reached 1.75 billion active users [2]. This massive amount of data is now also available (to some extent) to crawl. However, with limited resources and due to the complexity and speed in which new content is generated, there is a need for improved strategies on what content to focus on.
In previous studies, we have developed the Social Interaction Network Crawling Engine (SINCE) [3,4] that collects publicly available Facebook data. Over a period of four years, we have collected content generated by 700 million unique Facebook users interacting on 280 million posts through 5 billion comments and 35 billion likes. The SINCE crawler is novel and unique as it is the first crawler capable of gathering data in depth by covering all interactions within posts. An important challenge throughout the crawling process is how to measure the quality of the collected data as it is based on different aspects related to the application for which the data is intended to be used, e.g., whether the data are used for measuring similarity among users’ interactions; whether the data provide diversified perspectives on certain topics; or whether the data comprise a statistically representative sample of the complete data.
Crawling data from social media comes with two inherent problems. First, that the data volume is so large that it is close to impossible to continuously gather all content. Secondly, that only a subset of the data is relevant for a specific application, or is interesting to researchers. The crawler used for collecting social media data in this work (i.e., SINCE) struggles under both these inherent problems. This is why this work introduces and evaluates a novel method for crawling social media data more efficiently, without requiring any priori knowledge about the network itself.
This study considers publicly available data published in open pages and groups on Facebook. The aim of the study is to investigate how to efficiently and precisely crawl quality data from Facebook’s social network using the introduced User-guided Social Media Crawling (USMC) method. We investigate if the novel prioritization and ranking techniques in USMC can be used to exclude posts that are of less interest during crawling, in order to both reduce the crawling time and at the same time increase the number of included social interactions. USMC ranks posts based on metadata metrics, such as the number of likes and then selects the highest ranked posts. Thus, these metadata metrics are used for estimating the number interactions that posts are likely to receive.
The goal of the study is to evaluate to what extent the proposed USMC method is able to estimate the importance of content by relying on the wisdom of the crowd and without any a priori knowledge about the underlying social network. That is, utilizing the users’ interactions in the social network for pointing the crawler to which data that is of most importance. Using this approach we then investigate the trade-off between crawling speed and degree of data coverage in the crawling process. Finally, the proposed USMC method is evaluated against a random sampling without replacement approach [5,6] as well as a novel chronological crawling approach where posts are sampled based on their lifetime, i.e., for how long posts have been active.

1.1. Motivation

Social media interactions, and especially Facebook data, have been growing massively during the last decade [2], and there is an interest from the research community, industry and the society at large to be able to collect and analyze these data [7]. The interaction data and online data in general are essential for social media analytic solutions, reputation tracking systems, brand monitoring and other big data solutions [7,8]. The fact that Facebook has no intention to sell these data (due to its value) has been a motivation for developing and presenting a novel way to collect social interaction data from publicly available pages.
Data from social media are big, both in terms of volume and in terms of velocity (new data are constantly created and grow faster than we can crawl). Since it is unfeasible to collect all data, there is a need to address the issue of how to prioritize the data while crawling. Thus, as much data as possible are collected for future research. This is especially important as the crawling make use of both limited time and computational resources. However, data from social media can be treated differently depending on the requirements associated with the intended use of the data, i.e., the future application. The users’ interactions are highly relevant for social network analysis in different areas, such as identification of important users, or seed selection for information and influence spreading in complex networks. Thus, we adopt a quantitative data measuring strategy by regarding the quality of the crawled data as equivalent to the proportion of all available social interactions in the social media services.
The USMC method is interesting for anyone with limited resources that systematically wants to collect content from social media services, or similar web-based sources. However, future users need to be aware of the limitation and potential bias enforced with the USMC method, i.e., that the resulting data exclude low interaction volumes. Examples of analyses made possible using data crawled by the USMC method include community detection [9] and identification of influential users [10,11].

1.2. Limitations

The USMC method, similar to all types of sampling, introduces some limitation/bias to the collected data. Fortunately, with the USMC method the bias introduced is known in advance, since the method disregards posts with low interactions and will most likely omit outliers and special cases present in low interaction posts. For other sampling approaches, the bias on the resulting data might not be known a priori, e.g., for chronological sampling the most recent posts are collected but nothing is known about how much of the interactions that are captured.

1.3. Related Work

There is a lack of research concerning the quality of data in social media and social network research. There are studies on social media and social networks, mainly using data from Twitter. These data are, however, typically collected using Twitter’s free garden hose API with a risk of being unbalanced and an unrepresentative sample of the complete data. Studies that investigate the quality of social media data include [12,13], where the former addresses how social media data from online recommendation systems can be evaluated. Sampling studies of social networks are quite common, including [14,15] that uses the original graph sampling study by Leskovec and Faloutsos [16] as a baseline. Wang et al. presents an interesting study [17] on how to efficiently sample a social network with a limited budget. The study uses metrics of the graph to make informed decisions on how to transverse it. More recently Rezvanian and Meybodi [18] presents algorithms to sample weighted networks. Chiericetti et al. [19] further investigate network sampling methods and how to minimize the number of required queries.
On the topic of graph and social media crawling, Zafarani et al. [7] present ways to evaluate and understand the data generated in social media. However, many social media crawling studies are obsolete due to updates by Facebook regarding of the default privacy policy of users’ content, which makes it impossible by default to access Facebook users’ content [13,20,21,22]. Consequently, the amount of private Facebook data that can be collected is severely limited. Furthermore, since Facebook does not sell any of its data there is a need for crawling methods that collect social interaction data from publicly available sources, which is the main motivation for this work.
Buccafurri et al. [23] discuss different methods to transverse social networks from a crawling perspective by focusing on public groups rather than individual users’ profiles. Our approach mainly differs from this study in two ways. First, we do not create a social network to transverse and only treat the social media as data, i.e., our proposed method does not require any knowledge of the underlying network. Secondly, we focus on user interactions represented as so called Social Interaction Network (SIN) graphs [24] as it shows the interactions between users in various communities, i.e., SIN graphs can represent interactions of all users on one particular newsgroup or users interacting on a specific topic. To conclude, there are no prior studies, according to our literature review, that address the challenge of collecting data from Facebook after Facebook changed the default privacy policy of its end-users’ content. Most studies use online data repositories and do not address the issue of how to efficiently collect data directly from Facebook, or other social media sites.

2. Results

To prioritize data available for crawling, we need to define a set of quality measures which will allow to rank the posts on a page. In this section, we start by testing which of the metadata metrics most accurately assess the importance of a post in terms of how much new knowledge about users’ interactions on that page it will convey. Next, the identified metadata metrics are used when evaluating to what extent the USMC method can increase the number of interactions collected by the crawler. Finally, we create a posteriori social networks to validate our findings with network theory.
The SINCE crawler starts by performing an initial crawl of a page, followed by a full crawl of its data [3,4]. During the initial crawl, the SINCE crawler gathers metadata for all posts on a page. For each post, the following three metadata metrics are collected: post lifetime, number of comments, and number of likes. An Ordinary Least Square (OLS) regression test [25] is used to investigate which of the three measures most accurately assesses the total number of interactions (i.e., the total count of likes and comments) on posts based on the sample of 160 randomly sampled Facebook pages. The basic statistics of this dataset is available in Table 1 and detailed descriptions of each page are presented in the Supplementary Information Table S2.
Figure 1 shows the distribution of R 2 for the conducted OLS regression test, which indicates a high confidence, 0.80 ± 0.26 (std), that the number of likes can be used to predict the number of interactions of posts. A combination of the three metrics gives the most accurate assessment, 0.86 ± 0.24 (std), as illustrated in Figure 1. However, in a practical setting, a combined metric is not possible because of mainly the following two reasons. First, the ratio between the metrics is unknown a priori to the crawling, which spoil any attempt to create a well-balanced combination of two or more metrics. Further, such a balanced combination is required since the number of likes is much higher than number of comments (as shown in Table 1), which means a simple sum of both metrics will not work as the number of likes will overshadow the number of comments. The second reason is because each metric has different variance per page. That is, each page would require its own tailored version of such combined metrics. Therefore, we deem that combined metrics for prioritizing which content to crawl is practically infeasible, and therefore turn to investigate each metric individually. However, the use of combined metrics could be interesting for future work, as such metrics still show best performance in the OLS analysis.
In Figure 1, there is a clearly visible separation between the distributions of each of the three metrics. A Friedman’s test χ 2 = 299.73 , d f = 2 , p 0.001 shows that there is indeed a statistical difference between the distributions of the three metadata metrics. However, there is no statistically significant difference between the number of likes and combined metrics. Further, a Nemenyi post-hoc test shows (as expected from Figure 1) that the number of likes metric is the strongest predictor for the number of interactions, and that all three distributions are statistically different at significance level 0.001.
As identified in the OLS regression analysis, the number of likes is a suitable predictor of the number of interactions on a post. Thus, we use it to rank posts for each page and use that ranking to guide the SINCE crawler on which posts to crawl. We compare the results in terms of number of collected interactions with a traditional random sampling without replacement [5,6] approach as well as a chronological crawling approach. The results presented in Figure 2a show that by implementing the USMC method it is possible to cover a vast majority of the interactions in a page by considering only a fraction of all available posts. For example, on average, we need to crawl merely 20% of the posts in order to gather 75% of all interactions when using the USMC method to rank posts based on their number of likes. In addition, a sample size of 20% covers only 20% and 40% of the pages’ interactions using random sampling and chronological crawling approaches respectively. For individual results of all 160 Facebook pages please see Table S1 and Figure S1.
Figure 2b illustrates the fraction of crawling time (x-axis) needed to collect a desired proportion of the interactions. It shows that it is possible to collect just over 50% of the interactions in less than 25% crawling time. That is, approximately twice as many interactions than collected by the random sampling and chronological crawling methods given the same crawling time. The number of interactions collected at any given crawling time has a linear relationship for the random sampling method. For the USMC method this relationship is more favorable when below roughly 80% crawling time. For crawling times longer than 80% the gained efficiency over the random sampling method decreases since the USMC method is gathering the posts that received the least interactions from the crawled page.
The Cohen’s d scores for the findings in Figure 2a show that there are large ( d > 0.8 ) separation between the three metadata metrics for all sample sizes smaller than 95%. Regarding the crawling time, the Cohen’s d scores show large differences between the metadata metrics for all crawling times shorter than 80%, and medium differences for crawling times between 80–95%.
Both Figure 2a,b show that the most efficient approach for USMC is to use the number of likes metric for ranking posts. Therefore, the next experiments only consider number of likes when comparing the USMC method with both the random sampling and the chronological crawling approaches.
To further validate the proposed USMC method, we investigate how complete and useful the resulting social networks are when constructed from the gathered data. Please note that due to the limitation in computational power we had to exclude the two largest pages from this analysis.
We have created three social networks based on the social interactions collected by the USMC method as well as by the random sampling and chronological crawling methods. For this, we relied on the following sample sizes: 1%, 10%, 20%, 30%, 60% and 90% of all posts on each Facebook page. Figure 3 shows the number of nodes (a) (Facebook users) and edges (b) (interactions between users) in each social network. It is clear that the USMC method both collects content from significantly more users as well as more social relations between them, compared to the other two methods. Thus, the social network constructed from the data crawled by the USMC method is more complete. In fact, even with merely 20% of the collected posts, it is possible to create a network that contains more than 75% of the users and their interactions.
Next, we performed a social network analysis with respect to degree distribution for each created network. Figure 4 presents the degree distribution for the three social networks created from the three representative Facebook pages. These three pages are representatives of the first quantile ( Q 1 ), the Median and the third quantile ( Q 3 ) regarding the number of posts per page distribution for 158 pages. Figure 4 includes measurements for the following four sample sizes: 10%, 20%, 30% and 60% out of all posts on each of the three Facebook pages. Figure 4 shows that, even with a relatively small sample of 20%, the USMC method is able to create a social network with more than 75% of all users and interactions included, and with a degree distribution very similar to the complete social network created from all available data. This result can be seen for all 158 Facebook pages that were analyzed (see Supplementary Material Figure S2 for the details).

3. Discussion

Many times when considering large-scale data gathering from social media services, it is not possible to collect all available data, as they are too large, and the continuous influx is simply too fast to keep up with. In those situations, one needs to decide on one of two available data gathering strategies: The deadline-based and the coverage-based strategies. Each of these strategies considers when the dataset is “good enough” for the intended use of the data. The deadline-based data collection strategy should be adopted when the data collection process has a point in time when it has to be finished, e.g., an upcoming presidential election in four weeks. Following that example, as much data as possible need to be collected within the given time frame, say three weeks. That way, strategic decisions based on the collected online behavior can help pinpoint which national regions to focus on during the last week before the election day.
The second type of data collection strategy is the coverage-based that specifies a particular sample size of the full dataset that is needed, e.g., that 75% of the original data are required for credibility of particular study. As an example of this strategy, think of a particular page that would take 100 days to crawl in full length. By using random sampling, a 75% sample would be reached in approximately 75 days, or the USMC method could be used that would collect the required 75% of the interactions in about 45 days. That is, by using USMC, a time-saving of 30 days could be expected (Figure 2b) when compared to random sampling, which is equivalent to a 33% time saving. Further, a time saving of 55 days (or 70%) could be expected by using USMC when compared to collecting the full dataset. Some might object that it is just matter of adding the tight amount of additional resources to speed up the process to solve the data gathering problem. However, often, this is not possibility due to either API restrictions or the equipment available. Thus it is important to study how prioritization of posts could be handled in order to determine where the available resources could be used most efficiently. For the USMC method, this translated to benefiting from the wisdom of the crowd of social media users by relying on their online behaviors for pointing the crawler to which content to target, and in which order.
The goal of covering as many interactions as possible with the limited resources is evaluated in Figure 2, which show how the proportion of collected interactions correlate with sample sizes for each of the investigated crawling approaches. For instance, ranking posts based on number of comments covers 78.5 ± 16.7 % (std) of the available interactions on a given page, at a sample size of 40% of all posts on that given page. However, ranking posts based on number of likes provides an increased coverage with 86.7 ± 12.5 % (std) of all interactions at the same sample size. Figure 2b shows the interaction coverage with respect to crawling time for SINCE crawler. These results show that it is possible to decrease the overall crawling time if only the posts that covers the most number of interactions are being crawled, i.e., excluding the posts with least number of interactions.
The evaluation of the UMSC method has revealed that it is a suitable candidate for crawling high-volume data sources from social media services. However, it could be wise to consider other crawling approaches where the a posteriori data analysis is dependent on the interactions on posts with low number of interactions, e.g., Spam mitigation approaches, malicious content detection, or outlier analysis. However, for other application areas, the USMC approach is interesting to consider, e.g., community detection analysis, or identification of influential users.
In this study, the USMC method has been evaluated on data from public Facebook pages. However, it is likely that the same approach could be used for other social media services as well, e.g., Twitter, LinkedIn or ResearchGate. For Twitter, we could rank tweets using the number of re-tweets, likes and responses. Ranking by these attributes would probably allow the collection of social interactions from Twitter to be carried out more efficiently, compared to approaches used today, and at the same time produce representative samples. Similarly, USMC could be applied on the social network at ResearchGate by ranking content based on the number of comments, RG-score, h-index or average number of downloads per article. However, these suggestions need to be validated using research on other social media platforms.

4. Materials and Methods

In this section, the materials and methods used in this article are described. First, a detailed description of the proposed USMC method is given. Second, the dataset used in the evaluation of the proposed approach is presented. Third, the evaluation methods used in the study are detailed. Fourth, the process of creating a social network from the dataset as well as the social network analysis carried out on that network are being presented. Finally, we describe the various statistical tests used in this study.

4.1. User-Guided Social Media Crawling

As users interact on social media it is possible to use their actions (e.g., likes or comments) to rank posts. Evaluating data from social media can be made in various forms, but it is hard to computationally evaluate the content. This is why the work proposed in this study makes use of users’ actions in order to make more informed decisions about the social media data, i.e., benefiting from wisdom of the crowd. Users’ actions on posts could be used as indicators of how interesting posts are for the users in the particular community (different communities can have different values and understandings of the subject). The proposed USMC method therefore relies on “wisdom of the Facebook crowd” to find quality content in social networks as well as a way to rank the posts to capture. In general, the introduced crawling technique ranks content in the social network according to how much attention users give it, i.e., how much interaction each content receives.
In this work, we define social interactions as the type of actions users can take on content in the social network. To put it in Facebook’s terminology, the content is usually a post within a page and the actions are either a like on a post, a comment on a post or a like on a comment on a post. Figure 5 illustrates an example of different social interactions as well as how the three sampling methods evaluated in this study can be used for collecting those interactions. This example will be used throughout this section as a platform for describing and discussing various aspects in the crawling process.
In detail, the USMC enabled crawling process works as follows. First, the crawler makes a quick initial crawl of a page to gather the metadata for each post in that particular page. Next, the USMC method estimates the total numbers of interactions each post will receive during a given time interval and based on a chosen metric, and then sorts all posts in a list by decreasing order. Regarding the metric used for predicting the total number of interactions that post is likely to receive, the number of likes a post has received (which is available in the meta data) has proven to be a suitable metric. Next, the actual crawling of content from the page starts and continues until either the desired number of interactions has been reached, or the time limit is passed. For each iteration in this process the crawler selects the top most post from the list and carry out a full crawl for that particular post. A complete description of USMC enabled crawling process used by the SINCE crawler is shown in Algorithm 1.
Algorithm 1 USMC enabled crawling with the SINCE crawler.
  • Require: p a g e _ i d
  •    p o s t _ m e t a - d a t a collect_post_metadata ( p a g e _ i d )
  •   sort ( p o s t _ m e t a - d a t a ) based on USMC
  •  
      repeat
     
  •        p o s t _ d a t a makeFacebookRequest (/ p o s t _ i d )
  •        d a t a d a t a + p o s t d a t a
  •       repeat
     
  •             l i k e s makeFacebookRequest (/ p o s t _ i d /likes)
  •             d a t a d a t a + l i k e s
  •       until l i k e s is empty
  •       repeat
     
  •             c o m m e n t s makeFacebookRequest (/ p o s t _ i d /comments)
  •             d a t a d a t a + c o m m e n t s
  •            if c o m m e n t s has l i k e s then
  •  
                    repeat
     
  •            c o m m e n t L i k e s makeFacebookRequest (/ c o m m e n t _ i d /likes)
  •            d a t a d a t a + c o m m e n t L i k e s
  •                 until c o m m e n t L i k e s is empty
  •             end if
  •           until c o m m e n t s is empty
  •           saveData ( d a t a )
  •         until time is up required data is collected
The social interactions are exemplified in the toy example shown in Figure 5. The eight posts in Figure 5a include different number of interactions with regards to likes (shown in red next to the “thumb up” icon) and comments (shown in green number next to “speech bubble” icon). Figure 5b shows a bipartite network of the interactions between six users ( U 1 6 ) and each respective post, where green edges represent comments from users on a particular post and red edges represent likes. Figure 5c shows the aggregated network built on users’ interactions on posts collected by the following three sampling approaches: USMC, chronological and random sampling. The full network from all eight posts is shown as dashed edges, while the collected interactions are shown as solid lines. Red edges denote likes on posts and green edges denote comments on posts. Figure 5d shows the social networks created based on a 37.5% sample of all posts when collected by each of the three crawling methods, i.e., USMC, chronological and random sampling. The social network in Figure 5d is created as a projection from the bipartite networks shown in Figure 5b where the nodes are representing users, and where edges are present if the users have interacted on the same post.

4.2. Dataset

The dataset used for evaluating the USMC method was created by collecting 160 randomly selected open pages on Facebook. The dataset is available on Dataverse [26]. Table 1 shows descriptive statistics for these 160 pages included in the study. The SINCE crawler [3,4] was used for collecting the 160 Facebook pages between July 2014 and May 2016. SINCE is designed to collect publicly open pages using Facebook’s API. We adhere to Facebook’s data privacy policy [27] by anonymizing all data to an extent where it is only possible to backtrack the particular public page that is analyzed. The resulting dataset has a median page size of 5235 posts, 180,314 users, 45,592 comments and 442,424 likes. In total, the dataset includes some 368 million unique users interacting in little over 1.3 billion social interactions. However, it should be noted that 2 out of the 160 pages had to be excluded from the network analysis part as social networks could not be generated with the hardware resources available since 148 GB of RAM was not enough to fit the projected network. For complete statistics of all pages, please see Table S1 in the supplementary material.

4.3. Evaluation Methods

We evaluate the USMC method by comparing it to both traditional random sampling without replacement and chronological methods. Random sampling [5,6] in this context is about collecting posts at random, which gives a representative representation of the data (sampled data will represent the original dataset given the current sample size). During the evaluation each random sampling execution was iterated 100 times and the results report the mean and standard deviation as common in scientific work. The chronological method sorts all posts in decreasing order from oldest to the newest, and samples the oldest posts. Looking at the conceptual example in Figure 5 when having specified a sample size of 37.5% of all posts, i.e., 3 out of the 8 posts, the USMC will collect the three posts with highest number of likes, i.e., posts 1, 6 and 7, while the chronological method will collect posts 1, 2 and 3 and the random sampling collects for instance posts 2, 5 and 7.
Each page is evaluated with regards to the number of interactions they capture. Five different sample sizes (10%, 20%, 30%, 60% and 90% of all posts at the Facebook page) are used to represent the page. In the evaluation we also investigate the time it takes to crawl the 160 Facebook pages. In the example in Figure 5, each method produces a different set of posts: { 1 , 6 , 7 } for USMC, { 1 , 2 , 3 } for chronological and { 2 , 5 , 7 } for random sampling. These total number of interactions included in each set of posts differs, as can be seen in Figure 5c, where USMC captures 77% of all interactions while chronological and random sampling captures 32% and 27% respectively.

4.4. Social Network Analysis

The three methods are evaluated by comparing the social networks created from the interactions collected by each method. In these social networks, users are represented as nodes and the edges between them represent social interactions. A social network is created as a undirected graph as G = < N , E > , with a set of nodes N = { n 1 , , n n } to represent users and a set of edges E = { < n i , n j > : n i , n j N i j } representing relationship between the users i and j. The social network of interactions between users is projected from the bipartite network of users and posts, where an edge < n i , n j > is present if both of the users i and j have commented on the same post.
Figure 5d shows the resulting social networks created by three crawling methods (USMC, chronological and random) using the same sample size of 37.5% of all posts, i.e., of 3 out of the 8 posts. It is clear that USMC creates the most complete network since it includes all of the six existing users and 94% of edges. The chronological and random methods include 67% and 50% of the nodes (users), and 28% and 22% of the edges respectively. Please note that the example shows a multilayer social network with two types of edges based on: (i) likes; and (ii) comments represented in the form of a multi-graph Figure 5d. However, as we have mentioned earlier, the social network used in our experiments is a single layer network where edges are based on comments only.

4.5. Statistical Tests

The statistical tests used for evaluation purposes are as follows. First, we used an ordinary least square regression test [5] to investigate which metadata metrics (out of post lifetime, number of comments, and number of likes) was most accurate in predicting the number of interactions on posts. Secondly, the non-parametric Friedman test [6] was used to identify overall differences in the data since it is not normally distributed. Thirdly, a Nemenyi post-hoc test [6] was used to identify individual differences between metadata metrics. Fourth, Cohen’s d [28] was used to quantifying the difference between means. Finally, all reporting of results includes standard measurements such as the test statistic, p-value, mean/median and standard deviation.

5. Conclusions

This work introduces the novel USMC method for efficient crawling of data from social network services by utilizing a wisdom of the crowd approach by allowing the users’ interactions to guide the crawler to find content to crawl. The evaluation of the proposed USMC method, through social network analysis [8,29,30], shows that it can cover in excess of 80% of the social network by collecting merely 30% of the available posts on a page. The social networks constructed from the collected data are shown to have close to identical degree distribution already at as low sampling sizes as 20% compared to the whole page, which indicates that the social network structure of the 20% sample is nearly identical to the complete page. That is, both the number of users and the number of social relations are similar, as well as the network degree distribution. This corroborates that the USMC method could be a powerful addition to the already available methods for crawling social network services when as many social interactions as possible are needed, while still maintaining a close to identical social network in terms of the network degree distribution.

Supplementary Materials

The following are available online at www.mdpi.com/1099-4300/19/12/686/s1, Table S1: Statistics for each of the analyzed Facebook pages, Table S2: Interaction coverage for each of the analyzed Facebook pages, Figure S1: Ranking effect for each of the analyzed pages, Figure S2: Degree distribution for each of the analyzed Facebook pages using different sampling methods.

Acknowledgments

This work was partially supported by The Polish National Science Centre, the decision No. DEC-2016/21/D/ST6/02408 and by the European Union’s Horizon 2020 research and innovation programme under the Marie Skłodowska-Curie grant agreement No. 691152 (RENOIR) and the Polish Ministry of Science and Higher Education fund for supporting internationally co-financed projects in 2016-2019 (agreement No. 3628/H2020/2016/2).

Author Contributions

Fredrik Erlandsson created an initial concept of user-guided social media crawling; Fredrik Erlandsson and Piotr Bródka developed the concept to its current state; Fredrik Erlandsson and Piotr Bródka designed the experiments; Fredrik Erlandsson implemented and executed all experiments and simulations; Fredrik Erlandsson, Piotr Bródka and Martin Boldt analyzed data and discussed results; All authors drafted, critically reviewed the manuscript and approved the final version.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Twitter, Company | About. Available online: https://about.twitter.com/company/ (accessed on 12 December 2017).
  2. Facebook, Company Info | Facebook Newsroom. Available online: http://newsroom.fb.com/company-info/ (accessed on 3 October 2017).
  3. Erlandsson, F.; Nia, R.; Boldt, M.; Johnson, H.; Wu, S.F. Crawling Online Social Networks. In Proceedings of the Network Intelligence Conference, Karlskrona, Sweden, 21–22 September 2015; pp. 9–16. [Google Scholar]
  4. Erlandsson, F.; Wu, F.S. SocialCrawler 2.9. 2016. Available online: https://doi.org/10.5281/zenodo.153825 (accessed on 12 December 2017).
  5. Walpole, R.; Myers, R.; Sharon, M.; Ye, K. Probability & Statistics—For Engineers and Scientists; Pearson, Cambridge University Press: Upper Saddle River, NJ, USA, 2012. [Google Scholar]
  6. Sheskin, D.J. Handbook of Parametric and Nonparametric Statistical Procedures, 5th ed.; Chapman & Hall/CRC: Washington, DC, USA, 2011. [Google Scholar]
  7. Zafarani, R.; Abbasi, M.A.; Liu, H. Social Media Mining, An Introduction; Cambridge University Press: New York, NY, USA, 2014. [Google Scholar]
  8. Kietzmann, J.H.; Hermkens, K.; McCarthy, I.P.; Silvestre, B.S. Social media? Get serious! Understanding the functional building blocks of social media. Bus. Horiz. 2011, 54, 241–251. [Google Scholar] [CrossRef]
  9. Nia, R.; Erlandsson, F.; Johnson, H.; Wu, S.F. Leveraging social interactions to suggest friends. In Proceedings of the 2013 IEEE 33rd International Conference on Distributed Computing Systems Workshops (ICDCSW), Philadelphia, PA, USA, 8–11 July 2013; pp. 386–391. [Google Scholar]
  10. Erlandsson, F.; Borg, A.; Johnson, H.; Bródka, P. Predicting User Participation in Social Media. In Advances in Network Science; Springer International Publishing: Cham, Switzerland, 2016; pp. 126–135. [Google Scholar]
  11. Erlandsson, F.; Bródka, P.; Borg, A.; Johnson, H. Finding Influential Users in Social Media Using Association Rule Learning. Entropy 2016, 18, 164. [Google Scholar] [CrossRef]
  12. Agichtein, E.; Castillo, C.; Donato, D.; Gionis, A.; Mishne, G. Finding high-quality content in social media. In Proceedings of the 2008 international conference on web search and data mining, Palo Alto, CA, USA, 11–12 February 2008; pp. 183–194. [Google Scholar]
  13. Mislove, A.; Marcon, M.; Gummadi, K.P.; Druschel, P.; Bhattacharjee, B. Measurement and analysis of online social networks. In Proceedings of the 7th ACM SIGCOMM conference on Internet measurement, San Diego, CA, USA, 23–26 October 2007; pp. 29–42. [Google Scholar]
  14. Gjoka, M.; Kurant, M.; Butts, C.T.; Markopoulou, A. Walking in Facebook: A Case Study of Unbiased Sampling of OSNs. In Proceedings of the 2010 IEEE Conference on Computer Communications (INFOCOM), San Diego, CA, USA, 14–19 March 2010; pp. 1–9. [Google Scholar]
  15. Gjoka, M.; Butts, C.T.; Kurant, M.; Markopoulou, A. Multigraph Sampling of Online Social Networks. IEEE J. Sel. Areas Commun. 2011, 29, 1893–1905. [Google Scholar] [CrossRef]
  16. Leskovec, J.; Faloutsos, C. Sampling from large graphs. In Proceedings of the 12th ACM SIGKDD International Conference, New York, NY, USA, 20–23 August 2006; pp. 631–636. [Google Scholar]
  17. Wang, X.; Ma, R.T.B.; Xu, Y.; Li, Z. Sampling online social networks via heterogeneous statistics. In Proceedings of the 2015 IEEE Conference on Computer Communications (INFOCOM), Hong Kong, China, 26 April–1 May 2015; pp. 2587–2595. [Google Scholar]
  18. Rezvanian, A.; Meybodi, M.R. Sampling algorithms for weighted networks. Soc. Netw. Anal. Min. 2016, 6, 60. [Google Scholar] [CrossRef]
  19. Chiericetti, F.; Dasgupta, A.; Kumar, R.; Lattanzi, S.; Sarlós, T. On Sampling Nodes in a Network. In Proceedings of the 25th International Conference on World Wide Web (WWW’ 16), Montréal, QC, Canada, 11–15 April 2016; pp. 471–481. [Google Scholar]
  20. Catanese, S.A.; de Meo, P.; Ferrara, E.; Fiumara, G.; Provetti, A. Crawling Facebook for social network analysis purposes. In Proceedings of the International Conference on Web Intelligence, Mining and Semantics, Sogndal, Norway, 25–27 May 2011; pp. 52:1–52:8. [Google Scholar]
  21. Wilson, C.; Sala, A.; Puttaswamy, K.P.N.; Zhao, B.Y. Beyond social graphs: User interactions in online social networks and their implications. ACM Trans. Web 2012, 6, 17–31. [Google Scholar] [CrossRef]
  22. Crnovrsanin, T.; Muelder, C.W.; Faris, R.; Felmlee, D.; Ma, K.L. Visualization techniques for categorical analysis of social networks with multiple edge sets. Soc. Netw. 2014, 37, 56–64. [Google Scholar] [CrossRef]
  23. Buccafurri, F.; Lax, G.; Nocera, A.; Ursino, D. Moving from social networks to social internetworking scenarios: The crawling perspective. Inf. Sci. 2014, 256, 126–137. [Google Scholar] [CrossRef]
  24. Nia, R.; Erlandsson, F.; Bhattacharyya, P.; Rahman, M.R.; Johnson, H.; Wu, S.F. Sin: A platform to make interactions in social networks accessible. In Proceedings of the 2012 International Conference on Social Informatics (SocialInformatics), Lausanne, Switzerland, 14–16 December 2012; pp. 205–214. [Google Scholar]
  25. Davidson, R.; MacKinnon, J.G. Econometric Theory and Methods; Oxford University Press: New York, NY, USA, 2004. [Google Scholar]
  26. Erlandsson, F. Replication Data for: Do We Really Need to Catch Them All? A New User-Guided Social Media Crawling Method. 2017. Available online: http://dx.doi.org/10.7910/DVN/DCBDEP (accessed on 12 December 2017).
  27. Facebook. Facebook Data Policy. Available online: https://www.facebook.com/full_data_use_policy (accessed on 12 December 2017).
  28. Cohen, J. Statistical Power Analysis for the Behavioral Sciences, revised ed.; Academic Press: Cambridge, MA, USA, 1977. [Google Scholar]
  29. Safko, L. The Social Media Bible: Tactics, Tools, and Strategies for Business Success; John Wiley & Sons: Hoboken, NJ, USA, 2012. [Google Scholar]
  30. Saganowski, S.; Gliwa, B.; Bródka, P.; Zygmunt, A.; Kazienko, P.; Koźlak, J. Predicting Community Evolution in Social Networks. Entropy 2015, 17, 30–53. [Google Scholar] [CrossRef]
Figure 1. This box plot presents the R 2 distribution of OLS Regression test assessing the number of interactions on a post, using the following three metadata metrics: post lifetime (blue), number of comments (green), and number of likes (red). The three metrics are also shown as a combined metric (purple). All box-plots are created from a sample of 160 Facebook pages.
Figure 1. This box plot presents the R 2 distribution of OLS Regression test assessing the number of interactions on a post, using the following three metadata metrics: post lifetime (blue), number of comments (green), and number of likes (red). The three metrics are also shown as a combined metric (purple). All box-plots are created from a sample of 160 Facebook pages.
Entropy 19 00686 g001
Figure 2. Average interaction coverage for all 160 Facebook pages for: (a) different sample size (represented as a percentage of all post on Facebook a page); (b) crawling time. It is presented for the USMC method with rankings based on number of likes (red), number of comments (blue), post lifetime (green), chronological crawling (purple), and random sampling method (yellow). For the best and the worst approach, we included error bars showing the standard deviation. The individual results for all 160 Facebook pages are available in the supplementary material, see Figure S1.
Figure 2. Average interaction coverage for all 160 Facebook pages for: (a) different sample size (represented as a percentage of all post on Facebook a page); (b) crawling time. It is presented for the USMC method with rankings based on number of likes (red), number of comments (blue), post lifetime (green), chronological crawling (purple), and random sampling method (yellow). For the best and the worst approach, we included error bars showing the standard deviation. The individual results for all 160 Facebook pages are available in the supplementary material, see Figure S1.
Entropy 19 00686 g002
Figure 3. The fraction of all nodes (a) and edges (b) for the social networks constructed from the 1%, 10%, 20%, 30%, 60% and 90% samples of all available posts, using USMC (red), chronological crawling (purple), and random sampling (yellow) approaches. The plot is showing mean and standard deviation for all 160 pages.
Figure 3. The fraction of all nodes (a) and edges (b) for the social networks constructed from the 1%, 10%, 20%, 30%, 60% and 90% samples of all available posts, using USMC (red), chronological crawling (purple), and random sampling (yellow) approaches. The plot is showing mean and standard deviation for all 160 pages.
Entropy 19 00686 g003
Figure 4. The degree distribution for three social networks created out of three representative Facebook pages. Each column shows the sample size (10%, 20%, 30% and 60% of all posts on a page), and each row presents a representative page for each quantile in the number of posts per page distribution for 158 pages. The first row shows the Facebook page Chateau Elan Winery & Resort (https://www.facebook.com/chateauelan/) with 1131 posts, 25,008 users, and 4814 comments ( Q 1 ); the second Liftopia (https://www.facebook.com/liftopia/) with 3973 posts, 47,001 users, and 50,065 comments (Median); and the third San Francisco | The Official Guide (https://www.facebook.com/onlyinSF/) with 13,305 posts, 735,183 users, and 116,336 comments ( Q 3 ). As shown, a 20% sample of all pages collected by the USMC method allows for the creation of social networks that have almost identical degree distribution compared to social networks created from all available data.
Figure 4. The degree distribution for three social networks created out of three representative Facebook pages. Each column shows the sample size (10%, 20%, 30% and 60% of all posts on a page), and each row presents a representative page for each quantile in the number of posts per page distribution for 158 pages. The first row shows the Facebook page Chateau Elan Winery & Resort (https://www.facebook.com/chateauelan/) with 1131 posts, 25,008 users, and 4814 comments ( Q 1 ); the second Liftopia (https://www.facebook.com/liftopia/) with 3973 posts, 47,001 users, and 50,065 comments (Median); and the third San Francisco | The Official Guide (https://www.facebook.com/onlyinSF/) with 13,305 posts, 735,183 users, and 116,336 comments ( Q 3 ). As shown, a 20% sample of all pages collected by the USMC method allows for the creation of social networks that have almost identical degree distribution compared to social networks created from all available data.
Entropy 19 00686 g004
Figure 5. Example of interactions extracted from posts. (a) shows eight different posts with number of likes (“thumbs-up” icon), number of comments (“speech bubble” icon), and the age of the posts (“watch” icon); (b) shows the bipartite networks of interactions between the six users ( U 1 6 ) and the eight posts ( P 1 8 ), where red edges denote likes on posts and green edges denote comments on posts. The users are the same on all posts; (c) shows the aggregated networks of user’s interactions towards posts collected by three different crawling methods: USMC, chronological and random sampling. The full network from all eight posts is shown as dashed edges, while the collected interactions are shown as solid lines. Red edges denote likes on posts and green edges denote comments on posts; (d) shows the social networks created based on a 37.5% sample of all posts, collected by the three crawling methods: USMC, chronological and random sampling. The full social network from all eight posts is shown as dashed edges, while the collected edges are shown as solid lines.
Figure 5. Example of interactions extracted from posts. (a) shows eight different posts with number of likes (“thumbs-up” icon), number of comments (“speech bubble” icon), and the age of the posts (“watch” icon); (b) shows the bipartite networks of interactions between the six users ( U 1 6 ) and the eight posts ( P 1 8 ), where red edges denote likes on posts and green edges denote comments on posts. The users are the same on all posts; (c) shows the aggregated networks of user’s interactions towards posts collected by three different crawling methods: USMC, chronological and random sampling. The full network from all eight posts is shown as dashed edges, while the collected interactions are shown as solid lines. Red edges denote likes on posts and green edges denote comments on posts; (d) shows the social networks created based on a 37.5% sample of all posts, collected by the three crawling methods: USMC, chronological and random sampling. The full social network from all eight posts is shown as dashed edges, while the collected edges are shown as solid lines.
Entropy 19 00686 g005
Table 1. Descriptive statistics of the dataset of 160 pages.
Table 1. Descriptive statistics of the dataset of 160 pages.
MetricMeanStd.Min Q 1 Median Q 3 MaxSum  
Posts21,59060,78671313523514,758470,5283,454,456   
Users2,588,33811,460,28118226,589180,314897,564113,379,978414,134,086 *
Comments608,8732,584,4143710,63845,592230,20527,550,35297,419,710   
Likes7,640,85833,972,67438454,923442,4242,589,165308,495,9881,222,537,425   
Edges98,688,366395,490,320383,0801,331,53924,018,804   4,238,052,18915,592,761,948   
Nodes154,831382,03125561324,461115,2263,020,78624,463,392   
* Of which 368,094,952 are unique users, not overlapping on different pages.

Share and Cite

MDPI and ACS Style

Erlandsson, F.; Bródka, P.; Boldt, M.; Johnson, H. Do We Really Need to Catch Them All? A New User-Guided Social Media Crawling Method. Entropy 2017, 19, 686. https://doi.org/10.3390/e19120686

AMA Style

Erlandsson F, Bródka P, Boldt M, Johnson H. Do We Really Need to Catch Them All? A New User-Guided Social Media Crawling Method. Entropy. 2017; 19(12):686. https://doi.org/10.3390/e19120686

Chicago/Turabian Style

Erlandsson, Fredrik, Piotr Bródka, Martin Boldt, and Henric Johnson. 2017. "Do We Really Need to Catch Them All? A New User-Guided Social Media Crawling Method" Entropy 19, no. 12: 686. https://doi.org/10.3390/e19120686

APA Style

Erlandsson, F., Bródka, P., Boldt, M., & Johnson, H. (2017). Do We Really Need to Catch Them All? A New User-Guided Social Media Crawling Method. Entropy, 19(12), 686. https://doi.org/10.3390/e19120686

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop