1. Introduction
Although state-of-the-art sensors can detect various natural disasters in advance (e.g., Mexico City’s alarm system can timely sense earthquakes originating in the southern states) [
1], the devastating consequences of these events in urban areas are usually severe. The relief efforts during and after a disaster are essential for minimizing their negative impact. These efforts are largely the result of motivating the civil society to collaborate with rescue teams, public protection agencies, and security organizations to inform, rescue, and provide restoration. The active participation of civilians in the aftermath not only strengthens the society’s resiliency to a natural disaster but also improves the reliability of the information obtained from non-traditional sources [
2,
3]. For example, thanks to widespread wireless communication networks and mobile technologies, the dissemination of digital information now serves as a vital way to contact aid services and make appropriate decisions in a fast and more flexible manner [
4]. As an example, in the 2010 earthquake in Haiti, the use of instant messages sent by civilians from different locations facilitated the reporting of trapped individuals, the provision of medical assistance, and the delivery of basic needs, such as food, water, and shelter [
5]. Personal mobile phones can also be used by survivors to send messages to their relatives and the community at large about their current status, and this information can eventually be forwarded to rescue teams.
Figure 1 illustrates an example of an earthquake survivor using their mobile phone to communicate with relatives.
Personal mobile devices can be linked to Online Social Networks (OSNs) and enable synchronization among applications, e.g., Twitter, Facebook, and Instagram, which allows users to post and update their activities in real time [
7,
8]. The creation and prevalence of user-generated content [
9] may include temporal and spatial information associated with different events of interest [
10]. For the most part, this information is represented by georeferenced patterns that establish a relationship between the posted event and spatiotemporal characteristics of the publishing entity. As an example, an update (tweet) on Twitter that includes temporal and spatial information is shown in
Figure 2.
The dynamics of OSN users and their continuous status updates, along with numerous kinds of attachments, such as photos, videos, and documents, can be considered as a social sensor because the data generated on a large scale closely resembles that acquired by traditional sensor systems [
11,
12]. Below are some characteristics that reinforce the notion that OSNs can be treated as social sensors [
13,
14]:
Sensor operation: Sensors acquire data from various events as a result of observations. For example, smartphones are equipped with cameras, so users are able to obtain, process, and transmit data in real-time [
12,
15].
Processing of sensed data: When the information acquired by traditional sensing systems is processed, geographic information is available if navigation systems, e.g., GPS, are used. Information posted on OSNs may include either specific locations or textual descriptions of a place during an event. Moreover, users can reply, comment, and retransmit an update [
16].
Twitter has been popularized for the ease of reading, writing, and collecting data, which are published on a constant basis. Twitter allows users to publish opinions, sentiments, and observations, as well as update their statuses in an asymmetrical form (unlike other OSNs, such as Facebook, a Twitter user’s newsfeed, mentions, and replies remain public by default). Recently, Twitter has been the center of attention in different research fields related to Marketing, Social Sciences, Natural Language Processing (NLP), Opinion Mining, and Predictive analysis [
17]. Additionally, several applications are being developed to analyze Twitter data related to daily-life matters. For example, during electoral events, the work in [
18] confirmed that a high rate of tweets posted by users shows a correlation with the performance of candidates and the public’s preferences. Event prediction and monitoring can then be carried out by applying connective action theory that links a live event with the reactions of users [
19]. For example, it has been demonstrated that events with a negative impact on society can motivate hacker activists to perpetrate cyber attacks [
20]. Twitter can then be used as an alternative engine for exchanging information related to natural disasters, such as fires, floods, hurricanes, and earthquakes. Moreover, recent research has demonstrated [
21] that Twitter can also be a source of information for spreading awareness of ecological phenomena with well-defined temporal patterns.
In this work, a methodology is proposed that uses Twitter as a social sensor for natural disasters by exploiting the spatial and temporal information associated with the observations and experiences posted by users. The aim of our social sensor is to provide useful geo-temporal patterns that may appear during and after the occurrence of an event, which may be useful to assess the extent of the damages.
By default, tweets are short messages of a maximum of 280 characters in length. Tweets can include well-defined geographic data provided by GPS or manual check-ins. However, it has been reported that only a very small percentage of Twitter users use navigation systems or register places to reference their status [
22]. Given this difficulty in determining location, some studies have proposed estimating the location of a tweet by exploiting some of Twitter’s available features, including searching for updates related to certain events within a known geographical region [
23], grouping textual patterns associated with user language [
24], and parsing Twitter geo-objects to calculate the approximate coordinates from statuses that depict well-defined places [
25]. Further, to tackle these limitations, the textual content of a tweet can be examined to determine whether a location is mentioned.
An important contribution of our work is to expand on the idea of examining the textual content of tweets by inspecting the so-called
toponyms (places implicitly described in a text) from the surge of tweets that emerge during and after a natural disaster. To this end, our proposed approach employs Named Entity Recognition (NER), which is an information extraction method for finding and sorting named entities into pre-defined tags (persons, locations, and organizations) [
26,
27]. This is achieved by breaking down tweets into word units and classifying them into named entity tags so that a
toponym can be discovered and geocoded (estimating its spatial information in terms of latitude and longitude coordinates). Detecting places is not a trivial task, and major challenges associated with tweets must be addressed, such as the ungrammatical nature of tweets, as well as informal abbreviations and lexicons (for example, mentioning a location using a hashtag). With respect to temporal information, we cluster values of the time and duration of tweets connected to the event of interest by similarity within a window of time [
28]. To capture the semantic, morphological, and contextual richness of each word in a tweet, we perform a word-level analysis by using Word Embeddings [
29,
30], a widely used algorithm that transforms similar words into a continuous vector space. A sentence-level analysis is subsequently performed to extract semantic and syntactic information from each tweet by employing a Bidirectional Long Short-Term Memory (biLSTM) network [
31,
32], which is capable of using long-ranged symmetric sequence contexts. After training a Conditional Random Field (CRF) classifier [
33] with biLSTM output sequences and their corresponding NER target classes, our methodology predicts locations from tweets. Finally, it applies a Kernel Density Estimation (KDE) algorithm [
34] to the classified locations to compute various hotspot heat maps for the event of interest.
We have tested the proposed sensor with (Spanish) tweets from the 2017 Mexico City earthquake. Based on our evaluations, our sensor can accurately capture information that can help authorities, institutions, and volunteers to detect major risk areas and locate missing individuals and shelters.
2. Related Work
The detection of events related to natural disasters using OSN data has been the subject of recent research in the fields of sensors, natural language processing, and automatic and statistical learning. The common goal is to detect, monitor, and disseminate information about the event in a timely manner with some degree of trust. As described in [
35], Twitter has been recently used as a platform to post diverse information related to various natural disasters, such as wildfires [
36], floods [
35], hurricanes [
37], and earthquakes [
38], and it has resulted in situational awareness.
Table 1 summarizes the contributions of important works that employed data extracted from Twitter and other OSNs to sense natural disasters.
Although the state of the art provides significant advances in geoparsing
toponyms from OSNs using NER techniques [
39,
40,
41,
42,
43,
44], some challenges still exist. Important challenges are described below:
Vector space feature representations: a vector space model can capture the relevance of words by assigning them a numerical weight; then, each sentence can be represented as a sparse or dense vector of a vocabulary of size
V. Some algorithms include One-hot-Encoding, Bag-of-Words, and Tf–IDF (Term frequency–Inverse document frequency) [
47]. Such type of codification may fail to preserve semantic, syntactic, and linguistic features, as it cannot establish relationships and similarity patterns among words in a given corpora, making it difficult to examine transitions between contiguous data.
Algorithm selection: Geoparsing techniques based on NER require a suitable algorithm with minor preprocessing to train sequential structures such as tweets; more specifically, long contextual information should be considered in both directions of a word of interest. For this reason, approaches that employ SVM, Feed-Forward Neural Networks, Decision Trees, and single CRF classifiers may be unsuccessful as they assume that words are independent of each other, and rely on previous feature extraction steps. Recent approaches based on feed-forward algorithms for NER classification may have several disadvantages, as tabulated in
Table 2.
In order to clearly define each step of the proposed methodology, the following research questions are raised:
How can the semantic, morphological, and linguistic textual patterns that properly represent a word and its surrounding context be preserved?
How can a sequential labeling problem such as NER be addressed by capturing contextual information in both directions of a word of interest, and how can it be classified as a toponym?
Why is it important to scrutinize neighboring named entity tags as state sequences at a sentence level?
How should geocoded data and clustered temporal information be statistically scored to depict the dynamics during and after an event?
This work aims to answer these questions. Our main contributions are summarized as follows: (1) A text preprocessing module to remove noisy textual features; (2) Word embedding representations to depict each word of interest, keeping the semantic, syntactic, and linguistic relevance; (3) An NER-based geoparsing strategy (toponym extractor) based on a Recurrent Neural Network (RNN) with a CRF output layer to determine the word embeddings that form a tweet and their mapped states (named entity tags); (4) A Geocoder to query Google Maps API with each toponym, thus presenting results in latitude and longitude values; (5) A KDE algorithm to graphically depict hotspots from clusters of geocoded toponyms in the same spatial area during and after the event of interest.
5. Sensing Information: A Case Study of the 2017 Mexico City Earthquake
On 19 September 2017 at 1:14 p.m. CST, an earthquake with a 7.1 magnitude on the Richter scale with an epicenter in Axochiapan, Morelos, a state adjacent to Mexico City, impacted the urban infrastructure of the city and surrounding areas. Although the alarm system is efficient when epicenters occur on the Pacific Ocean coast, in the particular case of this natural disaster, the evacuations took place 11 s after the earthquake started because of the lack of sensors near the metropolitan area. It was not to be expected that Twitter users would report information related to the disaster zones. In addition to army and navy personnel, a large number of individuals took to the streets to offer humanitarian aid to people in major risk areas. Days later, a number of official and collaborative shelters were set up in churches, parks, schools, and other places to offer help to the victims.
Figure 6 shows a sample of tweets sent over a 3-day observation window.
To compare the proposed sensor with the recent state of the art, a survey was taken of recent works that aimed to address natural disaster monitoring using OSN data with open and available datasets.
Table 6 summarizes the works selected to be compared.
In [
74], the authors assess the impact of a natural hazard and evaluate different topics:
Caution and advice,
Displaced people and evacuations,
Donation needs or offers,
Infrastructure and utilities damage,
Injured or dead people,
Missing,
trapped or found people,
Sympathy emotional support,
Other useful information, and
Not related or irrelevant. Then, for each topic, they process tweets by removing noisy patterns, followed by tagging out-of-vocabulary words and normalizing them. Further, they weigh terms using Word2vec and use them for training three classifiers: NB, SVM, and RF. In [
76], the authors employ Word Embeddings of a fixed sized and a
simple linear kernel SVM to classify tweets into one of three topics:
Damage,
No damage, and
Not relevant. To compare these two works with our methodology, datasets provided by [
43,
74,
76] were annotated with the entity classes described in
Section 3.1 using Polyglot [
78], an NER tagger for multi-lingual purposes, along with other handcrafted rules. Thereafter, to evaluate classification performance, the tagged datasets and
X, the corpus of tweets used in this work (Mexico City Earthquake), were trained with the pool of algorithms used in [
74] and [
43,
76], as well as with the algorithm (biLSTM-CRF) used in our methodology. For each algorithm, it was assumed that words were preprocessed and transformed into word embedding representations. For biLSTM-CRF, only tags describing a
toponym (
and
) are considered. The results are listed in
Table 7 in terms of the
precision,
recall, and
F-1 score.
As observed in
Table 7, the biLSTM-CRF classifier used to build the proposed sensor performs better on average compared with the the RF, SVM, and NB algorithms. The biLSTM-CRF classifier achieves, on average, a precision =
, a recall =
, and an F1-score =
. Even though word embeddings are used in all approaches, only the biLSTM-CRF classifier can capture the maximum contextual information in both directions of a word embedding and its transitions between NER tags at the state level (sentence), thus improving performance results.
Visualizing the Social Dynamics via KDE
Figure 7a–c depict the hotspots obtained by KDE estimations from the geocoded
toponyms over a span of three days. These hotspots allow visualizing areas with the highest concentration of tweets reporting a specific topic and naming a
toponym; i.e.,
. To validate these results, these hotmaps are compared with two collaborative maps populated with official data verified by Google and the Mexican government (publicly available as
Mapeo Verificado19s [
79]). The information contained in Mapeo Verificado19s’ maps is divided into the following categories:
Official Damages: includes collapsed buildings, major and minor risks, and wall collapses.
Official Shelters: official government assistance and aid.
Collaborative Shelters: non-official collaborative assistance and aid.
In addition, sources of information that contributed in a collaborative way to the population of maps during and after the earthquake are listed below:
Mexico City’s Monitor System: includes major risks, collapsed buildings, and gas hazards.
Harvard-Massachusetts Institute of Technology (MIT): collaborative data gathered from social media sources.
It is important to emphasize that official and collaborative maps, e.g.,
Mapeo Verificado19s, neither allow for determining the spatial density of the topic of interest nor account for missing persons. This can be a crucial disadvantage in cases where it is necessary to examine the dynamics and evolution of an event of interest on the basis of incoming reports. The authenticity of
toponyms is tested by searching official addresses published by Mexico’s federal government [
80]. This information can be collected only after civil protection units verify the geographical areas of the disaster and issue an official statement. Unfortunately, there were no oficial data for this natural disaster related to aid, shelter, and missing persons. The proposed sensor has then the potential to assist in estimating in real-time the geographical regions with the largest density of tweets associated with a specific topic of interest, enabling information to be disseminated without subjecting responders to the risks associated with on-site verification.
Table 8 lists the most common geocoded
toponyms transformed into Google API addresses found by our sensor. These locations have also been officially declared as disaster areas by Mexico’s federal government.
6. Conclusions
In this work, a methodology that uses Twitter as a social sensor is proposed. This is accomplished by employing an information sequential extraction procedure known as Named Entity Recognition (NER), which aims to describe mentioned entities, such as places, persons, and organizations. The methodology considers the semantic, morphological, and contextual information about each word composing a tweet and its surrounding context, thus allowing to properly identify a named place (toponym). To achieve this, words are tokenized and transformed into word embeddings to represent them as vectors with rich syntactic and semantic relationships that are established by neighboring words. To ensure that a high classification accuracy of the sequential data is achieved with out heavily relying on handcrafted feature extraction techniques, a Recurrent Neural Network variant, i.e., a Bidirectional Long Short-Term Memory (biLSTM) network, is used. Specifically, the biLSTM network deals with long-distance dependencies, which feed-forward algorithms, such as NB, SVM, and RF, cannot handle. This is achieved by considering contextual information in both directions of a word in a tweet. By using a CRF output layer with the biLSTM network, NER tag transitions over the word embeddings are accounted for.
In the presented case study, geo-queries related to the earthquake of 19 September 2017 in Mexico City were used to retrieve tweets with specific keyword-based hashtags. After classifying Tweets with NER tags and joining them to form useful toponyms, these toponyms were geocoded in terms of addresses and latitude and longitude coordinates by means of Google’s API. Finally, a KDE algorithm was computed to visualize the spatial density of geocoded toponyms from topics related to disaster areas, missing individuals, and shelters. Our results show that addresses and coordinates obtained by our methodology coincide with the ones reported by civil protection units and with official data from Mexico’s federal government. Collaborating with the government and civil organizations to improve the timely detection of disaster areas, finding missing individuals, and locating shelters in real-time by using our proposed methodology is part of our future work.