<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Statistical Semantic Classi cation of Crisis Information</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Prashant Khare</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Miriam Fernandez</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Harith Alani</string-name>
          <email>h.alanig@open.ac.uk</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Knowledge Media Institute, Open University</institution>
          ,
          <country country="UK">UK</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>The rise of social media as an information channel during crisis has become key to community response. However, existing crisis awareness applications, often struggle to identify relevant information among the high volume of data that is generated over social platforms. A wide range of statistical features and machine learning methods have been researched in recent years to automatically classify this information. In this paper we aim to complement previous studies by exploring the use of semantics as additional features to identify relevant crisis information. Our assumption is that entities and concepts tend to have a more consistent correlation with relevant and irrelevant information, and therefore can enhance the discrimination power of classi ers. Our results, so far, show that some classi cation improvements can be obtained when using semantic features, reaching +2.51% when the classi er is applied to a new crisis event (i.e., not in training set).</p>
      </abstract>
      <kwd-group>
        <kwd>semantics</kwd>
        <kwd>crisis informatics</kwd>
        <kwd>tweet classi cation</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>As per the 2016 World Humanitarian Data and Trends report by UNOCHA,1
there were around 102 million people, from 114 countries, a ected by natural
disasters in the year of 2015 alone, causing an estimated damage of $90 billion.
During such disasters there is normally a surge of real time content across
multiple social media platforms. For example, during the 2011 Japan earthquake,
there was a 500% increase in the number of tweets.2 All these messages
constitute a critical source of information for relief and search teams, communities
and individuals.</p>
      <p>
        However, it is almost impossible to manually absorb and process the sheer
volume of social media reports generated during a crisis, and to e ciently
lter any relevant and actionable information [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. Tools to automatically identify
relevant information are largely unavailable, and the characteristics of social
media messages (short length, use of colloquialisms, ill-formed words and syntactic
structures) increases the challenges of automatically processing and
understanding of such messages.
1 https://data.humdata.org/dataset/world-humanitarian-data-and-trends
      </p>
    </sec>
    <sec id="sec-2">
      <title>2 https://blog.twitter.com/official/en_us/a/2011/global-pulse.html</title>
      <p>
        Much research explored methods for the classi cation of social media data
into crisis-related or unrelated, based on supervised [
        <xref ref-type="bibr" rid="ref10 ref16 ref20 ref8">10,8,16,20</xref>
        ] and
unsupervised [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] machine learning (ML) methods. These methods tend to identify
relevant data based on n-grams and statistical features (message length, URLs,
Hashtags, etc.). This paper aims to complement previous works by investigating
the impact of semantic features to identify relevant information from Twitter
data during crisis situations. The semantic features explored in our work include
entities (e.g., \London", \Colorado", \Fire") extracted from tweets, as well as
their hypernyms from BabelNet, which is an external knowledge base[
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]. Our
hypothesis is that entities and concepts may have a more consistent correlation
with relevant and irrelevant crisis information, and therefore can be used to
better interpret the content of the tweets and to enhance the discrimination power
of classi ers.
      </p>
      <p>We explore the e ectiveness of semantic features by creating and testing
classi ers to identify relevant crisis information, as well as by testing these classi ers
with previously unseen information from di erent crisis events. The dataset used
in our research is a small subset of CrisisLexT26;3 a library of 205K annotated
tweets posted during 26 real crisis events in 2012 and 2013. Our subset consists
of a balanced related-unrelated set of 3.2K tweets on 9 crisis events (detailed in
Section 3.1). Our results show that using semantic information can indeed help
to enhance classi cation results, but only by a small margin. When the classi er
is applied to a new crisis event, results show that the use of semantic
annotations of concepts and entities in itself is e ective, and the use of semantically
expanded concepts (i.e., entities and their hypernyms) further improves over it
slightly. However, the use of hypernyms also sometimes introduces generic
concepts, such as \person", that appear in both, crisis related and non-crisis related
posts, and thus e ects the discrimination power of semantic features.</p>
      <p>The contributions of this work can be summarised as follows:
{ Demonstrating the impact of using a variety of semantic features for
identifying crisis-related information from social media posts.
{ Showing that adding semantic features is especially useful when classifying
new crisis events that were not seen during the model training phase.
{ Testing using annotated data from CrisisLexT26 of 9 real crisis events.
{ Discussing and re ecting on the potential use of semantics to identify
crisisrelevant information.</p>
      <p>The rest of the paper is structured as follows. Section 2 summarises the
related work on processing social media data for identifying crisis related content.
Section 3 describes our approach, including the selected semantic features and
how they are used to created various types of classi ers. The experiments are
results are reported in Section 4. Section 5 discusses the lessons learned from
this work, as well as its limitations and the future lines of work. Conclusions are
reported in Section 6.</p>
    </sec>
    <sec id="sec-3">
      <title>3 crisislex.org</title>
      <sec id="sec-3-1">
        <title>Related Work</title>
        <p>
          During a crisis, a very large number of messages are often posted on various
social media platforms. Processing all such messages requires substantial time and
e ort to ensure that crisis related messages are e ciently spotted and handled,
since a good percentage of messages posted about a crisis tends to be
irrelevant and unrelated. Olteanu and colleagues observed that crisis reports could be
classi ed into three main categories: related and informative, related but not
informative, and not related [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ]. In this work, we focus primarily on the automatic
identi cation of crisis related information. The identi cation of informativeness
in crisis scenarios is a complex task that requires a deeper re ection and
investigation of the meaning of informativeness and its dimensions (freshness, novelty,
location, scope). It is therefore an important part of our future work.
        </p>
        <p>
          To identify crisis related messages from social media data, several works have
proposed the use of supervised [
          <xref ref-type="bibr" rid="ref10 ref16 ref20 ref8">10,8,16,20</xref>
          ] and unsupervised [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ] ML classi
cation methods. Supervised methods tend to make use of n-grams as well as of
linguistic and statistical features such as part of speech (POS), number of
hashtags, mentions, or message length. They also highlight the use of location as
an important indicator, since people tend to create and retweet messages with
locally actionable information [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ]. These works make use of various supervised
classi cation methods, from traditional classi cation algorithms such as Naive
Bayes, Support Vector Machines or Conditional Random Field [
          <xref ref-type="bibr" rid="ref13 ref16 ref6">13,16,6</xref>
          ] to more
novel techniques such as deep learning [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ]. Unsupervised methods, on the other
hand, are mainly based on keyword processing and clustering [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ]. Our work
aims to complement these studies by investigating the use of semantics, and
particularly the use of entities extracted from tweets, and their hypernyms, as
additional features to boost classi cation. As previously done by [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] we not only
aim to generate classi ers able to identify crisis-related information, but we also
aim to test the generated classi ers on crisis events that the classi ers have not
previously seen.
        </p>
        <p>
          While semantic models have been developed and used to represent and
capture the information that emerge from crisis events (e.g., MOAC - Management
of a Crisis 4, or HXL - Humanitarian eXchange Language5), few works in the
literature have explored the use of semantics to identify and lter crisis-related
information. In [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ], Abel and colleagues presented Twitcident, a system that uses
semantic information to facilitate ltering and search of crisis related
information. The system extracts semantic information from social media data in the
form of entities using Name Entity Recognisers (NER) and external knowledge
bases. However, as opposed to our work, they do not explore the use of entities
as features for classi cation. Instead, they develop similarity models in which
the crisis event and the posts are pro led based on this semantic enrichment,
and the Jaccard similarity coe cient6 is used to compute whether the content
of the posts is similar or not to the event.
        </p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4 http://www.observedchange.com/moac/ns.</title>
    </sec>
    <sec id="sec-5">
      <title>5 http://hxlstandard.org/</title>
    </sec>
    <sec id="sec-6">
      <title>6 https://en.wikipedia.org/wiki/Jaccard_index</title>
      <p>Classi cation: Identifying Crisis Related Information
Our approach for identifying crisis related information among tweets functions
by generating binary classi ers to di erentiate crisis-related from non-related
posts. In this section, we explain (i) the dataset used in our experiments, (ii) the
two set of features (statistical and semantic) that we use to build the classi ers,
and (iii) our classi er selection process.
3.1</p>
      <sec id="sec-6-1">
        <title>Data Selection</title>
        <p>
          To conduct our study we selected the CrisisLexT267 dataset [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ], an annotated
dataset of 205K tweets posted during 26 crisis events occurring between 2012 and
2013. The search keywords used to construct CrisesLexT26 were selected
following the standard practices of hashtags and/or terms often paired with canonical
forms of a disaster name and impacted location (e.g., Queensland oods) or a
meteorological term (e.g., Hurricane Sandy). For each of the 26 crisis events,
around 1,000 tweets are annotated (Related and Informative, Related but not
Informative, Not Related, or Not Applicable). Given our focus on English tweets,
we selected 9 events for which the content was predominantly in English: West
Texas Explosion(WTE), Colorado WildFire(CWF), Colorado Flood (CFL),
Australia Bush re(ABF), Boston Bombing (BB), LA Shooting (LAS), Queensland
Flood (QFL), Savar Building Collapse (SBC), and Singapore Haze(SGH).
        </p>
        <p>We merged those tweets labelled as Not Related and Not Applicable under
the class Not Related, obtaining a total of 1539 non crisis-related tweets. We also
merged those tweets labelled as Related and Informative and Related but not
Informative under the class Related, obtaining a total of 7461 crisis related tweets.
In line with common practice, we balanced the dataset to remove classi cation
bias towards the bigger class Related, by randomly selected 1667 crisis related
tweets. This gives us a balanced and annotated dataset of 3206 of Related and
Not Related tweets.
3.2</p>
      </sec>
      <sec id="sec-6-2">
        <title>Feature Engineering</title>
        <p>
          To generate classi ers able to identify crisis-related posts, we explore two
distinct feature sets, statistical and semantic features. Statistical features have been
widely studied in the literature [
          <xref ref-type="bibr" rid="ref10 ref16 ref20 ref8">10,8,16,20</xref>
          ] and are used as the baseline for our
experiments. They capture the linguistic and quanti able attributes of posts.
Semantic features, on the other hand, capture the di erent named entities that
emerge from tweets, as well as their hierarchical information which we extract
from an external knowledge source.
3.2.1 Statistical Features (SA) For each social media post, we extract the
following statistical features:
{ Number of nouns: nouns generally refer to locations, resources, or actors
involved in the crisis event.
7 http://crisislex.org/data-collections.html#CrisisLexT26
{ Number of verbs: verbs are an indication of the di erent actions that are
occurring during the crisis event.
{ Number of pronouns: as with nouns, pronouns may be an indication of the
actors, locations, or resources that are named during the crisis event.
{ Tweet Length: number of characters contained in the posts. The longer the
post is, the higher the amount of information it may contain.
{ Number of words: number of words may be another indication of the amount
of information the post may have.
{ Number of Hashtags: hashtags indicate the themes of the post and are
manually generated by the posts' authors.
{ Readability: Gunning fog index using average sentence length (ASL) and
the percentage of complex words (PCW): 0.4 * (ASL + PWC). This feature
gauges how hard the post is to parse by humans.8
{ Unigrams: unigrams provide a keyword-based representation of the content
of the posts
        </p>
        <p>To extract the unigrams from social media posts we make use of the Weka
data mining software9, and speci cally its StringToWord functionality,
including lower case conversion for all tokens, stemming (using Lovins' algorithm)10,
stopword removal, and tf*idf transformation. The total number of unigrams, or
vocabulary size, for the complete dataset is 10655. To extract the Part of Speech
(POS) tags and the statistical features listed above (top ve), we make use of
the Stanford Core NLP software.11 Hashtags are identi ed by the use of the #
character, and readability is computed using the Gunning fog index.
3.2.2 Semantic Features (SemF) The semantic feature extraction process
is summarised in Figure 3.2.2 and consists of three main steps: (i) semantic
annotation, (ii) semantic expansion, and (iii) semantic ltering. Each of these
three steps generates a di erent set of semantic features that we explore,
individually and in combination, when generating binary classi ers to distinguish
crisis-related posts from unrelated ones.</p>
      </sec>
      <sec id="sec-6-3">
        <title>Semantic Annotation Features (SemAF): In the initial step (semantic</title>
        <p>
          annotation) semantic entities are extracted from the posts by using Babelfy.12
This Name Entity Recogniser (NER) identi es the di erent entities that appear
in the text, disambiguates them, and links them to the BabelNet[
          <xref ref-type="bibr" rid="ref11">11</xref>
          ] knowledge
base, providing a unique identi er (SynsetID) for each of the identi ed entities.
For example (Figure 1), for the post \A 15-year-old High River boy is missing
due to the ood. Call police if you see Eric St. Denis #ab ood " Babel y identi es
entities such as High River, Boy, Flood, etc. The annotation of the entire dataset
(see Section 3.1) resulted in 12,006 unique concepts.
        </p>
      </sec>
    </sec>
    <sec id="sec-7">
      <title>8 https://en.wikipedia.org/wiki/Gunningfogindex</title>
    </sec>
    <sec id="sec-8">
      <title>9 http://www.cs.waikato.ac.nz/ml/weka/</title>
      <p>10 http://www.mt-archive.info/MT-1968-Lovins.pdf
11 https://stanfordnlp.github.io/CoreNLP/
12 http://babelfy.org</p>
      <p>Semantic Expansion Features (SemEF): In the second step (semantic
expansion) the BabelNet knowledge base is used to extract every direct
hypernym (distance-1) of these entities. Our hypothesis for considering hypernyms is
that, by introducing upper level concepts, we might be able to better encapsulate
the semantics of crisis-related tweets. For example, if the entities reman and
policeman appear often in crisis related posts. These entities have a common
hypernym, defender. As a result, a post with the entity MP (Military Police),
is more likely to also be crisis-related since this entity also has the hypernym
defender. The semantic expansion process resulted in an additional 7032 unique
concepts.</p>
      <p>
        Semantic Filtering Features (SemFF): When semantically expanding
the initially extracted entities, we could sometimes introduce very generic
concepts with low discrimination power. For example, the hypernym Person appears
in both crisis and non-crisis related posts, and thus does not help the classi ers
to identify crisis-related information. Our ltering process aims to discard such
semantic annotations that might be too generic and hence are likely to reduce
the discrimination power of semantics. Our proposed ltering process is based
on the computation of the depth of a concept in the hierarchy of BabelNet.
To determine the depth of concepts, we query iteratively through the hierarchy
of BabelNet. Abstract concepts, i.e., concepts with a lower depth are therefore
removed. To determine the shortest depth of a concept in the hierarchy of
BabelNet, we used nearly 4 million relations extracted by iteratively querying for
hypernyms and generated a Directed Graph. The node with highest betweenness
centrality (SynSetID `bn:00031027n', which relates to the main sense `Entity')
was determined to be the most abstract concept. The NetworkX13 graph library
for Python was used for this task. We then computed the Shortest path between
the node `Entity' and all the extracted hypernyms. The maximum depth found
was 21, where level 0 is assigned to the concept `Entity'. By performing an
em13 https://networkx.github.io/
pirical analysis of the concepts using Information Gain, we observed that the
most informative concepts are those whose depth is between 3 and 7. Those are
therefore the ones selected as features for classi cation. This ltering process
resulted in 574 concepts ltered out from the semantics across 9 events.
When selecting a classi cation method for the problem at hand we considered
the high dimensionality of features, particularly the high number of unigrams
and semantic features, the limited set of labelled data (3,206 posts) and the
importance of avoiding over- tting. Given the large set of features in comparison
with the number of training examples, we opted for selecting the Support Vector
Machine (SVM) classi cation model [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] with Linear Kernel. SVM has proven
e ective for problems with these characteristics.14
4
      </p>
      <sec id="sec-8-1">
        <title>Experiments</title>
        <p>In this section we describe our experimental set up, and particularly the design of
our model selection and testing experiments. We report on the obtained results
and later discuss how semantic features can help enhancing the performance
of classi ers based on statistical features, and especially when the classi er is
applied to cross-crisis scenarios.
4.1</p>
        <sec id="sec-8-1-1">
          <title>Experimental Setup</title>
          <p>We designed two main experiments where we train and test our classi cation
models on (i) all 9 crisis events, and (ii) on 8 events, and retest on the 9th event,
i.e., cross-crisis testing.
14 http://www.csie.ntu.edu.tw/~cjlin/papers/guide/guide.pdf Radial Basis
Function (rbf) kernel or a Polynomial Kernel may cause an over- tting
problem,hence we opted for a linearly separable hyperplane.
{ Crisis Classi cation Model: In our rst experiment we compare the
performance of classi ers generated with statistical features vs. classi ers enhanced
with semantic features and analyse if the use of semantics does indeed help
boosting the performance of binary classi ers when identifying crisis-related
posts. We compare the performance of four di erent classi ers generated
using the complete dataset (see Section 3.1), and tested using 10-fold cross
validation. We use the WEKA software (v.3.8)15 to generate the classi ers.</p>
          <p>SF: A classi er generated with statistical features; our baseline.</p>
          <p>SF+SemAF: A classi er generated with statistical features, and semantic
annotation features.</p>
          <p>SF+SemAF+SemEF: A classi er generated with statistical features,
semantic annotations, and their hypernyms, i.e., the Semantic Expanded
Features.</p>
          <p>SF+SemFF: A classi er generated with statistical features, and ltered
semantic annotations, along with their hypernyms, i.e., the Semantic
Filtered Features
{ Cross-crisis Classi cation: In our second experiment we retest the classi ers
above by applying them to a new crisis data, i.e., on data from a new crisis
event that was not part of the training set. For this experiment, we generate
the same four classi ers described in the previous task. However, rather than
using the complete dataset to generate the model, we use 8 out of the 9 crisis
events to generate the model, and then apply the models to the remaining
event for validation. We therefore generate 36 di erent classi cation models
for this experiment.
4.2</p>
        </sec>
        <sec id="sec-8-1-2">
          <title>Results: Crisis Classi cation</title>
          <p>The results of our rst experiment (each model statistically evaluated with 10
iterations of 10-fold cross validation) are presented in Table 1. The table presents
F-measure(F) value (from 10-fold cross validation), mean of F-measure (Fmean)
of 100 results from 10 iterations, standard deviation in F-measure ( ),and the
increment of Fmean over the baseline F=F . Precision and Recall values where
equal to F in this experiment, and hence were omitted from table. As we can
see in this table, the use of semantic features helps to enhance classi cation
results in all cases, but almost negligibly (less than 0.6%). However, the use of
annotations alone (SF+SemAF) produces slightly better results than the use of
annotations and hypernyms (SF+SemAF+SemEF).</p>
          <p>To better understand the impact of semantics in this context, we manually
analyse some of the tweets that were misclassi ed by the statistical baseline
model, but were correctly classi ed when using semantics (see Table 2)In
addition, we perform feature selection using Information Gain (IG) over the
generated classi ers to determine which are the most discriminative statistical and
semantic features when identifying crisis related posts.
15 http://www.cs.waikato.ac.nz/ml/weka/</p>
          <p>When applying IG over the attributes of the baseline classi er, the number
of hashtags was the most relevant feature. By manually checking some of the
tweets, we observe that Not Related posts tend to either have no hashtags (see
as example Post2) or contain many hashtags (see Post1). The number of nouns
and pronouns is also a high discriminative feature. As we hypothesized, crisis
related posts generally contain more nouns and pronouns mentioning persons,
resources or locations relevant to the crisis event. When including semantics,
we observe that the hypernyms and annotations are among the highly ranked
features, based on IG. Apart from highly ranked statistical features, hypernyms
such as `Happening ' and `Event ' (which, in BabelNet, are hypernyms of concepts
such as `Incident', `Fire', `Crisis',`Disaster',and `Death'), were among the top 10
attributes (out of almost 800 positive IG attributes).</p>
          <p>Post3 was misclassi ed when using only statistical features. Although it
contains the relevant term burn, it barely appears in the training data.
However, the post is correctly classi ed by SF+SemAF, because the term burn
returns the concept Fire as part of its semantic annotation. Post 4 was
misclassi ed by SF+SemAF, but correctly classi ed when adding semantic expansion
(SF+SemAF+SemEF).The original tweet was annotated with the concept
Wildre, which has the hypernym Fire; a feature with high IG and strongly associated
with the class crisis-related. Therefore, in this case, the use of hypernyms helped
to obtain the additional information needed to correctly categorise the post.</p>
          <p>Post 5 was misclassi ed by SF+SemAF+SemEF but correctly classi ed by
SF+SemFF. Annotations such as Thanks and Meet semantically expanded to
hypernyms such as Virtue, and Desire, which have a very low discrimination
power, and hence weakens the classi er. We observe that removing such less
informative abstract concepts results in increasing the discriminative power of
the remaining, more informative, concepts, such as `Volunteer ' and hypernym
(of `donor') `Benefactor '.
4.3</p>
        </sec>
        <sec id="sec-8-1-3">
          <title>Results: Cross-crisis Classi cation</title>
          <p>The results of this experiment are reported in Table 3. In this experiment, we
compile 9 di erent datasets, where in each dataset 1 out of the 9 crisis events is
entirely left out of the training sample used to train and test the classi cation
model.16 Each row is named after the the crisis event that was left out of the
dataset during its creation (see Section 3.1). The data split for each dataset (train
on 8 event/test on 9th event) is presented in the second column of the table.
For each of these 9 datasets we created the four di erent classi ers described in
Section 4.2. The results of each of these models for the 9 di erent datasets are
reported in table along with their values of Precision (P), Recall (R), F1-measure
(F) and the increment of F measure over the baseline, F=F .</p>
          <p>As we can see, the use of semantics enhances classi cation results in all
cases. We observe that SF+SemAF improves the classi cation over the baseline
SF, in 6 out of 9 case, with an average of 0.9% increase in F-1 measure. As
opposed to our previous experiment (10-fold cross-validation), however, the use
of hypernyms makes the model more adaptable to unknown data in 6 out of 9
cases, with an average improvement of 1.94% over the baseline(SF). Semantic
expansion (SemEF) improves over the annotation model (SemAF) in 5 out of 9
cases. Also, it is worth noting that ltering out the abstract concepts resulted
in an improved performance of SF+SemFF over SF+SemAF+SemEF model (
average of 0.6%), in 7 out of 9 cases. This validates the argument (Sec 3.2.2) that
certain concepts tend to appear in both, crisis related and non-related tweets, and
16 Each model was tested on the 8 event dataset it was trained on using 10 fold
crossvalidation to ensure its accuracy before applying it to the 9th event data. There
accuracy drops around 17% on average when applied to new events.
therefore introduce noise rather than helping with the classi cation. Filtering out
such concepts enhances the classi cation. SF+SemFF model improves over the
baseline by an average of 2.51%.
5</p>
        </sec>
      </sec>
      <sec id="sec-8-2">
        <title>Discussion and Future Work</title>
        <p>Our ndings show potential in mixing statistical and semantic features for
classifying crisis-related and unrelated tweets. The highest, and more worthy,
improvement is achieved when using this hybrid model to classify data of a new
crisis event that the model was not trained on. This is due to the use of semantic
knowledge graphs to expand the vocabulary into semantic concepts and
hypernyms, and thus capturing the essence of the tweets and their terms. However,
we showed that such a semantic expansion could introduce noise in the form of
abstract concepts, which requires ltering to maximise bene t.</p>
        <p>An issue we encountered was the unsymmetrical mappings of
HypernymHyponym relationship in BabelNet, which e ected the hierarchical expansion of
semantics and hierarchy generation. As a future work, we plan to refer to more
symmetrically mapped resources, such as WordNet17, and extend to the types
and categories of semantics through external knowledge base such as DBpedia18.</p>
        <p>One of the limitations of this study is the small size of the dataset (3206
annotations) and type of crisis events (5 di erent types), which we plan to expand
in future work. We also need to investigate whether the discriminative features
di er across the various type of crisis, and languages. Additionally, we will
investigate whether adding semantic features incorrectly classi es some tweets that
are correctly classi ed by the statistical approach.
6</p>
      </sec>
      <sec id="sec-8-3">
        <title>Conclusion</title>
        <p>This work presents an approach to leverage semantic enrichment for classifying
unseen crisis Twitter data. The two approaches of semantic enrichment;
annotation and semantic expansion, exhibit an improvement in classi cation
performance over the statistical features by 0.9%-2.51%. We have also demonstrated
empirically that more abstract concepts are less discriminative, and proposed a
method that lters the concepts which are less likely to be discriminative.
17 https://wordnet.princeton.edu/
18 http://wiki.dbpedia.org/</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Abel</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Celik</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Houben</surname>
            ,
            <given-names>G.J.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Siehndel</surname>
            ,
            <given-names>P. Leveraging</given-names>
          </string-name>
          <article-title>the semantics of tweets for adaptive faceted search on twitter</article-title>
          .
          <source>Int. Semantic Web Conf. (ISWC)</source>
          , Bonn, Germany,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Abel</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hau</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Houben</surname>
            ,
            <given-names>G. J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stronkman</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Tao</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <article-title>Semantics+ ltering+ search= twitcident. exploring information in social web streams</article-title>
          .
          <source>Conf. Hypertext and Social Media (Hypertext)</source>
          ,
          <source>WI., USA</source>
          ,
          <year>2012</year>
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Burel</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Saif</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fernandez</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Alani</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          (
          <year>2017</year>
          ).
          <article-title>On Semantics and Deep Learning for Event Detection in Crisis Situations</article-title>
          . Workshop on Semantic Deep Learning (SemDeep),
          <string-name>
            <surname>at</surname>
            <given-names>ESWC</given-names>
          </string-name>
          , Portoroz, Slovenia,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Cristianini</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Shawe-Taylor</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <article-title>An introduction to support vector machines and other kernel-based learning methods</article-title>
          . Cambridge university press,
          <year>2000</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Gao</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Barbier</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Goolsby</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <article-title>Harnessing the crowdsourcing power of social media for disaster relief</article-title>
          .
          <source>IEEE Intelligent Systems</source>
          ,
          <volume>26</volume>
          (
          <issue>3</issue>
          ),
          <fpage>10</fpage>
          -
          <lpage>14</lpage>
          ,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Imran</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Elbassuoni</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Castillo</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Diaz</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Meier</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <article-title>Practical extraction of disaster-relevant information from social media</article-title>
          .
          <source>Int. World Wide Web Conf. (WWW)</source>
          , Rio de Janeiro, Brazil,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Jadhav</surname>
            ,
            <given-names>A.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Purohit</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kapanipathi</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Anantharam</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ranabahu</surname>
            ,
            <given-names>A.H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nguyen</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mendes</surname>
            ,
            <given-names>P.N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Smith</surname>
            ,
            <given-names>A.G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cooney</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Sheth</surname>
            ,
            <given-names>A.P.</given-names>
          </string-name>
          <article-title>Twitris 2.0: Semantically empowered system for understanding perceptions from social data</article-title>
          , http://knoesis.wright.edu/library/download/Twitris ISWC 2010.pdf,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Karimi</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yin</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          and Paris, C. December.
          <article-title>Classifying microblogs for disasters</article-title>
          .
          <source>Australasian Document Computing Symposium</source>
          , Brisbane,
          <string-name>
            <surname>QLD</surname>
          </string-name>
          , Australia,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Kogan</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Palen</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Anderson</surname>
            ,
            <given-names>K.M.</given-names>
          </string-name>
          <string-name>
            <surname>February</surname>
          </string-name>
          .
          <article-title>Think local, retweet global: Retweeting by the geographically-vulnerable during Hurricane Sandy</article-title>
          . Conf.
          <article-title>on computer supported cooperative work &amp; social computing (</article-title>
          <source>CSCW '15)</source>
          , Vancouver, Canada,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lei</surname>
            ,
            <given-names>K.H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Khadiwala</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Chang</surname>
            ,
            <given-names>K.C.C.</given-names>
          </string-name>
          ,
          <year>2012</year>
          , April. Tedas:
          <article-title>A twitterbased event detection and analysis system</article-title>
          .
          <source>IEEE 28th Int. Conf. on Data Engineering (ICDE)</source>
          , Washington, DC, USA,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Navigli</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Ponzetto</surname>
            ,
            <given-names>S.P.</given-names>
          </string-name>
          <article-title>BabelNet: The automatic construction, evaluation and application of a wide-coverage multilingual semantic network</article-title>
          .
          <source>Arti cial Intelligence</source>
          ,
          <volume>193</volume>
          , pp.
          <fpage>217</fpage>
          -
          <lpage>250</lpage>
          ,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Olteanu</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vieweg</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Castillo</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>February</surname>
          </string-name>
          .
          <article-title>What to expect when the unexpected happens: Social media communications across crises</article-title>
          .
          <source>Conf. on Computer Supported Cooperative Work &amp; Social Computing (CSCW '15)</source>
          , Vancouver, Canada,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Power</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Robinson</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Colton</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Cameron</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <article-title>Emergency situation awareness: Twitter case studies</article-title>
          .
          <source>Int. Conf. on Info. Systems for Crisis Response and Management in Mediterranean Countries (ISCRAM)</source>
          , Toulouse, France,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Rogstadius</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vukovic</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Teixeira</surname>
            ,
            <given-names>C.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kostakos</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Karapanos</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Laredo</surname>
            ,
            <given-names>J.A.</given-names>
          </string-name>
          <article-title>CrisisTracker: Crowdsourced social media curation for disaster awareness</article-title>
          .
          <source>IBM Journal of Research and Development</source>
          ,
          <volume>57</volume>
          (
          <issue>5</issue>
          ), pp.
          <fpage>4</fpage>
          -
          <lpage>1</lpage>
          ,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Sakaki</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Okazaki</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Matsuo</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          <article-title>Earthquake shakes Twitter users: real-time event detection by social sensors</article-title>
          .
          <source>Int. Conf. World Wide Web (WWW)</source>
          , Raleigh, North Carolina USA,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>Stowe</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Paul</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Palmer</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Palen</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Anderson</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Identifying</surname>
            and
            <given-names>Categorizing</given-names>
          </string-name>
          <string-name>
            <surname>Disaster-Related Tweets</surname>
          </string-name>
          .
          <source>Workshop on Natural Language Processing for Social Media</source>
          ,
          <string-name>
            <surname>In</surname>
            <given-names>EMNLP</given-names>
          </string-name>
          , Austin, Texas, USA,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <surname>Vieweg</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hughes</surname>
            ,
            <given-names>A.L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Starbird</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Palen</surname>
            ,
            <given-names>L..</given-names>
          </string-name>
          <article-title>Microblogging during two natural hazards events: what twitter may contribute to situational awareness</article-title>
          .
          <source>Proc. SIGCHI Conf. on Human Factors in Computing Systems (CHI)</source>
          , Atlanta,
          <string-name>
            <surname>GA</surname>
          </string-name>
          , USA,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          18.
          <string-name>
            <surname>Vieweg</surname>
            ,
            <given-names>S.E.</given-names>
          </string-name>
          ,
          <year>2012</year>
          .
          <article-title>Situational awareness in mass emergency: A behavioral and linguistic analysis of microblogged communications (Doctoral dissertation</article-title>
          , University of Colorado at Boulder), https://works.bepress.com/vieweg/15/
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          19.
          <string-name>
            <surname>Yin</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lampert</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cameron</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Robinson</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Power</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <article-title>Using social media to enhance emergency situation awareness</article-title>
          .
          <source>IEEE Intelligent Systems</source>
          ,
          <volume>27</volume>
          (
          <issue>6</issue>
          ), pp.
          <fpage>52</fpage>
          -
          <lpage>59</lpage>
          ,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          20.
          <string-name>
            <surname>Zhang</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Vucetic</surname>
            ,
            <given-names>S..</given-names>
          </string-name>
          <article-title>Semi-supervised Discovery of Informative Tweets During the Emerging Disasters</article-title>
          .
          <source>arXiv preprint arXiv:1610.03750</source>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>