<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Detecting the Need for Resources and their Availability</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Nelleke Oostdijk*</string-name>
          <email>n.oostdijk@let.ru.nl</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ali Hürriyetoǧlu</string-name>
          <email>a.hurriyetoglu@cbs.nl</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>CLS, Radboud University</institution>
          ,
          <addr-line>Nijmegen and Statistics, Netherlands, CBS-Weg 11, Heerlen</addr-line>
          ,
          <country country="NL">The Netherlands</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Centre for Language Studies (CLS), Radboud University</institution>
          ,
          <addr-line>Erasmusplein 1, Nijmegen 6525 HT</addr-line>
          ,
          <country country="NL">The Netherlands</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>In this working note we describe our submission to the FIRE2017 IRMiDiS track. We participated in both sub-tasks, the first of which was directed at identifying the need or availability of specific resources and the second at matching tweets expressing the need for a resource with tweets mentioning their availability. Our linguistically motivated approach using patern matching of word n-grams achieved an overall average MAP score of 0.2458 for sub-task (1), outperforming our machine-learning approach (MAP 0.1739) while being surpassed by two other (automatic) systems. The linguistic approach was also used in sub-task (2). There it was the bestperforming approach with an f-score of 0.3793.</p>
      </abstract>
      <kwd-group>
        <kwd>Language tag</kwd>
        <kwd>English</kwd>
        <kwd>Hindi</kwd>
        <kwd>Nepali</kwd>
        <kwd>Othera</kwd>
        <kwd>All</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>CCS CONCEPTS
•Computing methodologies →Information extraction;
Natural language processing; Support vector machines; •Information
systems →Content ranking; Social tagging; Social networks;
Content analysis and feature selection; Information extraction; Expert
search; Clustering and classification; Web and social media search;</p>
      <p>
        INTRODUCTION
hTe FIRE2017 IRMiDiS track[ 1] was directed at microblogs and the
aim was to identify actionable information such as what resources
are needed or available during a disaster. In this track there were
two sub-tasks: (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ) identifying need tweets and availability tweets
and (
        <xref ref-type="bibr" rid="ref2">2</xref>
        ) matching need tweets and availability tweets. The dataset
consisted of tweets posted during the Nepal 2015 earthquake.
      </p>
      <p>
        Our submission consisted of the results of two runs each using
a diferent semi-automatic approach addressing sub-task (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ), and
one run with one of our approaches addressing sub-task (
        <xref ref-type="bibr" rid="ref2">2</xref>
        ).
      </p>
      <p>
        hTe structure of the text below is as follows: The data are
described in more detail in Section 2. In Section 3 we describe the
approaches used to address subtask (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ) and the results obtained.
In Section 4 the approach used for subtask 2 is described as well as
the results obtained. Section 5 summarizes the main findings and
includes suggestions for future work.
      </p>
    </sec>
    <sec id="sec-2">
      <title>DATA</title>
      <p>hTe data for this track were tweets posted during the 2015 Nepal
Earthquake and include tweets in English as well as tweets in local
languages such as Hindi and Nepali and code-mixed tweets. The
data were provided by the organizers in the form of tweet IDs,
together with a script for downloading them.
*hTis is the corresponding author</p>
      <p>Tweet %
79.72
6.84
9.34
4.10
100.00
aOther is used to refer to a wide range of languages/language tags as well as the</p>
      <p>UND tag which occurred with 976 tweets.</p>
      <p>Participants were given two datasets which they could use to
develop their methods. The first set (development data) included
roughly 18,000 tweets without any class labels. The second set
(train data) contained some 800 tweets with class labels identifying
them as need tweet (211) or availability tweet (718). The test data
contained 46,920 tweets of which we managed to download 44,759
tweets. The composition of the dataset we downloaded is shown
in Table 1.
3</p>
    </sec>
    <sec id="sec-3">
      <title>SUB-TASK (1): IDENTIFYING NEED AND</title>
    </sec>
    <sec id="sec-4">
      <title>AVAILABILITY</title>
      <p>
        For sub-task (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ) participants were expected to develop
methodologies for identifying need tweets and availability tweets. The
following descriptions were provided: 1
      </p>
      <sec id="sec-4-1">
        <title>Need-tweets: Tweets which inform about the need</title>
        <p>or requirement of some specific resource such as
food, water, medical aid, shelter, mobile or
internet connectivity, etc.</p>
        <p>Availability-tweets: Tweets which inform about
the availability of some specific resources. This
class includes both tweets which inform about
potential availability such as resources being
transported or dispatched to the disaster-struck area,
as well as tweets informing about the actual
availability in the disaster-struck area, such as food
being distributed, etc.
3.1</p>
        <p>hTe Linguistic Approach
hTe approach we used here was based on that applied on a previous
occasion.[3] In this approach a (monolingual English) lexicon and
a set of hand-crafted rules are used to tag the relevant n-grams.
hTe tagged output is then used for assigning class labels. As for
the current track approaches were expected to be able to handle
also non-English data, a pre-processing step was introduced for
dealing with these data. In order to produce the results in a ranked
fashion, where this is not part of the approach as such, we used
some heuristics. More details on each of the steps outlined above
are given below.</p>
        <p>Pre-processing. As a pre-processing step we translated part of
the non-English tweets to English by means of Google Translate
(htps://translate.google.com), that is, only tweets that carried the
(Twiter) language tag HI or NE were translated. All other
nonEnglish tweets (again, according to the language tag provided) were
left untouched, as were all tweets that the language tag identified
as English where in fact they were either mixed tweets (English
and some other language, possibly Hindi or Nepali) or a completely
diferent language.</p>
        <p>An example of a tweet where Twiter identified the language
as Hindi is given below (Example 1). Its translation as produced
by Google Translate is shown in Example 1a. As neither of the
authors has any knowledge of Hindi or Nepali, we had no way of
knowing to what extent the language tag and the translation were
reliable and we simply had to take the translation for what it is
worth. However, based on what we saw with some other tweets in
languages that we do know, we could tell that the language tag was
sometimes completely of. Yet, we decided that we were not going
to spend any time on this issue within the scope of the present
task.</p>
        <p>[Example 1]
[Example 1a]</p>
        <p>Tagging relevant n-grams. A lexicon was constructed
containing some 1,400 items. The lexicon contains mostly unigrams and
bigrams that are considered relevant for the task at hand. All
lexical entries are typed. Main types are V(erb)-like or N(oun)-like.
V-like items typically are verb forms (e.g. distribute) and
nominalizations (e.g. distribution) which express some action. With each
V-like item a tag is associated which indicates whether the items
expresses need or availability (tags act-n and act-a respectively).
N-like items are typically nouns (e.g. food). With each N-like item
a tag is associated which indicates whether the item is a specific
(e.g. water) or more general resource (e.g. aid). Tags here are
res-a and res. In addition to the lexicon there is a small rule set
(currently comprising 10 rules) which specifies how lexical items
may combine to form multiword n-grams. The lexicon and the
rules are used by a software module we have developed in another
project. The module assigns the appropriate tags to the relevant
word n-grams.</p>
        <p>Example 2 shows the result of the tagging of the (translated)
tweet shown in Example 1a.</p>
        <p>[Example 2]
food [res-a] water [res-a] medicine [res-a] lack of
[act-n] electricity [res-a]</p>
        <p>Assigning class labels. The labeling module we developed
automatically assigns a class label (’Nepal-Need’ or ’Nepal-Avail’) to
each tweet. Input for this module is the output of the tagging
described above. For each tweet, the tags assigned are uniqued,
that is a single instance of each tag type is maintained. Example
3 shows the result of this step for our example tweet. The module
makes use of the label patern list which specifies which labels are
to be associated with the (combinations of) tags that can occur in a
tweet (the tag paterns). For our example this means that given the
tag patern [act-n][res-a] the label ’Nepal-Need’ is assigned
(Example 3a). A typical patern for a tweet expressing the availability of
a resource is [act-a] [res-a].</p>
        <p>[Example 3]
[Example 3a]
[act-n] [res-a]
[act-n] [res-a] Nepal-Need</p>
        <p>Ranking the output. As this method does not yield any
confidence or likelihood scores, a ranking was obtained in the following
manner. The output was first ranked based on a human-estimated
confidence of specific class label + tag patern combinations. This
resulted in an initial ranking of the sets of tweets that showed a
particular tag patern. The final ranking was obtained by ordering
the tweets within these ranked sets according to their tweet ID.</p>
        <p>Results. The results as evaluated by the organizers are shown
in Tables 2 and 3 as run ID Radboud_CLS_task1_1. The results
were the best of all submissions using a semi-automatic approach
on all counts (Precision@100, Recall@100, and MAP, both for the
need tweets and the availability tweets). The average MAP of the
Availability and Necessity tweets are 0.2458 and 0.1736 for
Radboud_CLS_task1_1 and Radboud_CLS_task1_2 respectively.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>3.2 hTe Relevancer Approach</title>
      <p>
        For sub-task (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ) we also applied the Relevancer approach[2].
Relevancer was used to generate 194 clusters for the tweets tagged
as English or Hindi. English clusters, which are one third of the
clusters, were annotated and used as training data for the
support vector machines (SVM) based classifier. The cluster
annotation yielded 272 availability and 38 need tweets. The training data
was extended with additional data from [3], the gold annotations
released by the organization team, and the development data
released in the scope of this shared task. The final classifier was
used to predict label of the test tweets. The classifier confidence
was used to rank the results.
      </p>
      <p>Results. The results as evaluated by the organizers are shown
in Table 2 as run ID Radboud_CLS_task1_2.
4</p>
    </sec>
    <sec id="sec-6">
      <title>SUB-TASK (2): MATCHING NEED AND</title>
    </sec>
    <sec id="sec-7">
      <title>AVAILABILITY</title>
      <p>
        In sub-task (
        <xref ref-type="bibr" rid="ref2">2</xref>
        ) participants were required to develop
methodologies for matching need tweets with appropriate availability tweets.
For this task we used the tagging output we had obtained in the
processing the tweets for sub-task 1 using the linguistic approach.
For every need tweet all the word n-grams that had been tagged as
identifying a resource, we would atempt to find an exact match in
the availability tweets. In both cases (need and availability tweets),
the ranked list was used and the software program would work its
way down. Since the task was to find up to 5 availability tweets
with each need tweet and the algorithm would always start at the
top of the list of ranked availability tweets, only a small portion
of the availability tweets actually appears in the matching results.
Only when no exact match could be found, would the software
attempt to find near-matches. This was typically the case for tweets
where at best tweets could be found that yielded a partial match
(e.g. matching 2 out of 4 requested resources).
      </p>
      <p>To illustrate the approach described above, let us get back to the
example we have been using throughout this paper. Our example
tweet was identified as a need tweet. From the tagging we obtained
(shown in Example 2) we would only keep those word n-grams and
their tags that identified a resource. The result of this is shown in
Example 2b.
food [res-a] water [res-a] medicine [res-a]
electricity [res-a]</p>
      <p>In order to find matching availability tweets we would look for
tweets where the same word n-grams could be found, regardless of
the order in which they occurred. The five matching tweets found
amongst the highest ranking availability tweets in the case of our
example were those shown in Examples 4-8.2</p>
      <p>food [res-a] water [res-a]
food [res-a] water [res-a]</p>
      <sec id="sec-7-1">
        <title>RT @abpnewshindi: Food, water and blankets are</title>
        <p>sent to Nepal in the aircraft. s. See Jaishankar
#NepalEarthquake Live- htp://t.co/MG3hLqR5bO
food [res-a] water [res-a] blankets [res-a]</p>
      </sec>
      <sec id="sec-7-2">
        <title>Under the service of religion, Christian mission</title>
        <p>aries are distributing the Bible in food, water, and
clothing in Nepal. Htp://t.co/4E6IHcEqM4 via
@thelapine</p>
        <p>clothing [res-a] food [res-a] water [res-a]
Results.</p>
        <p>hTe results were submited under run ID Radboud_CLS_task2_1
and evaluated by the organizers as follows: precision@5 0.3305,
recall 0.4450 and f-score 0.3793. Thus this approach was found to
outperform all other approaches.
5</p>
      </sec>
    </sec>
    <sec id="sec-8">
      <title>DISCUSSION AND CONCLUSIONS</title>
      <p>
        Our participation in the IRMiDiS track was rather successful: we
achieved a third and sixth place in the overall (MAP) ranking for
3hTe language tag for this tweet was ’und’, therefore no preprocessing was applied.
4hTe language tag here was for Hindi. Example 7a shows the English translation
obtained from Google Translate.
5hTe language tag here was for Hindi. Example 8a shows the English translation
obtained from Google Translate.
sub-task (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ), and scored best of all approaches on sub-task (
        <xref ref-type="bibr" rid="ref2">2</xref>
        ).6 Our
success can in part be atributed to the experience gained through
our participation in previous shared tasks.
      </p>
      <p>However, there are some issues that we struggled with and where
we expect we might do beter on a future occasion. Thus we found
that for the current task the definitions of what constitutes a need
tweet and what an availability tweet were somewhat unclear, more
particularly in specifying what exactly was meant by ’a specific
resource’. The examples in the task description were all clear-cut
cases, including food, water, medicine, electricity, and blood donors.
But what about donations or support which is what we also
encounter in the data.</p>
      <p>
        Looking at the results obtained through our linguistic approach
for sub-task (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ), we note that the distribution of the need and
availability tweets over the various languages is rather uneven. While
on average 7.73% of the data is classified as availability tweet and
3.07% as need tweet, the ratios especially for other language tweets
are much lower (Table 4). We speculate that the fact that the data
comprised multiple languages has afected our recall obtained for
sub-task (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ). While the use of Google Translate for tweets tagged
as Hindi or Nepali was reasonably efective, we expect that beter
results could be achieved if we put more efort into the
preprocessing of the tweets. This would involve both improving the language
identification and finding a way to handle code-mixed tweets.
      </p>
      <p>
        hTe results obtained with the linguistic approach for sub-task (
        <xref ref-type="bibr" rid="ref1">1</xref>
        )
might have been beter if we had allowed for multiple class labels
to be associated with a given tweet. However, as we expected
subtask (
        <xref ref-type="bibr" rid="ref2">2</xref>
        ) to be easier if a tweet carried only a single label, we opted
for a forced choice for one of the two classes and we ignored tweets
where we could not decide for either class.
      </p>
      <p>
        In matching need and availability of resources for sub-task (
        <xref ref-type="bibr" rid="ref2">2</xref>
        )
we restricted ourselves to only exact (literal) matches which means
that we fail to match instances such as the need for shelter and the
availability of tents, or the need for food and the availability of
packed meals. In future work we might include synonyms,
hypernyms, and hyponyms.
      </p>
      <p>hTe Relevancer approach sufered from the fact that the training
set of labelled data was rather small (929 tweets) and quite
unbalanced (211 need vs 718 availability tweets). However, we think that
a combination of the linguistic approach and Relevancer approach
has the potential to overcome such limitations. We are currently
conducting experiments in which we aim to combine the strengths
of the two approaches. So stay tuned!
6hTe organizers distinguish between automatic and semi-automatic approaches
without specifying how to discriminate between them. We consider our approaches
semiautomatic but actually in their execution they operate automatically.</p>
    </sec>
    <sec id="sec-9">
      <title>ACKNOWLEDGMENTS</title>
      <p>
        hTe authors are grateful to Peter Beinema for his help with the
software used in sub-task (
        <xref ref-type="bibr" rid="ref2">2</xref>
        ).
      </p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>Moumita</given-names>
            <surname>Basu</surname>
          </string-name>
          , Saptarshi Ghosh, Kripabandhu Ghosh, and
          <string-name>
            <given-names>Monojit</given-names>
            <surname>Choudhury</surname>
          </string-name>
          .
          <year>2017</year>
          .
          <article-title>Overview of the FIRE 2017 track: Information Retrieval from Microblogs during Disasters (IRMiDis)</article-title>
          .
          <source>In Working notes of FIRE</source>
          <year>2017</year>
          <article-title>- Forum for Information Retrieval Evaluation (CEUR Workshop Proceedings</article-title>
          ).
          <source>CEUR-WS.org.</source>
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>Ali</given-names>
            <surname>Hürriyetoǧlu</surname>
          </string-name>
          , Nelleke Oostdijk, Mustafa Erkan Başar, and Antal van den Bosch.
          <year>2017</year>
          . Supporting Experts to Handle
          <source>Tweet Collections About Significant Events</source>
          . Springer International Publishing, Cham,
          <fpage>138</fpage>
          -
          <lpage>141</lpage>
          . https://doi.org/10. 1007/978-3-
          <fpage>319</fpage>
          -59569-6_
          <fpage>14</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>Ali</given-names>
            <surname>Hürriyetoğlu</surname>
          </string-name>
          and
          <string-name>
            <given-names>Nelleke</given-names>
            <surname>Oostdijk</surname>
          </string-name>
          .
          <year>2017</year>
          .
          <article-title>Extracting Humanitarian Information from Tweets</article-title>
          .
          <source>In Proceedings of the First International Workshop on Exploitation of Social Media for Emergency Relief and Preparedness</source>
          . Aberdeen, United Kingdom. http://ceur-ws.
          <source>org/</source>
          Vol-1832/SMERP-2017
          <string-name>
            <surname>-DC-</surname>
          </string-name>
          RU-Retrieval.pdf
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>