<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Monitoring Social Media to Identify Environmental Crimes through NLP A Preliminary Study</article-title>
      </title-group>
      <contrib-group>
        <aff id="aff0">
          <label>0</label>
          <institution>Raffaele Manna, Antonio Pascucci, Wanda Punzi Zarino, Vincenzo Simoniello, Johanna Monti UNIOR NLP Research Group University “L'Orientale” Naples</institution>
          ,
          <country country="IT">Italy</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>This paper presents the results of research carried out on the UNIOR Eye corpus, a corpus which has been built by downloading tweets related to environmental crimes. The corpus is made up of 228,412 tweets organized into four different subsections, each one concerning a specific environmental crime. For the current study we focused on the subsection of waste crimes, composed of 86,206 tweets which were tagged according to the two labels alert and no alert. The aim is to build a model able to detect which class a tweet belongs to.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1 Introduction</title>
      <p>
        In the current era, social media represent the
most common means of communication,
especially thanks to the speed with which a post can
go viral and reach in no time every corner of the
globe. The speed with which information is
produced creates an abundance of (linguistic) data,
which can be monitored and handled with the use
of hashtags (#). Hashtags are user-generated
labels, which allow other users to track posts with a
specific theme on Twitter. Moreover, social media
such as Twitter can be powerful tools for
identifying a variety of information sources related to
people’s actions, decisions and opinions before,
during and after broad scope events, such as
environmental disasters like earthquakes, typhoons,
volcanic eruptions, floods, droughts, forest fires,
landslides
        <xref ref-type="bibr" rid="ref11 ref3 ref4">(Imran et al., 2015; Maldonado et al.,
2016; Corvey et al., 2010)</xref>
        . In light of the above,
      </p>
      <p>Copyright c 2020 for this paper by its authors. Use
permitted under Creative Commons License Attribution 4.0
International (CC BY 4.0).
our aim is to monitor social media in order to
detect environmental crimes.</p>
      <p>
        Our research is guided by the following question:
can Natural Language Processing (NLP) represent
a valuable ally to identify these kinds of crimes
through the monitoring of social media? For this
purpose, we compiled a corpus of tweets starting
from a list of 41 terms related to environmental
crimes, e.g. combustione illecita (illicit
combustion), rifiuti radioattivi (radioactive waste),
discarica abusiva (illegal dumping), and we used the
Twitter API to download all the tweets
(specifically 228,412) related to these terms introduced
by hashtag. In this research, a special focus is
dedicated to the tweets related to La terra dei
fuochi (literally the Land of Fires)
        <xref ref-type="bibr" rid="ref13">(Peluso, 2015)</xref>
        ,
a large area located between Naples and Caserta
(in the South of Italy) victim of illegal toxic wastes
dumped by organized crime for about fifty years
and routinely burned to make space for new toxic
wastes.
      </p>
      <p>In order to achieve our purpose, we trained
different machine learning algorithms to classify report
emergency text and user-generated reports. The
paper is organized as follows: in Section 2 we
discuss Related Work, in Section 3 we present the
UNIOR Earth your Estate (UNIOR Eye) corpus.
The case study is described in Section 4 and
Results are discussed in Section 5. Conclusions are
in Section 6 along with directions for Future Work.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Related Work</title>
      <p>As previously mentioned, hashtags are one of the
most important resources - if not the most
important - in text data such as those of Twitter. The
possibility to aggregate data according to their content
allows users to monitor all the discussion about a
specific subject in real-time (an emblematic case
is the hashtag #Covid 19).</p>
      <p>Concerning the topic of our research, namely
environmental issues, the most representative and
productive hashtags have proved to be
#terradeifuochi and #rifiuti (respectively with a frequency
of 92,322 and 62,750 occurrences), that directly
refer to circumstances that have a strong impact
on the environment and people’s health. The use
of hashtags proved to be useful in monitoring
natural disasters, such as earthquakes, flood and
hurricane.</p>
      <p>
        For a survey on information processing and
management of social media contents to study
natural disasters, see
        <xref ref-type="bibr" rid="ref5">(Imran et al., 2016)</xref>
        .
        <xref ref-type="bibr" rid="ref12">(Neubig et
al., 2011)</xref>
        focused on the 2011 East Japan
earthquake. The scholars built a system able to
extract the status of people involved in the disaster
(e.g. if they declared to be alive, they request for
help, their information requests, information about
missing people). About one hundred scholars
participated spontaneously in the project ANPI NLP
(ANPI means Safety in Japanese) and the results
show convincing performances by the classifier
they built.
        <xref ref-type="bibr" rid="ref11">(Maldonado et al., 2016)</xref>
        investigated
natural disasters in Ecuador, monitoring Twitter
to filter contents according to four different
categories: volcanic, telluric, fires and climatological.
The filtering process is based on keywords related
to the four categories. The scholars released a web
application that graphically shows the database
evolution. The efficiency of the tweet filtering
algorithm that they developed is expressed in terms
of precision (%93.55).
        <xref ref-type="bibr" rid="ref17">(Tarasconi et al., 2017)</xref>
        investigated tweets related to eight different event
types (floods, wildfires, storms, extreme weather
conditions, earthquakes, landslides, drought and
snow) in Italian, English and Spanish. The
corpus is composed of 9,695 tweets and can be
extremely useful to perform information extraction
in the aforementioned three languages.
        <xref ref-type="bibr" rid="ref15">(Sit et
al., 2019)</xref>
        used the Hurricane Irma, which
devastated Caribbean Islands and Florida in September
2017, as a case-study: the scholars demonstrate
that by monitoring tweets it is possible to detect
potential areas with high density of affected
individuals and infrastructure damage throughout the
temporal progression of the disaster. By
focusing on tweets generated before, during, and after
Hurricane Sandy, a superstorm which severely
impacted New York in 2012,
        <xref ref-type="bibr" rid="ref16">(Stowe et al., 2016)</xref>
        proposed an annotation schema to identify
relevant Twitter data (within a corpus of 22.2M unique
tweets from 8M unique Twitter users),
categorizing these tweets into fine-grained categories,
such as preparation and evacuation.
        <xref ref-type="bibr" rid="ref5">(Imran et al.,
2016)</xref>
        presented Twitter corpora composed of over
52 million crisis-related tweets, collected during
19 different crises that took place from 2013 to
2015. These corpora were manually-annotated
by volunteers and crowd-sourced workers
providing two types of annotations, the first one related
to a set of categories, the second one
concerning out-of-vocabulary words (e.g. slangs, places
names, abbreviations, misspellings). The
scholars then built machine-learning classifiers in
order to demonstrate the effectiveness of the
annotated datasets, also publishing word2vec word
embeddings trained on more than 52 million
messages. The preliminary results of this study posit
that a classification with a high precision of tweets
relevant to the disaster is possible to assist crisis
managers and first responders. Our study is not
devoted to monitor natural disasters but to
monitor natural human-caused disasters. More
specifically, the aim is to exploit NLP techniques to
contribute to the identification of intentional
environmental crimes through social media analysis. To
the best of our knowledge, this perspective of
investigation is rather novel in the field.
3
      </p>
    </sec>
    <sec id="sec-3">
      <title>The UNIOR Eye Corpus</title>
      <p>This section outlines the way the UNIOR Eye
corpus was created and how it is internally structured.
The research has been carried out in the
framework of the C4E - Crowd for the Environment
(Progetto PON Ricerca e Innovazione 2014-2020)
project2.</p>
      <p>The UNIOR Eye corpus is made up of 228,412
tweets related to environmental crimes
downloaded through Twitter API, covering the period
from 01 January 2013 to 06 August 2020. The
compilation phase of the corpus was divided into
two steps: the creation of a vocabulary containing
keywords related to environmental crimes and the
creation of the corpus. During this work phase, the
data was structured and organized according to the
different keywords, obtained from glossaries and
documents specific to the topic.</p>
      <p>Precisely, the following resources</p>
      <p>
        Glossario di termini sull’ambiente (FIMP, 2017) (a
guide from A to Z concerning the complex issue of
environmental pollution);
2http://www.unior.it/ateneo/20574/1/c4e-crowd-for-theenvironment-progetto-pon-ricerca-e-innovazione-20142020.html
Glossario dinamico per l’Ambiente ed il Paesaggio
        <xref ref-type="bibr" rid="ref6">(ISPRA, 2012)</xref>
        (a glossary supplied by the Italian Institute
for Environmental Protection and Research);
Glossario ambientale3 (a glossary supplied by the
national agency for the environmental protection of
Tuscany);
BeSafeNet4 (a glossary based on the Glossary on
Emergency Management, which has been developed in 2001
by European Centre of Technological Safety (TESEC)
of Euro-Mediterranean network of Centres EUR-OPA
Major Hazard Agreement of Council of Europe in
collaboration with other centres of network);
HERAmbiente5 (a glossary provided by Herambiente,
the largest company in the waste management sector);
Enciclopediambiente6 (the first freely available online
Encyclopedia on the Environment, designed by a group
of four engineers with the aim of spreading
“environmental knowledge”)
and the following two web sources
a dossier containing important provisions aimed at
dealing with environmental and industrial emergencies
and encouraging the development of the affected
areas7;
a document on environmental crimes and
environmental protection8.
were consulted. All of these language
resources contain information and definitions of
the basic terms related to environmental
disasters and crimes, e.g. Rifiuti pericolosi (hazardous
waste): waste products which can generate
potential/substantial risk to human health/the
environment if handled improperly. Hazardous waste
contains at least one of these characteristics:
flammability, corrosivity, or toxicity,9 and is included in
special lists. Here are some examples.
      </p>
      <p>HASHTAG HASHTAG Fiumicino: eternit e rifiuti
pericolosi al Passo della Sentinella URL HASHTAG
(HASHTAG HASHTAG Fiumicino: eternit and
hazardous waste in Passo della Sentinella URL
HASHTAG);
Cani in gabbia in discarica abusiva: Due animali tra
rifiuti pericolosi, amianto e bombole gas URL (Caged</p>
      <sec id="sec-3-1">
        <title>3http://www.arpat.toscana.it/glossario-ambientale</title>
        <p>4http://www.besafenet.net/it-it/glossary
5http://ha.gruppohera.it/glossario ambiente/
6http://www.enciclopediambiente.com
7https://www.senato.it/japp/bgt/showdoc/17/
DOSSIER/0/740667/index.html?part=dossier
dossier1sezione sezione12-h2 h28</p>
        <p>8https://scuola21.fermi.mn.it/documenti/reati ambientali.
pdf</p>
        <p>9https://eur-lex.europa.eu/legal-content/EN/TXT/PDF/
?uri=OJ:C:2018:124:FULL&amp;from=IT
dogs in illegal dump: two pets among hazardous waste,
asbestos and gas cylinders URL)
After this phase it was possible to create the corpus
by downloading from Twitter all the tweets
containing these keywords preceded by the hashtag.
These hashtags helped us to gather the
information needed to detect crimes against the
environment. More specifically, the corpus is internally
divided into semantic areas, each one concerning
a specific environmental crime: rifiuti e terra dei
fuochi (waste and Terra dei fuochi); reati contro le
acque (water-related crimes); materiali e sostanze
pericolose (hazardous substances and materials);
incendi e roghi ambientali (environmental fires).
These sets are further divided into more specific
subsets, e.g. the folder reati contro le acque
(water-related crimes) contains the subsets acque
di scarico, acque reflue, fiumi inquinati, liquami
(sewage, wastewater, polluted rivers, slurry). The
resulting corpus contains, therefore, a total of
228,412 tweets, 22,780,746 tokens, 569,905 types
with a type/token ratio (TTR) of 0.025.
4</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Case Study</title>
      <p>This section describes the steps taken to perform
the preliminary experiments on a selected part
of the UNIOR Eye corpus. First, the dataset on
which the experiments and data preparation
operations were carried out is presented, then the
preprocessing steps are listed and, finally, the
different machine learning approach used are described.</p>
      <sec id="sec-4-1">
        <title>4.1 Dataset</title>
        <p>As described in Section 3, the UNIOR Eye
corpus is divided into four semantic areas related to
the most common crimes against the environment.
Among the four semantic areas, we decided to use
the waste crimes subsection to test a specific use
case: whether an NLP system can understand and
report emergency text and user-generated reports.
Therefore, for the experiments described in this
paper, we focus our investigation on a sub-section
of the UNIOR Eye corpus, namely tweets about
waste related crimes and tweets with the hashtag
#terradeifuochi contained in the corresponding
semantic area: waste and Terra dei fuochi. This
subsection of the corpus contains 86,206 tweets. First,
for the total number of tweets, hashtags, mentions
and URLs are replaced with placeholder words.
Then tweets were annotated by the paper authors
on the basis of two labels: i) alert and ii) no alert,
i.e. if the tweet contains or not a message aimed at
reporting and locating a waste related crime.
Below, we provide a sample of annotated tweets
following our two labels, alert - no alert:
Ore 11:40 autostrada A1 altezza Afragola Acerra
direzione Roma. Roghi Tossici indisturbati, la
HASHTAG... URL HASHTAG HASHTAG (11:40 am A1
motorway near Afragola Acerra towards Rome.
Undisturbed toxic fires, the HASHTAG ... URL HASHTAG
HASHTAG) — ALERT
MENTION ministro, piuttosto che pensare alla
HASHTAG pensi ai continui roghi MENTION (MENTION
Minister, rather than thinking about the HASHTAG
think about the continuous fires MENTION) — NO
ALERT
During the annotation phase, we noted that the no
alert class is the one which contains the majority
of tweets and includes examples of hate speech,
satirical texts, news about emergency actions as
well as politically oriented texts. Consequently,
our dataset built in this way is unbalanced for the
two classes, counting 81,235 tweets for the no
alert class and 4,970 alert tweets. In order to
visualize alert tweets, we exploit Carto10, a cloud
computing platform that provides a geographic
information system, web mapping, and spatial data
science tools11.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2 Inter-annotator Agreement</title>
        <p>When different annotators label a corpus, it is
important to calculate the inter-annotator agreement
(IAA) with a twofold objective: i) make sure that
annotators agree and ii) test the clarity of
guidelines. As previously mentioned, the dataset
(composed of 86,206 tweets) has been annotated by
four of the paper authors on the basis of two labels:
i) alert and ii) no alert. This implies that each
author annotated about 21,000 tweets. Then, to
calculate inter-annotator agreement we randomly
selected 10% of the tweets (i.e. 8,620) which were
tagged by all annotators.</p>
        <p>
          The agreement among the four annotators is
measured using Krippendorff’s coefficient; instead,
to estimate the agreement between pairs of
annotators, we use Cohen’s coefficient
          <xref ref-type="bibr" rid="ref1">(Artstein and
Poesio, 2008)</xref>
          . Taking into account the
recommendations set out in
          <xref ref-type="bibr" rid="ref1 ref8">(Artstein and Poesio, 2008;
Krippendorff, 2004)</xref>
          , we interpret the values obtained
10carto.com
11A map showing toxic fires alert tweets in the UNIOR
Eye corpus is available at this link https://uniornlp.carto.com/
builder/04f2cca9-08cd-4b9f-90cd-79fc0d93af42/embed
in IAA according to the strength of agreement
criteria described in
          <xref ref-type="bibr" rid="ref9">(Landis and Koch, 1977)</xref>
          for
each pair of annotators; whereas, for agreement
among four annotators, we follow the suggested
standard in
          <xref ref-type="bibr" rid="ref8">(Krippendorff, 2004)</xref>
          . The calculated
value of Krippendorff’s is 0.706. Considering
the standard value in
          <xref ref-type="bibr" rid="ref8">(Krippendorff, 2004)</xref>
          , our
value of =0.706 is considered as acceptable and
expressing a good data reliability. In Table 1 we
show the results for pairs of annotators.
        </p>
        <p>
          Pair of annotators
a1 - a2
a1 - a3
a1 - a4
a2 - a3
a2 - a4
a3 - a4
According to
          <xref ref-type="bibr" rid="ref9">(Landis and Koch, 1977)</xref>
          , five out of
six Cohen’s values show a “substantial” strength
of agreement for each pair; while a pair (a1-a4)
show a value considered “almost perfect” in the
research cited.
4.3
        </p>
      </sec>
      <sec id="sec-4-3">
        <title>Preprocessing</title>
        <p>Before feeding the machine learning algorithms,
some pre-processing steps are performed. Since
the majority of mentions and hashtags are shared
by both alert and no-alert samples, we focus on
the tweet itself, by removing any reference to
people, entities and organizations conveyed through
hashtags and mentions. Therefore, the placeholder
words related to hashtags, URLs and mentions
are removed. Then, punctuation is removed from
the tweets along with a custom list of function
words such as determiners, prepositions and
conjunctions. Finally, the tweets are lower-cased and
the tokenization is performed.
4.4</p>
      </sec>
      <sec id="sec-4-4">
        <title>Machine Learning Approaches</title>
        <p>We set the problem of tweets related to waste
crimes as a supervised binary classification
problem between different textual content.</p>
        <p>
          To tackle the problem as first task within the C4E
Project, we select a machine learning approach
using Support Vector Machines (SVM) with
linear kernel and C=1 and Multinomial Naive Bayes
(MNB) as classification algorithms
          <xref ref-type="bibr" rid="ref4">(Imran et al.,
2015)</xref>
          . Since the task concerns the classification
of tweets belonging to the alert class, to deal with
the unbalanced dataset, we use the undersampling
technique by automatically reducing the number
of samples for the majority class (no alert)
          <xref ref-type="bibr" rid="ref10">(Li et
al., 2009)</xref>
          , until they were balanced with the
samples of the alert class. We used the tf-idf technique
to extract the features used by both algorithms. To
build algorithms and extract features, we used the
Python scikit-learn library.
        </p>
        <p>
          In addition to the MNB and SVM with tf-idf
technique, we built two models with sentence
embeddings as features and SVM with the tuning of C
parameter as a classification algorithm. In the first
model (FT-SVM), we used the Italian pre-trained
word vectors from fastText12
          <xref ref-type="bibr" rid="ref2">(Bojanowski et al.,
2017)</xref>
          to build our sentence embeddings by
averaging word embeddings for all tokens for each
tweet; then, C=10 is found as the best C parameter
value using GridSearchCV13 instance. In the
second model (mDB-SVM), we generated sentence
embeddings using the pretrained multilingual
DistilBERT
          <xref ref-type="bibr" rid="ref14">(Sanh et al., 2019)</xref>
          model from
Transformers14. To accomplish this, each tweet is
represented as a list of tokens and, then, each list
is padded to the same size (max len = 94). The
attention mask is used. Before fitting the
sentence embeddings thus constructed in the SVM
classifier, it is searched for the best value of the
C parameter set to C=0.1. For both models
(FTSVM and mDB-SVM) the pre-processing steps
described above are performed.
5
        </p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Results</title>
      <p>In this section, we show the results obtained by our
models in terms of Precision, Recall, F-Measure
and Accuracy. For all models, the results are
obtained on 30% of the dataset set aside as a test
set, keeping the samples balanced between the two
classes. Furthermore our models were evaluated
using a 10-Fold Cross-Validation15.</p>
      <p>As a baseline to compare with, we used Dummy
classifier which achieves an accuracy of 0.501. On
the test set, the SVM classifier achieves an
accuracy of 0.870, while for the MNB classifier it is
0.839. Regarding the evaluation by 10-fold cross
12https://fasttext.cc/docs/en/pretrained-vectors.html
13https://scikit-learn.org/stable/modules/generated/
sklearn.model selection.GridSearchCV.html#</p>
      <p>14https://huggingface.co/transformers/pretrained models.
html</p>
      <p>15https://scikit-learn.org/stable/modules/generated/
sklearn.model selection.KFold.html
validation, our SVM reaches an accuracy of 0.868
with the mean and standard deviation of 0.008,
instead the accuracy of the MNB is 0.841 with the
mean and standard deviation of 0.010. In Table 2
we show the performances achieved by both
models.</p>
      <p>MNB
alert
no alert</p>
      <p>SVM
alert
no alert</p>
      <sec id="sec-5-1">
        <title>Precision</title>
        <p>0.871
0.807
Precision
0.857
0.883</p>
        <p>Both classifiers with tf-idf achieve good accuracy
and seem to have a good ability to classify a
considerable amount of tweets providing good
results in terms of precision and recall. One of the
reasons for these performances may be ascribed
to a discriminating lexical composition regarding
the samples belonging to the alert and no alert
classes.</p>
        <p>Regarding the accuracy of sentence embeddings
models on the test set, FT-SVM reaches an
accuracy of 0.822, while mDB 0.774. By evaluating
the predictive performance of the two models with
10-fold cross-validation, FT-SVM achieves an
accuracy of 0.825 with the mean and standard
deviation of 0.011, while mDB-SVM reaches the
accuracy of 0.773 with the mean and standard
deviation of 0.013. In Table 3, the results in terms of
Precision, Recall and F-Measure are shown.</p>
        <p>FT-SVM</p>
        <p>alert
no alert
mDB-SVM</p>
        <p>alert
no alert</p>
      </sec>
      <sec id="sec-5-2">
        <title>Precision</title>
        <p>0.826
0.818
Precision
0.785
0.765</p>
      </sec>
      <sec id="sec-5-3">
        <title>Recall</title>
        <p>0.817
0.827
Recall
0.766
0.783</p>
      </sec>
      <sec id="sec-5-4">
        <title>F-Measure 0.821</title>
        <p>0.822
F-Measure
0.775
0.774</p>
        <p>Both models fed with sentence embeddings
constructed with different techniques, seem to
perform well in this classification task. In
particular, the FT-SVM model based on sentence
embeddings built with FastText seems to have better
scores in terms of Precision and F-measure than
those achieved by the mDB-SVM model. One
of the reasons could be that sentence embeddings
built with FastText benefit from a resource tailored
on the Italian language compared to a multilingual
one used in DBert-SVM. Specifically, mDB-SVM
achieved good results in terms of precision and
fmeasure for the alert class. Instead, in terms of
Recall, both models have a high proportion of
relevant instances for the no alert class.
5.1</p>
        <sec id="sec-5-4-1">
          <title>Confusion Matrices</title>
          <p>
            In this section we show the four confusion
matrices in order to graphically display the
performances achieved by the different models. In
Figure 1 we show the confusion matrix of the MNB
model, while in Figure 2 that of the SVM model.
The confusion matrices of the FT-SVM and the
mDB-SVM model are shown respectively in
FigWe presented a case study within the C4E project
aimed at monitoring social media to provide
support against environmental crimes. In particular,
we described the UNIOR Eye corpus, in some
sections still in progress, on which we tested four
models with three different features extraction and
construction techniques on a part of the corpus.
We proposed two classifiers, namely SVM and
MNB, with tf-idf features as the first experiment;
then, SVM with C parameter tuning fed with
sentence embeddings. These embeddings were built
both using Italian pre-trained fastText model and
using pre-trained DistilBert multilingual model.
Our purpose was to classify alert tweets related
to waste crimes vs no alert tweets. Future
research will include the enlargement of the corpus,
applications of NLP in the field of
environmental protection as well as the analysis of contextual
features related to environmental issues used as a
medium to polarize public opinion
            <xref ref-type="bibr" rid="ref7">(Karol, 2018)</xref>
            .
          </p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgements</title>
      <p>This research has been carried out within the
framework of two Innovative Industrial PhD
projects supported by the PON Ricerca e
Innovazione 2014/20 and the POR Campania FSE
2014/2020 funds and two research grants
supported by the PON Ricerca e Innovazione 2014/20
in the context of the C4E project.</p>
      <p>Authorship contribution is as follows: Raffaele
Manna is author of section 4. Section 2 is by
Antonio Pascucci. Section 5 is by Raffaele Manna
and Antonio Pascucci. Sections 1, 3 and 6 are
by Wanda Punzi Zarino and Vincenzo Simoniello.
We are grateful to Prof. Johanna Monti for
supervising the research.
FIMP. 2017. Fimp ambiente - federazione italiana
medici pediatriglossario di termini sull’ambiente.
una guida dalla a alla z per orientarsi nel complesso
tema dell’inquinamento ambientale.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <string-name>
            <given-names>Ron</given-names>
            <surname>Artstein</surname>
          </string-name>
          and
          <string-name>
            <given-names>Massimo</given-names>
            <surname>Poesio</surname>
          </string-name>
          .
          <year>2008</year>
          .
          <article-title>Inter-coder agreement for computational linguistics</article-title>
          .
          <source>Computational Linguistics</source>
          ,
          <volume>34</volume>
          (
          <issue>4</issue>
          ):
          <fpage>555</fpage>
          -
          <lpage>596</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <string-name>
            <given-names>Piotr</given-names>
            <surname>Bojanowski</surname>
          </string-name>
          , Edouard Grave, Armand Joulin, and
          <string-name>
            <given-names>Tomas</given-names>
            <surname>Mikolov</surname>
          </string-name>
          .
          <year>2017</year>
          .
          <article-title>Enriching word vectors with subword information</article-title>
          .
          <source>Transactions of the Association for Computational Linguistics</source>
          ,
          <volume>5</volume>
          :
          <fpage>135</fpage>
          -
          <lpage>146</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <string-name>
            <surname>William J Corvey</surname>
            , Sarah Vieweg, Travis Rood, and
            <given-names>Martha</given-names>
          </string-name>
          <string-name>
            <surname>Palmer</surname>
          </string-name>
          .
          <year>2010</year>
          .
          <article-title>Twitter in mass emergency: What nlp can contribute</article-title>
          .
          <source>In Proceedings of the NAACL HLT 2010 Workshop on Computational Linguistics in a World of Social Media</source>
          , pages
          <fpage>23</fpage>
          -
          <lpage>24</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <string-name>
            <given-names>Muhammad</given-names>
            <surname>Imran</surname>
          </string-name>
          , Carlos Castillo, Fernando Diaz, and
          <string-name>
            <given-names>Sarah</given-names>
            <surname>Vieweg</surname>
          </string-name>
          .
          <year>2015</year>
          .
          <article-title>Processing social media messages in mass emergency: A survey</article-title>
          .
          <source>ACM Computing Surveys (CSUR)</source>
          ,
          <volume>47</volume>
          (
          <issue>4</issue>
          ):
          <fpage>1</fpage>
          -
          <lpage>38</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <string-name>
            <given-names>Muhammad</given-names>
            <surname>Imran</surname>
          </string-name>
          , Prasenjit Mitra, and
          <string-name>
            <given-names>Carlos</given-names>
            <surname>Castillo</surname>
          </string-name>
          .
          <year>2016</year>
          .
          <article-title>Twitter as a lifeline: Humanannotated twitter corpora for nlp of crisis-related messages</article-title>
          .
          <source>arXiv preprint arXiv:1605</source>
          .
          <fpage>05894</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <string-name>
            <surname>ISPRA.</surname>
          </string-name>
          <year>2012</year>
          .
          <article-title>Ispra - l'istituto superiore per la protezione e la ricerca ambientale, glossario dinamico per l'ambiente ed il paesaggio</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          <string-name>
            <given-names>David</given-names>
            <surname>Karol</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>Party polarization on environmental issues: toward prospects for change</article-title>
          . Research Paper. Niskanen Center, Washington, DC.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          <string-name>
            <given-names>Klaus</given-names>
            <surname>Krippendorff</surname>
          </string-name>
          .
          <year>2004</year>
          .
          <article-title>Reliability in content analysis: Some common misconceptions and recommendations</article-title>
          .
          <source>Human communication research</source>
          ,
          <volume>30</volume>
          (
          <issue>3</issue>
          ):
          <fpage>411</fpage>
          -
          <lpage>433</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          <string-name>
            <given-names>J Richard</given-names>
            <surname>Landis</surname>
          </string-name>
          and Gary G Koch.
          <year>1977</year>
          .
          <article-title>The measurement of observer agreement for categorical data</article-title>
          .
          <source>biometrics</source>
          , pages
          <fpage>159</fpage>
          -
          <lpage>174</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          <string-name>
            <given-names>Yaoyong</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Kalina</given-names>
            <surname>Bontcheva</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Hamish</given-names>
            <surname>Cunningham</surname>
          </string-name>
          .
          <year>2009</year>
          .
          <article-title>Adapting svm for data sparseness and imbalance: a case study in information extraction</article-title>
          .
          <source>Natural Language Engineering</source>
          ,
          <volume>15</volume>
          (
          <issue>2</issue>
          ):
          <fpage>241</fpage>
          -
          <lpage>271</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          <string-name>
            <given-names>Miguel</given-names>
            <surname>Maldonado</surname>
          </string-name>
          , Darwin Alulema, Derlin Morocho, and Marida Proan˜o.
          <year>2016</year>
          .
          <article-title>System for monitoring natural disasters using natural language processing in the social network twitter</article-title>
          .
          <source>In 2016 IEEE International Carnahan Conference on Security Technology (ICCST)</source>
          , pages
          <fpage>1</fpage>
          -
          <lpage>6</lpage>
          . IEEE.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          <string-name>
            <given-names>Graham</given-names>
            <surname>Neubig</surname>
          </string-name>
          , Yuichiroh Matsubayashi, Masato Hagiwara, and
          <string-name>
            <given-names>Koji</given-names>
            <surname>Murakami</surname>
          </string-name>
          .
          <year>2011</year>
          .
          <article-title>Safety information mining-what can nlp do in a disaster-</article-title>
          .
          <source>In Proceedings of 5th International Joint Conference on Natural Language Processing</source>
          , pages
          <fpage>965</fpage>
          -
          <lpage>973</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          <string-name>
            <given-names>Pasquale</given-names>
            <surname>Peluso</surname>
          </string-name>
          .
          <year>2015</year>
          .
          <article-title>Dalla terra dei fuochi alle terre avvelenate: lo smaltimento illecito dei rifiuti in italia</article-title>
          .
          <source>Rivista di Criminologia</source>
          ,
          <source>Vittimologia e Sicurezza</source>
          ,
          <volume>9</volume>
          (
          <issue>2</issue>
          ):
          <fpage>13</fpage>
          -
          <lpage>30</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          <string-name>
            <given-names>Victor</given-names>
            <surname>Sanh</surname>
          </string-name>
          , Lysandre Debut, Julien Chaumond, and
          <string-name>
            <given-names>Thomas</given-names>
            <surname>Wolf</surname>
          </string-name>
          .
          <year>2019</year>
          .
          <article-title>Distilbert, a distilled version of bert: smaller, faster, cheaper and lighter</article-title>
          . arXiv preprint arXiv:
          <year>1910</year>
          .01108.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          <string-name>
            <given-names>Muhammed</given-names>
            <surname>Ali</surname>
          </string-name>
          <string-name>
            <surname>Sit</surname>
          </string-name>
          , Caglar Koylu, and
          <string-name>
            <given-names>Ibrahim</given-names>
            <surname>Demir</surname>
          </string-name>
          .
          <year>2019</year>
          .
          <article-title>Identifying disaster-related tweets and their semantic, spatial and temporal context using deep learning, natural language processing and spatial analysis: a case study of hurricane irma</article-title>
          .
          <source>International Journal of Digital Earth</source>
          ,
          <volume>12</volume>
          (
          <issue>11</issue>
          ):
          <fpage>1205</fpage>
          -
          <lpage>1229</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          <string-name>
            <given-names>Kevin</given-names>
            <surname>Stowe</surname>
          </string-name>
          , Michael Paul, Martha Palmer, Leysia Palen, and
          <string-name>
            <surname>Kenneth M Anderson</surname>
          </string-name>
          .
          <year>2016</year>
          .
          <article-title>Identifying and categorizing disaster-related tweets</article-title>
          .
          <source>In Proceedings of The fourth international workshop on natural language processing for social media</source>
          , pages
          <fpage>1</fpage>
          -
          <lpage>6</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          <string-name>
            <given-names>Francesco</given-names>
            <surname>Tarasconi</surname>
          </string-name>
          , Michela Farina, Antonio Mazzei, and
          <string-name>
            <given-names>Alessio</given-names>
            <surname>Bosca</surname>
          </string-name>
          .
          <year>2017</year>
          .
          <article-title>The role of unstructured data in real-time disaster-related social media monitoring</article-title>
          .
          <source>In 2017 IEEE International Conference on Big Data (Big Data)</source>
          , pages
          <fpage>3769</fpage>
          -
          <lpage>3778</lpage>
          . IEEE.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>