<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Lang. NOT HOF HATE OFFN PRFN Total
English</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Overview of the HASOC track at FIRE 2019: Hate Speech and O ensive Content Identi cation in Indo-European Languages</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Prasenjit Majumder</string-name>
          <email>prasenjit.majumder@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Daksh Patel</string-name>
          <email>dakshpatel68@gmail.com</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>DA-IICT</institution>
          ,
          <addr-line>Gandhinagar</addr-line>
          ,
          <country country="IN">India</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>LDRP-ITR</institution>
          ,
          <addr-line>Gandhinagar</addr-line>
          ,
          <country country="IN">India</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>University of Hildesheim</institution>
          ,
          <country country="DE">Germany</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>3591</year>
      </pub-date>
      <volume>2261</volume>
      <abstract>
        <p>The identi cation of Hate Speech in Social Media has received much attention in research recently. There is a particular demand for research for languages other than English. The rst edition of the HASOC track creates resources for Hate Speech Identi cation in Hindi, German, and English. Three datasets were developed from Twitter, and Facebook and made available. HASOC intends to stimulate research and development for Hate Speech classi cation for di erent languages. The datasets allow the development and testing of supervised machine learning systems. Binary classi cation and more ne-grained sub-classes were o ered in 3 sub tasks. For all sub-tasks, 321 experiments were submitted. For the classi cation task, models based on deep learning methods have proved to be adequate. The approaches used most often were Long-ShortTerm memory (LSTM) networks with distributed word representation of the text. The performance of the best system for identi cation of Hate Speech for English, Hindi, and German was a Marco-F1 score of 0.78, 0.81, and 0.61, respectively. This overview provides details insights and analyzes the results.</p>
      </abstract>
      <kwd-group>
        <kwd>Hate Speech Text Classi cation Evaluation Deep Learning</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>The large fraction of Hate Speech and other o ensive and objectionable content
online poses a huge challenge to societies. O ensive language such as insulting,
hurtful, derogatory, or obscene content directed from one person to another
person and open for others undermines objective discussions. There is a growing
need for research on the classi cation of Hate Speech into di erent categories
of o ensive content on di erent platforms of social media without human
assistance.</p>
      <p>In October 2019, the European Court of Justice decided that platforms need
to take down content worldwide even after national decisions. In a particular
case, the EU court debated defamatory posts on Facebook. Even posts similar in
tone need to be addressed and the ruling explicitly mentions automatic systems.
This shows that automatic systems are of high social relevance. Recently, also
the founder of Facebook proposed ideas for the regulation of the Internet. He
demanded standards and baselines for the de nition of harmful content. Such
clear de nitions have not been provided and are unlikely to be developed in the
near future. This makes research and annotated corpora even more necessary.</p>
      <p>The identi cation of Hate Speech within a collection or a stream of tweets
is a challenging task because systems cannot rely on the text content. Based
on content, text classi cation systems have been successful. However, Hate text
might have many issues. Hate often has no clear signal words, and word lists, as
in sentiment analysis, are expected to work less well.</p>
      <p>In order to contribute to this research, w this overview paper presents the
1st edition of HASOC Hate Speech and O ensive Content Identi cation in
IndoEuropean Languages, namely: German, English, and Hindi. The dataset for
all three languages was created from Twitter and Facebook. HASOC consists
of three tasks, a coarse-grained binary classi cation task, and two ne-grained
multi-class classi cations. Of course, freedom of speech needs to be guaranteed
in democratic societies for future development. Nevertheless, the o ensive text
which hurts others' sentiments needs to be restricted. As there is such an
increase in the usage of abuse on many internet platforms, technological support
for the recognition of such posts is necessary. The use of supervised learning with
the annotated dataset is a key strategy for advancing such systems. There has
been signi cant work in several languages in particular for English. However,
there is a lack of research on this recent and relevant topic for most other
languages. This track intends to develop data and evaluation resources for several
languages. The objectives are to stimulate research for these languages and to
nd out the quality of hate speech detection technology in other languages.</p>
      <p>The HASOC dataset provides several thousands labeled social media posts
for each language. The entire dataset was annotated and checked by the
organizers of the track. The annotation architecture is designed to create data for 3
di erent sub tasks.
1. SUB-TASK A: classi cation of Hate Speech (HOF) and non-o ensive
content.
2. SUB-TASK B: If the post is HOF, sub-task B is used to identify the type of
hate.
3. SUB-TASK C: it decides the target of the post.</p>
      <p>
        Hate Speech detection is of great signi cance and attracting many researchers.
Recent overview papers provide a good introduction to the scienti c issues that
are involved in Hate Speech identi cation [
        <xref ref-type="bibr" rid="ref12 ref36">12,36</xref>
        ].
      </p>
      <p>https://www.nytimes.com/2019/10/03/technology/facebook-europe.html
https://www.faz.net/aktuell/wirtschaft/diginomics/facebook-ceo-zuckerberg-ideasto-regulate-the-internet-16116032.html
2</p>
    </sec>
    <sec id="sec-2">
      <title>Related Forum and Dataset</title>
      <p>
        Collections are an important asset for any supervised classi cation methods.
For Hate Speech, several previous initiatives have created corpora that have
been used for research. There has been signi cant work in several languages,
in particular for English. However, for other languages, such as Hindi standard
datasets are not available and HASOC is an attempt to create the labeled dataset
for such low resource language. HASOC is primarily inspired by two previous
evaluation forums, GermEval [
        <xref ref-type="bibr" rid="ref44">44</xref>
        ], and O ensEval [
        <xref ref-type="bibr" rid="ref47">47</xref>
        ], and tries to leverage the
synergies of these initiatives.
      </p>
      <p>
        Data sampling is a paramount task for any data challenges competition. Some
of the corpora focuses in speci c on certain targets, like immigrants, women
(HateEval) [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]or racism (e.g. [
        <xref ref-type="bibr" rid="ref39">39</xref>
        ]). Others focus on Hate Speech in general (e.g.
HaSpeeDe [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]) or other unacceptable text types. A recent trend is to introduce
a more ne-grained classi cation. Some data challenges require detailed analysis
for the hateful comments, like detection of the target (HateEval and O ensEval)
or the type of Hate Speech (GermEval). Others focus on the severity of the
comment (Kaggle Toxic [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]). A recent and very interesting collection is CONAN.
It o ers Hate Speech and the reactions to it [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. This could open opportunities
for detecting Hate Speech by analyzing it jointly with the following posts. Table
1 summarize standard Hate speech dataset available at various forum.
      </p>
      <p>There is a huge demand for many languages other than English. HASOC is
the rst shared task which developed a resource for three languages together
and which encourages multilingual research.
3</p>
    </sec>
    <sec id="sec-3">
      <title>Task Description</title>
      <p>HASOC and most other collections provide the text of a post and require systems
to detect hateful content. No context or meta-data like time related features or
the network of the actors are given which might make these tasks somewhat
unrealistic. Platforms can obviously use all meta-data of a post and a user.
However, the distribution of such data poses legal issues. The following tasks
have been proposed in HASOC 2019:
Sub-task A : Sub-task A focuses on Hate speech and O ensive language
identi cation and is o ered for English, German, Hindi. Sub-task A is coarse-grained
binary classi cation in which participating system are required to classify tweets
into two class, namely: Hate and O ensive (HOF) and Non- Hate and o ensive.
1. (NOT) Non Hate-O ensive - This post does not contain any Hate speech,
o ensive content.
2. (HOF) Hate and O ensive - This post contains Hate, o ensive, and profane
content.</p>
      <p>During our annotation, we labeled posts as HOF in case they contained
any form of non-acceptable language such as hate speech, aggression, profanity;
otherwise they were labeled as NOT.
Sub-task B : Sub-task B represents a ne-grained classi cation. Hate-speech
and o ensive posts from the sub-task A are further classi ed into three
categories.
1. (HATE) Hate speech: Posts contain Hate speech content.
2. (OFFN) O ensive: Posts contain o ensive content.
3. (PRFN) Profane: These posts contain profane words
HATE SPEECH : Describing negative attributes or de ciencies to groups of
individuals because they are members of a group (e.g. all poor people are stupid).
Hateful comment toward groups because of race, political opinion, sexual
orientation, gender, social status, health condition or similar.</p>
      <p>OFFENSIVE : Posts which are degrading, dehumanizing, insulting an
individual, threatening with violent acts are categorized into this category.
PROFANITY : Unacceptable language in the absence of insults and abuse. This
typically concerns the usage of swearwords (Schei e, Fuck etc.) and cursing (Hell!
Verdammt! etc.). Such posts are categorized into this category. As expected, most
posts are in the category NOT, some are HATE and the other two categories
are less frequent. Dubious cases which are di cult to decide even for humans,
were left out.</p>
      <sec id="sec-3-1">
        <title>Sub-task C (only for English and Hindi) : Sub-task C considers the type</title>
        <p>of o ense. Only posts labeled as HOF in sub-task A are included in sub-task C.
The two categories in sub-task C are the following:
1. Targeted Insult (TIN): Posts containing an insult/threat to an individual,
group, or others.
2. Untargeted (UNT): Posts containing non targeted profanity and swearing.</p>
        <p>Posts with general profanity are not targeted, but they contain non-acceptable
language.
4</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Data Set and Collection</title>
      <p>
        The following sections explain how the data set was created and enriched by
annotations. First, the authors searched with heuristics for typical Hate Speech
in online fora. They identi ed topics for which many hate posts can be expected.
Di erent hashtags and keywords were used for all three languages. For some of
the found posts, the id of the author was recorded. For a number of such users,
the timeline was collected. Based on tweets found, we crawled the last posts of
the authors to increase variety. The systems are less likely to classify individual
textual style when they have a rich set of posts from an author. This procedure
was intended to decrease bias and was inspired by GermEval [
        <xref ref-type="bibr" rid="ref43">43</xref>
        ].
      </p>
      <p>The HASOC dataset was subsequently sampled from Twitter and partially
from Facebook for all the three languages. The Twitter API gives a large number
Classes
NOT
HATE
HATE
OFFN
PRFN
PRFN
UNT
OFFN
TIN
Sample tweet from the class
4 matches were can't play due to rain and many more will
be not played r the same reason . Conclusion this world
cup is no more world cup. #ShameOnICC #RainCup
Are Muslims, in general a nuisance to be tolerated by
the rest of the world ? #SaveBengal #DoctorsFightBack
#DoctorsStrike #MamtaBanerjee
#TerroristNationPakistan 90% Pakistanis wants war with
India and 10% said war should not be. And Those 10%
belongs to Pakistans Armed Forces
#TerroristNationPakistan
#Just a daily reminder to @realDonaldTrump that he is
a National Disgrace. #TraitorTrump #TrumpIsADisgrace
#TrumpIsATraitor
@cizzacampbell Didn't realise you were an expert
#dickhead
Who voted for a no-deal? Tell me, who the fuck voted for
a no deal? The way I see it, the referendum was a corrupt
vote between remain and leave. Not remain, leave, deal,
no deal. Nobody voted for no deal!!
@realDonaldTrump Will it be worse than killing children?
Worse than selling your country to the Russians? Worse
than saying you love a ruthless dictator? Probably not.
#TrumpIsATraitor
of recent tweets which resulted in an unbiased dataset. Thus the tweets were
acquired using hashtags and keywords that contained o ensive content. The
collection was provided to participants without metadata. We have developed
Twitter and Facebook plugins to fetch the posts without using the API. The
size of the data corpus is shown in tables 2 and 3.</p>
      <p>During the labeling process, several juniors for each language engaged with an
online system to judge the tweets. The system can be seen in gure 1 and gure 2.
They were given short guidelines that contained the information as mentioned
in section 3.1. The process is highly subjective, and even after discussions of
questionable often no agreement could be reached. This lies in the nature of
Hate Speech.</p>
      <p>
        As pointed out in the study by Ross et al. [
        <xref ref-type="bibr" rid="ref32">32</xref>
        ], not even with providing
written guidelines can improve the agreement. Consequently, and to be sure that
people can see them on one page, we tried to keep the guidelines short. The
guidelines for HASOC are listed in the annex. A study by Salminen et al. [
        <xref ref-type="bibr" rid="ref35">35</xref>
        ]
showed that the dubious and questionable cases led to much more disagreement
than clear cases with obvious Hate Speech characters. Jhaver et al. [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] and
colleagues interviewed both the receivers and the senders of some posts which were
considered to be aggressive. They revealed that the senders often did not agree
with the judgment of readers. Among other arguments, they brought forward
that some messages were regarded as hateful because people did not want to be
confronted with the arguments. Again, this study shows that there is a great
deal of subjectivity involved and that also context matters.
      </p>
      <p>
        The di culties during assessment in HASOC were often related to the use
of language registers like youth talk and irony or indirectness which might not
be understood by all readers. A more detailed analysis of the issues encountered
during the HASOC annotation for German has been carried out[
        <xref ref-type="bibr" rid="ref42">42</xref>
        ]. u
      </p>
      <p>The overlap between annotators for task A for English, Hindi, and German
for a subset to tweets and posts annotated twice was 89%, 91%, and 32%,
respectively. Further statistical details of the annotation process can be seen in
table 5. The e ects of such disagreement need to be analyzed in the future.
The values show that the labeling task is hard overall. The second sub-task
can only be solved with a lower quality. For the sub-task C, the quality does not
drop much of is even higher than that of sub-task B .</p>
      <p>We also calculated the (Kappa) coe cient due to the high imbalance of
the data sets. Using the scikit-Learn package, the inter annotator agreement for
the rst two annotators for a tweet was determined. Table 6 shows values of
in sub-task A for all three languages</p>
      <p>
        The degree of disagreement might also result from the topics present in the
collection[
        <xref ref-type="bibr" rid="ref46">46</xref>
        ]. The issues and the level of disagreement need to be analyzed in
the future.
5
      </p>
    </sec>
    <sec id="sec-5">
      <title>Evaluation Metrics</title>
      <p>The metrics for classi cation should combine both recall and precision. The
F1score has many variants like weighted F1, Macro-F1 or micro-F1. For multi-class
classi cation, the distribution of class labels is often unbalanced. The weighted
F1-score calculates the F1 score for each class independently. When it adds them,
it uses a weight based on the number of true labels of each class. Therefore, it
gives a bias for the majority class. The 'macro' calculates the F1 separately
for each class but does not use weights for the aggregation. This results in a
stronger penalization when a system does not perform well for the minority
classes. Choice of the variant of F1-measure depends on the objective of the tasks
and the distribution of label in the dataset. Hate Speech related classi cation
problems su er from class imbalance. Therefore, the macro F1 is the natural
choice for the evaluation.
6</p>
    </sec>
    <sec id="sec-6">
      <title>Results</title>
      <p>Overall, 103 registrations were submitted for the track. 37 teams submitted runs
and 25 teams have submitted papers. 321 runs were submitted by 37 teams in
all the sub-tasks.</p>
      <p>The following sections show the sub-tasks of HASOC. The approaches of
all teams are brie y summarized in the annex of this paper. For details on
the technical implementation, the reader is referred to the descriptions of the
participating teams in this volume.
6.1</p>
      <sec id="sec-6-1">
        <title>English Dataset</title>
        <p>
          In the English language, Total 174 runs were submitted across 3 sub-tasks. The
YNU wb team [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] used an LSTM approach with ordered neurons and applied an
attention mechanism. The absolute di erences between the top runs are rather
small. Table 8 presents the results of the top 10 teams of the English sub-task
A.
        </p>
        <p>The plot of the performance of all systems in Figure 3 shows that the Median
of the runs lies quite close to the top performance.</p>
        <p>Despite the similar performance of many teams, the recall-precision graph in
gure 4 shows that there are considerable di erences between the systems which
the F1 measures do not reveal.</p>
        <p>
          The overall F1 measures for sub-task B and C are much lower than for
subtask A. Table 9 and 10 shows the results of these tasks. The best performing
team [
          <xref ref-type="bibr" rid="ref21">21</xref>
          ] for sub-task B and sub-task C used the relatively new BERT model for
classi cation. This shows that it performed well for both sub-task A with more
training samples as well as for sub-task B with much fewer training instances.
        </p>
        <p>
          The performance for task C shows that the weighted F1 values are very close
together and that run number 10 has even a higher values than run number 1.
The careful selection of metrics is crucial. The boxplots in gure 5, and 6 show
that the Median lies again close to the top performing run for sub-tasks B and
C.
In the Hindi language, total 93 runs were submitted across 3 sub-tasks. The
QutNocturnal team [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ] used a CNN base approach with Word2vec embedding.
The absolute di erences between the top runs are rather small. Table 11 presents
results of the top team of Hindi sub-task A The absolute values for Hindi
subtask A are comparable to the English sub-task and the top-performing systems
are again close to each other.
In the German language, total 54 runs were submitted across 2 sub-tasks and
only, the rst two sub-tasks were possible. The Macro F1 score is lower than
for the other two languages. For sub-task A, the best team used BERT sentence
embedding and the multilingual sentence embedding LASER. Table 14 and 15
present result of sub-task A and B.
        </p>
        <p>
          The LSV team [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ] performed second and rst for sub-task B. They apply
the BERT model and use additional corpora for similar tasks. Boxplots of the
performance of all the participants team are shown in gure 10 and 11.
7
        </p>
      </sec>
    </sec>
    <sec id="sec-7">
      <title>Approaches</title>
      <p>
        The top performance for the sub-task A for English and and Hindi three
languages is delivered by systems based on Deep neural models. Even new
architectures for which little experience is available like BERT have been applied
Standing Team name
1 LSV-UdS [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]
2 LSV-UdS [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]
3 HateMonitors [
        <xref ref-type="bibr" rid="ref34">34</xref>
        ]
4 3Idiots [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ]
5 Cs
6 3Idiots [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ]
7 FalsePostive [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ]
8 FalsePostive [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ]
9 FalsePostive [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ]
10 LSV-UdS [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]
with great success. There is even true for sub-task B for German where only
few training examples were available. There needs to be considered that most
systems applied a Deep Learning system (see annex B). However, for Hindi the
top performance comes from a traditional machine learning system. Even for the
other two languages, we can observe that some of the few non-Deep Learning
systems lead to a performance quite close to the top performance. For example,
Team A3-108 [
        <xref ref-type="bibr" rid="ref23">23</xref>
        ] reaches a result close to the top performance for the Hindi
subtask B. Also the run IRLAB@IITBHU [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] achieves a higher weighted F1 value
than the top run for sub-task B for English. It seems that the size of HASOC
is small enough that traditional approaches can still prevail. There might not
be enough data to train Deep architectures with many parameters. Future
improvement for such systems might lie in the intelligent use of external resources.
Participants were allowed to use external resources and other datasets for this
task. For German, this seems to have boosted the top performing team LSV-UdS
for the sub-task B for which only few training examples were available.
      </p>
      <p>Several teams have adopted an open code policy and published their code in
Github repositories. This policy allows repeat-ability and reproducibility of the
experiments.
8</p>
    </sec>
    <sec id="sec-8">
      <title>Performance Analysis</title>
      <p>Some of the participants have conducted an interesting analysis in order to
explore the behavior of their systems. We tried to explore the performance of all
systems on each tweet. We ranked the tweets for sub-task A in English based
on the number of systems that classi ed them. The following gure shows the
distribution of the values.</p>
      <p>
        We can observe that only 30% of the systems agree a post is an o ensive (class
HOF) considering the Median. On the other hand, 70% of the systems vote for
NOT in the Median for the class NOT. However, the distributions are quite
scattered. This shows that for the systems there seem to be no clear and obvious
cases. Considering the analysis of Salminen et al. in which humans agreed much
on obvious Hate Speech tweets [
        <xref ref-type="bibr" rid="ref35">35</xref>
        ], there seems to be less agreement by systems.
As a consequence, voting approaches might not work well. Another consequence
could be that it is hard to explain and understand the decision of a classi er in
this domain. This may lead to a lack of ability to explain decisions and a lack
of transparency. This can result in a low degree of acceptance in society. More
analysis of the results is necessary for the future.
9
      </p>
    </sec>
    <sec id="sec-9">
      <title>Conclusion and Outlook</title>
      <p>
        The submissions for HASOC have shown that deep learning representations
seem to be the state of the art approach for Hate Speech classi cation. After
analyzing the results, the best method for Hate speech classi cation is dependent
on the corpus language, classi cation granularities, and distribution of each
classlabels. In other words balance, an unbalanced training dataset might a ect the
performance of the classi cation system. In the long run, the HASOC track aims
at supporting researchers to develop robust technology which can cope with
multilingual data and to develop transfer learning approaches that can exploit
learning data across languages. For future editions, we envision the integration of
further languages. The potential bias in the data collection needs to be analyzed
and monitored [
        <xref ref-type="bibr" rid="ref43">43</xref>
        ].
10
      </p>
    </sec>
    <sec id="sec-10">
      <title>Acknowledgements</title>
      <p>We thank all participants for their submissions and the work involved. We thank
all the jurors who labeled the tweets in a short period of time. We also thank
the FIRE organizers for their support in organizing the track.
A
A.1</p>
      <sec id="sec-10-1">
        <title>Annotation Guidelines for HASOC 2019</title>
        <p>HATE SPEECH Ascribing negative attributes or de ciencies to groups of
individuals because they are members of a group (e.g. all poor people are stupid).
Hateful comment toward groups because of race, political opinion, sexual
orientation, gender, social status, health condition or similar.</p>
        <p>OFFENSIVE Degrading, dehumanizing or insulting an individual. Threatening
with violent acts.</p>
        <p>PROFANITY Unacceptable language in the absence of insults and abuse. This
typically concerns the usage of swearwords (Schei e, Fuck etc.) and cursing (Zur
Holle! Verdammt! etc.).</p>
        <p>OTHER Normal content, statements, or anything else. If the utterances are
considered to be \normal" and not o ending to anyone, they should not be
labeled. This could be part of youth language or other language registers.</p>
        <p>We expect most posts to be OTHER, some to be HATE and the other two
categories to be less frequent.</p>
        <p>Dubious cases which are di cult to decide even for humans, should be left
out.
B
B.1</p>
      </sec>
      <sec id="sec-10-2">
        <title>Systems and Approaches at HASOC 2019</title>
        <p>The following tables summarize the approaches used by the teams. The last
col-umn has an entry when the team compared several approaches and clearly
identi- ed a best one. The rst table shows the approaches which used technology
without Deep Learning or for which a traditional approaches performed best.
The second table shows the approaches which used Deep Learning.</p>
        <p>Text Representa- Best
tion and Classi er (when</p>
        <p>cable)
CNN, fastText fastText,</p>
        <p>Hot</p>
        <p>
          run
appliLSTM
Amrita [
          <xref ref-type="bibr" rid="ref37">37</xref>
          ] Amrita Vishwa CNN, LSTM
fast
        </p>
        <p>
          Vidyapeetham Text
LGI2P [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ] Univ Montpellier fastText
CIT Kokrajhar [
          <xref ref-type="bibr" rid="ref33">33</xref>
          ] University of Ed- LSTM
monton &amp; CIT
Kokrajhar
        </p>
        <p>Yunnan Univ.</p>
        <p>Stacked</p>
        <p>
          CNN
RALIGRAPH [
          <xref ref-type="bibr" rid="ref19">19</xref>
          ] Univ. of Montral BERT, Graph VCGN-BERT
CNNPre-trained
with external Founta
corpus
3Idiots [
          <xref ref-type="bibr" rid="ref21">21</xref>
          ] Univ. of Illinois &amp; BERT, All 3 tasks in BERT cased
        </p>
        <p>
          IIT Kanpur one
BRUMS [
          <xref ref-type="bibr" rid="ref29">29</xref>
          ] Univ of Wolver- Several deep learn- BERT
hampton, ing architectures
Rochester In- including LSTM,
stitute of Techn. &amp; GRU, Attention, 2D
Birmingham City Convolution,light
        </p>
        <p>
          Univ pre-processing
KMI-Panlingua [
          <xref ref-type="bibr" rid="ref31">31</xref>
          ] Bhimrao Ambed- BERT, Char and BERT
kar Univ. &amp; Pan- word ngrams + SVM
lingua &amp; Charles
        </p>
        <p>
          Univ. Prague
LSV-UdS [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ] Saarland Univer- BERT, BERT
sity SVM,External
collections
        </p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Kaggle</surname>
          </string-name>
          (
          <year>2017</year>
          ):
          <article-title>Toxic comment classi cation challenge: Identify and classify toxic online comments</article-title>
          , https://www.kaggle.com/c/jigsaw-toxic
          <article-title>-comment-classi cationchallenge</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Anita</surname>
            <given-names>Saroj</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>R.K.M.</given-names>
            ,
            <surname>Pal</surname>
          </string-name>
          ,
          <string-name>
            <surname>S.</surname>
          </string-name>
          : Irlab@iitbhu at hasoc
          <year>2019 2019</year>
          <article-title>:traditional machine learning for hate speech and o ensive content identi cation</article-title>
          .
          <source>In: Proceedings of the 11th annual meeting of the Forum for Information Retrieval Evaluation (December</source>
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Baruah</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Barbhuiya</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dey</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          :
          <article-title>IIITG-ADBU at HASOC 2019: Automated Hate Speech and O ensive Content Detection in English and Code-Mixed Hindi Text</article-title>
          . In:
          <article-title>Proceedings of the 11th annual meeting of the Forum for Information Retrieval Evaluation (December</article-title>
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Bashar</surname>
            ,
            <given-names>M.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nayak</surname>
          </string-name>
          , R.: QutNocturnalHASOC'19:
          <article-title>CNN for Hate Speech and Offensive Content Identi cation in Hindi Language</article-title>
          . In:
          <article-title>Proceedings of the 11th annual meeting of the Forum for Information Retrieval Evaluation (December</article-title>
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Basile</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bosco</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fersini</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nozza</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Patti</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pardo</surname>
            ,
            <given-names>F.M.R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rosso</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sanguinetti</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Semeval-2019 task 5: Multilingual detection of hate speech against immigrants and women in twitter</article-title>
          .
          <source>In: Proceedings of the 13th International Workshop on Semantic Evaluation</source>
          . pp.
          <volume>54</volume>
          {
          <issue>63</issue>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Bin</surname>
            <given-names>Wang</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yunxia Ding</surname>
            ,
            <given-names>S.L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhou</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          : YNU Wb at HASOC 2019:
          <article-title>Ordered Neurons LSTM with Attention for Identifying Hate Speech and O ensive Language</article-title>
          . In:
          <article-title>Proceedings of the 11th annual meeting of the Forum for Information Retrieval Evaluation (December</article-title>
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Bosco</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Felice</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Poletto</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sanguinetti</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Maurizio</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>Overview of the evalita 2018 hate speech detection task</article-title>
          .
          <source>In: EVALITA 2018-Sixth Evaluation Campaign of Natural Language Processing and Speech Tools for Italian</source>
          . vol.
          <volume>2263</volume>
          , pp.
          <volume>1</volume>
          {
          <issue>9</issue>
          .
          <string-name>
            <surname>CEUR</surname>
          </string-name>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Casavantes</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lopez</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gonzalez</surname>
            ,
            <given-names>L.C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Montes-y Gomez</surname>
            ,
            <given-names>M.:</given-names>
          </string-name>
          <article-title>UACh-INAOE at HASOC 2019: Detecting Aggressive Tweets by Incorporating Authors' Traits as Descriptors</article-title>
          .
          <source>In: Proceedings of the 11th annual meeting of the Forum for Information Retrieval Evaluation (December</source>
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Chung</surname>
            ,
            <given-names>Y.L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kuzmenko</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tekiroglu</surname>
            ,
            <given-names>S.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Guerini</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Conan{counter narratives through nichesourcing: a multilingual dataset of responses to ght online hate speech</article-title>
          . arXiv preprint arXiv:
          <year>1910</year>
          .
          <volume>03270</volume>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <given-names>Dana</given-names>
            <surname>Ruiter</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.A.R.</given-names>
            ,
            <surname>Klakow</surname>
          </string-name>
          ,
          <string-name>
            <surname>D.</surname>
          </string-name>
          : LSVUdS at HASOC 2019:
          <article-title>The Problem of De ning Hate? In: Proceedings of the 11th annual meeting of the Forum for Information Retrieval Evaluation (December</article-title>
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Davidson</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Warmsley</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Macy</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Weber</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          :
          <article-title>Automated Hate Speech Detection and the Problem of O ensive Language</article-title>
          .
          <source>In: Proceedings of ICWSM</source>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Fortuna</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nunes</surname>
            ,
            <given-names>S.:</given-names>
          </string-name>
          <article-title>A Survey on Automatic Detection of Hate Speech in Text</article-title>
          .
          <source>ACM Computing Surveys (CSUR) 51(4)</source>
          ,
          <volume>85</volume>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Jean-Christophe</surname>
            <given-names>Mensonides</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pierre-Antoine Jean</surname>
            ,
            <given-names>A.T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Harispe</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          : IMT Mines Ales at HASOC 2019:
          <article-title>Automatic Hate Speech detection</article-title>
          .
          <source>In: Proceedings of the 11th annual meeting of the Forum for Information Retrieval Evaluation (December</source>
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Jhaver</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ghoshal</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bruckman</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gilbert</surname>
          </string-name>
          , E.:
          <article-title>Online harassment and content moderation: The case of blocklists. ACM Transactions on Computer-Human Interaction (TOCHI) 25(2</article-title>
          ),
          <volume>12</volume>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Jiang</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>QMUL-NLP at HASOC 2019: O ensive Content Detection and Classication in Social Media</article-title>
          . In:
          <article-title>Proceedings of the 11th annual meeting of the Forum for Information Retrieval Evaluation (December</article-title>
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>Kaushik Amar Das</surname>
            ,
            <given-names>F.A.B.</given-names>
          </string-name>
          : FalsePostive at HASOC 2019:
          <article-title>Transfer-Learning for Detection and Classi cation of Hate Speech</article-title>
          . In:
          <article-title>Proceedings of the 11th annual meeting of the Forum for Information Retrieval Evaluation (December</article-title>
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <given-names>Kirti</given-names>
            <surname>Kumari</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.P.S.</given-names>
            :
            <surname>AI ML NIT</surname>
          </string-name>
          <article-title>Patna at HASOC 2019: Deep Learning Approach for Identi cation of Abusive Content</article-title>
          . In:
          <article-title>Proceedings of the 11th annual meeting of the Forum for Information Retrieval Evaluation (December</article-title>
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          18.
          <string-name>
            <surname>Kwok</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          :
          <article-title>Locate the hate: Detecting Tweets Against Blacks</article-title>
          .
          <source>In: TwentySeventh AAAI Conference on Arti cial Intelligence</source>
          (
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          19.
          <string-name>
            <surname>Lu</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nie</surname>
          </string-name>
          , J.Y.: RALIGRAPH at HASOC 2019:
          <article-title>VGCN-BERT: Augmenting BERT with Graph Embedding for O ensive Language Detection</article-title>
          . In:
          <article-title>Proceedings of the 11th annual meeting of the Forum for Information Retrieval Evaluation (December</article-title>
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          20.
          <string-name>
            <surname>Mishra</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pal</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          : IIT Varanasi at HASOC 2019 :
          <article-title>Hate Speech and O ensive Content Identi cation in Indo-European Languages</article-title>
          . In:
          <article-title>Proceedings of the 11th annual meeting of the Forum for Information Retrieval Evaluation (December</article-title>
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          21.
          <string-name>
            <surname>Mishra</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mishra</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          : 3Idiots at HASOC 2019:
          <article-title>Fine-tuning Transformer Neural Networks for Hate Speech Identi cation in Indo-European Languages</article-title>
          . In:
          <article-title>Proceedings of the 11th annual meeting of the Forum for Information Retrieval Evaluation (December</article-title>
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          22.
          <string-name>
            <surname>Mubarak</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Darwish</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Magdy</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          :
          <article-title>Abusive language detection on arabic social media</article-title>
          .
          <source>In: Proceedings of the First Workshop on Abusive Language Online</source>
          . pp.
          <volume>52</volume>
          {
          <issue>56</issue>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          23.
          <string-name>
            <surname>Mujadia</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mishra</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sharma</surname>
            ,
            <given-names>D.M.</given-names>
          </string-name>
          :
          <string-name>
            <surname>IIIT-Hyderabad</surname>
            <given-names>at HASOC</given-names>
          </string-name>
          2019:
          <article-title>Hate Speech Detection</article-title>
          . In:
          <article-title>Proceedings of the 11th annual meeting of the Forum for Information Retrieval Evaluation (December</article-title>
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          24.
          <string-name>
            <surname>Nayel</surname>
            ,
            <given-names>H.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shashirekha</surname>
            ,
            <given-names>H.L.</given-names>
          </string-name>
          :
          <article-title>DEEP at HASOC2019 : A Machine Learning Framework for Hate Speech and O ensive Language Detection</article-title>
          . In:
          <article-title>Proceedings of the 11th annual meeting of the Forum for Information Retrieval Evaluation (December</article-title>
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          25.
          <string-name>
            <surname>Nina-Alcocer</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          : Vito at HASOC 2019:
          <article-title>Detecting Hate Speech and O ensive Content through Ensembles</article-title>
          .
          <source>In: Proceedings of the 11th annual meeting of the Forum for Information Retrieval Evaluation (December</source>
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          26.
          <string-name>
            <surname>Parikh</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Desai</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bisht</surname>
            ,
            <given-names>A.S.</given-names>
          </string-name>
          : DA Master at HASOC 2019:
          <article-title>Identi cation of Hate Speech using Machine Learning and Deep Learning approaches for social media post</article-title>
          .
          <source>In: Proceedings of the 11th annual meeting of the Forum for Information Retrieval Evaluation (December</source>
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          27.
          <string-name>
            <surname>Pedro</surname>
            <given-names>Alonso</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>R.S.</given-names>
            ,
            <surname>Kovacs</surname>
          </string-name>
          , G.: TheNorth at HASOC 2019:
          <article-title>Hate Speech Detection in Social Media Data</article-title>
          . In:
          <article-title>Proceedings of the 11th annual meeting of the Forum for Information Retrieval Evaluation (December</article-title>
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          28.
          <string-name>
            <given-names>R.</given-names>
            <surname>Rajalakshmi</surname>
          </string-name>
          ,
          <string-name>
            <surname>Y.R.</surname>
          </string-name>
          :
          <source>DLRG@HASOC</source>
          <year>2019</year>
          :
          <article-title>An Enhanced Ensemble Classi er for Hate and O ensive Content Identi cation</article-title>
          .
          <source>In: Proceedings of the 11th annual meeting of the Forum for Information Retrieval Evaluation (December</source>
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          29.
          <string-name>
            <surname>Ranasinghe</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zampieri</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hettiarachchi</surname>
          </string-name>
          , H.: BRUMS at HASOC 2019:
          <article-title>Deep Learning Models for Multilingual Hate Speech and O ensive Language Identi - cation</article-title>
          . In:
          <article-title>Proceedings of the 11th annual meeting of the Forum for Information Retrieval Evaluation (December</article-title>
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          30.
          <string-name>
            <surname>Ritesh</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>N.</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.A.</given-names>
            ,
            <surname>Akshit</surname>
          </string-name>
          ,
          <string-name>
            <surname>B.</surname>
          </string-name>
          ,
          <article-title>MaheshwariTushar: Aggression-annotated corpus of hindi-english code-mixed data</article-title>
          .
          <source>In: Proceedings of the 11th Language Resources and Evaluation Conference (LREC)</source>
          . pp.
          <volume>1</volume>
          {
          <fpage>11</fpage>
          .
          <string-name>
            <surname>Miyazaki</surname>
          </string-name>
          ,
          <string-name>
            <surname>Japan</surname>
          </string-name>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          31.
          <string-name>
            <given-names>Ritesh</given-names>
            <surname>Kumar</surname>
          </string-name>
          ,
          <string-name>
            <surname>A.K.O.:</surname>
          </string-name>
          KMI-Panlingua
          <source>at HASOC</source>
          <year>2019</year>
          :
          <article-title>SVM vs BERT for Hate Speech and O ensive Content Detection</article-title>
          . In:
          <article-title>Proceedings of the 11th annual meeting of the Forum for Information Retrieval Evaluation (December</article-title>
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          32.
          <string-name>
            <surname>Ross</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rist</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          , Carbonell, G.,
          <string-name>
            <surname>Cabrera</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kurowsky</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wojatzki</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Measuring the Reliability of Hate Speech Annotations: The Case of the European Refugee Crisis</article-title>
          .
          <source>In: Proceedings of the Workshop on Natural Language Processing for Computer-Mediated Communication (NLP4CMC)</source>
          . Bochum,
          <string-name>
            <surname>Germany</surname>
          </string-name>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          33.
          <string-name>
            <surname>Saha</surname>
            ,
            <given-names>B.N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Senapati</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>CIT Kokrajhar Team: LSTM based Deep RNN Architecture for Hate Speech and O ensive Content (HASOC) Identi cation in IndoEuropean Languages</article-title>
          . In:
          <article-title>Proceedings of the 11th annual meeting of the Forum for Information Retrieval Evaluation (December</article-title>
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref34">
        <mixed-citation>
          34.
          <string-name>
            <surname>Saha</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mathew</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Goyal</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mukherjee</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          : HateMonitors at HASOC 2019:
          <article-title>Language Agnostic Online Abuse Detection</article-title>
          . In:
          <article-title>Proceedings of the 11th annual meeting of the Forum for Information Retrieval Evaluation (December</article-title>
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref35">
        <mixed-citation>
          35.
          <string-name>
            <surname>Salminen</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Almerekhi</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kamel</surname>
            ,
            <given-names>A.M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jung</surname>
            ,
            <given-names>S.g.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jansen</surname>
            ,
            <given-names>B.J.:</given-names>
          </string-name>
          <article-title>Online hate ratings vary by extremes: A statistical analysis</article-title>
          .
          <source>In: Proceedings of the 2019 Conference on Human Information Interaction and Retrieval</source>
          . pp.
          <volume>213</volume>
          {
          <fpage>217</fpage>
          .
          <string-name>
            <surname>ACM</surname>
          </string-name>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref36">
        <mixed-citation>
          36.
          <string-name>
            <surname>Schmidt</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wiegand</surname>
            ,
            <given-names>M.:</given-names>
          </string-name>
          <article-title>A Survey on Hate Speech Detection Using Natural Language Processing</article-title>
          .
          <source>In: Proceedings of the Fifth International Workshop on Natural Language Processing for Social Media. Association for Computational Linguistics</source>
          . pp.
          <volume>1</volume>
          {
          <fpage>10</fpage>
          .
          <string-name>
            <surname>Valencia</surname>
          </string-name>
          ,
          <string-name>
            <surname>Spain</surname>
          </string-name>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref37">
        <mixed-citation>
          37.
          <string-name>
            <surname>Sreelakshmi</surname>
          </string-name>
          .K, P., K.P, S.: AmritaCEN at HASOC 2019:
          <article-title>Hate SpeechDetection in Roman and Devanagiri Scripted Text</article-title>
          . In:
          <article-title>Proceedings of the 11th annual meeting of the Forum for Information Retrieval Evaluation (December</article-title>
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref38">
        <mixed-citation>
          38.
          <string-name>
            <surname>Stru</surname>
            ,
            <given-names>J.M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Siegel</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ruppenhofer</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wiegand</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Klenner</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Overview of germeval task 2, 2019 shared task on the identi cation of o ensive language (</article-title>
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref39">
        <mixed-citation>
          39.
          <string-name>
            <surname>Tulkens</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hilte</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lodewyckx</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Verhoeven</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Daelemans</surname>
            ,
            <given-names>W.:</given-names>
          </string-name>
          <article-title>The automated detection of racist discourse in dutch social media</article-title>
          .
          <source>Computational Linguistics in the Netherlands Journal</source>
          <volume>6</volume>
          ,
          <issue>3</issue>
          {
          <fpage>20</fpage>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref40">
        <mixed-citation>
          40.
          <string-name>
            <surname>Tulkens</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hilte</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lodewyckx</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Verhoeven</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Daelemans</surname>
            ,
            <given-names>W.:</given-names>
          </string-name>
          <article-title>A dictionarybased approach to racism detection in dutch social media</article-title>
          .
          <source>arXiv preprint arXiv:1608.08738</source>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref41">
        <mixed-citation>
          41.
          <string-name>
            <given-names>Urmi</given-names>
            <surname>Saha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.D.</given-names>
            ,
            <surname>Bhattacharyya</surname>
          </string-name>
          ,
          <string-name>
            <surname>P.</surname>
          </string-name>
          : IIT Bombay at HASOC 2019:
          <article-title>Supervised Hate Speech and O ensive Content Detection in Indo-European Languages</article-title>
          . In:
          <article-title>Proceedings of the 11th annual meeting of the Forum for Information Retrieval Evaluation (December</article-title>
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref42">
        <mixed-citation>
          42.
          <string-name>
            <surname>Wagner</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bumann</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>Challenges in annotating a corpus for automatic hate speech detection</article-title>
          . In: BOBCATSSS Paris, January (
          <year>2020</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref43">
        <mixed-citation>
          43.
          <string-name>
            <surname>Wiegand</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ruppenhofer</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kleinbauer</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>Detection of abusive language: the problem of biased datasets</article-title>
          .
          <source>In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</source>
          , Volume
          <volume>1</volume>
          (Long and Short Papers). pp.
          <volume>602</volume>
          {
          <issue>608</issue>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref44">
        <mixed-citation>
          44.
          <string-name>
            <surname>Wiegand</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Siegel</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ruppenhofer</surname>
          </string-name>
          , J.:
          <article-title>Overview of the germeval 2018 shared task on the identi cation of o ensive language (</article-title>
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref45">
        <mixed-citation>
          45.
          <string-name>
            <surname>Zampieri</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Malmasi</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nakov</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rosenthal</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Farra</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kumar</surname>
          </string-name>
          , R.:
          <article-title>Predicting the Type and Target of O ensive Posts in Social Media</article-title>
          .
          <source>In: Proceedings of NAACL</source>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref46">
        <mixed-citation>
          46.
          <string-name>
            <surname>Zampieri</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Malmasi</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nakov</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rosenthal</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Farra</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kumar</surname>
          </string-name>
          , R.:
          <article-title>Predicting the type and target of o ensive posts in social media</article-title>
          . arXiv preprint arXiv:
          <year>1902</year>
          .
          <volume>09666</volume>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref47">
        <mixed-citation>
          47.
          <string-name>
            <surname>Zampieri</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Malmasi</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nakov</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rosenthal</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Farra</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kumar</surname>
          </string-name>
          , R.:
          <article-title>Semeval-2019 task 6: Identifying and categorizing o ensive language in social media (o enseval)</article-title>
          .
          <source>arXiv preprint arXiv:1903</source>
          .
          <volume>08983</volume>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>