<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Leveraging Bias in Pre-Trained Word Embeddings for Unsupervised Microaggression Detection</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Tol u´l o.p e´. O`g u´nre`.m´ı</string-name>
          <email>tolulope@stanford.edu</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Nazanin Sabri</string-name>
          <email>nazanin.sabrii@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Valerio Basile</string-name>
          <email>valerio.basile@unito.it</email>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Tommaso Caselli</string-name>
          <email>t.caselli@rug.nl</email>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>. Independent Researcher</institution>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>. Stanford University</institution>
          ,
          <country country="US">United States</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>. University of Groningen</institution>
          ,
          <country country="NL">Netherlands</country>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>. University of Turin</institution>
          ,
          <country country="IT">Italy</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Microaggressions are subtle manifestations of bias (Breitfeller et al., 2019). These demonstrations of bias can often be classified as a subset of abusive language. However, not as much focus has been placed on the recognition of these instances. As a result, limited data is available on the topic, and only in English. Being able to detect microaggressions without the need for labeled data would be advantageous since it would allow content moderation also for languages lacking annotated data. In this study, we introduce an unsupervised method to detect microaggressions in natural language expressions. The algorithm relies on pre-trained wordembeddings, leveraging the bias encoded in the model in order to detect microaggressions in unseen textual instances. We test the method on a dataset of racial and gender-based microaggressions, reporting promising results. We further run the algorithm on out-of-domain unseen data with the purpose of bootstrapping corpora of microaggressions “in the wild”, and discuss the benefits and drawbacks of our proposed method.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1 Introduction</title>
      <p>The growth of Social Media platforms has been
accompanied by an increased visibility of
expressions of socially unacceptable language online. In
a 2016 Eurobarometer survey, 75% of people who
follow or participate in online discussions have
witnessed or experienced abuse or hate speech.
With this umbrella term, different phenomena can</p>
      <p>
        Copyright © 2021 for this paper by its authors. Use
permitted under Creative Commons License Attribution 4.0
International (CC BY 4.0).
be identified ranging from offensive language to
more complex and dangerous ones, such as hate
speech or doxing. Recently, there has been a
growing interest by the Natural Language Processing
community in the development of language
resources and systems to counteract socially
unacceptable language online. Most previous work has
focused on few, easy to model phenomena,
ignoring more subtle and complex ones, such as
microaggressions
        <xref ref-type="bibr" rid="ref10 ref5">(Jurgens et al., 2019)</xref>
        .
      </p>
      <p>
        Microaggressions are brief, everyday
exchanges that denigrate stigmatised and culturally
marginalised groups
        <xref ref-type="bibr" rid="ref15">(Merriam-Webster, 2021)</xref>
        .
They are not always perceived as hurtful by
either party, and they can often be detected as
positive statements by current hate-speech detection
systems
        <xref ref-type="bibr" rid="ref5">(Breitfeller et al., 2019)</xref>
        . The
occasionally unintentional hurt caused by such comments
is a reflection of how certain stereotypes of
others are baked into society. Sue et al. (2007)
deifne microaggressions in the racial context,
particularly when directed toward people of color, as
“brief and commonplace daily verbal, behavioral,
or environmental indignities”, such as: “you are a
credit to your race.” (intended message: it is
unusual for someone of your race to be intelligent)
or “do you think you’re ready for college?”
(indented message: it is unusual for people of color to
succeed). The need for moderation of hateful
content has previously been explored. For instance,
Mathew et al. (2019b) analyses the temporal
effects of allowing hate speech on Gab, and finds
that the language of users tends to become more
and more similar to that of hateful users over time.
Mathew et al. (2019a) further highlights that the
spreading speed and reach of hateful content is
much higher than with the non-hateful content. As
a result, being able to remove instances of
hateful language, such as microaggressions, is of great
importance.
      </p>
      <p>Previous work on microaggressions with
computational methods is quite recent. Breitfeller et
al. (2019) is one of the first work to address
microaggressions in a systematic way, also
introducing a first dataset, SelfMA. A further
contribution specifically focused on racial microaggression
is Ali et al. (2020), where the authors focus on the
development of machine learning systems.</p>
      <p>In this study we introduce an unsupervised
method for microaggression detection. Our
method utilizes the existing bias in
wordembeddings to detect words with biased
connotations in the message. Although unsupervised
approaches tend to be less competitive than their
supervised counterparts, our method is
languageindependent and thus it can be applied to any
language for which embedding representations exist.
Furthermore, the reliance of our methods on
specific lexical items and their context of occurrence
makes transparent the flagging of a message as an
instance of a microaggression. In addition to the
usefulness of our method in languages with no
labeled data, the reliance of our model on words in
the sentences would make it interpretable as it
allow human moderators to understand what the
system has based its decision on.</p>
      <p>Our contributions can be summarised as
follows:
• we introduce a new unsupervised method
for the detection of microaggressions which
builds on top of pre-trained word
embeddings;
• we compare the performance of our model
using different pre-trained word embeddings
(Glove, FastText, and Word2Vec) and discuss
the potential reasons behind the differences;
• we test the proposed algorithm on unseen
data from a different domain (i.e., Twitter),
in order to qualitatively evaluate its efficacy
in discovering new instances of
microaggression.</p>
      <p>The rest of this paper is structured as follows:
we introduce our method in Section 2. The data
and our results are reported in Section 3. We
deploy our model and discuss its limitations in
Section 4. Finally, we present the conclusion and
future work in Section 5.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Use the Bias Against the Bias</title>
      <p>
        Embedded representations, either from pre-trained
word embeddings or pre-trained language models,
have been shown to contain and amplify the biases
present in the data used to generate them
        <xref ref-type="bibr" rid="ref10 ref11 ref13 ref3 ref4 ref5">(Bolukbasi et al., 2016; Lauscher and Glavasˇ, 2019;
Bhardwaj et al., 2020)</xref>
        . As such, they often
exhibit gender and racial bias
        <xref ref-type="bibr" rid="ref20">(Swinger et al.,
2019)</xref>
        . Many studies have attempted to reduce
this bias
        <xref ref-type="bibr" rid="ref12 ref21 ref22 ref3">(Yang and Feng, 2020; Zhao et al., 2018;
Manzini et al., 2019)</xref>
        . In this work, we take a
different turn by using this bias to our advantage:
rather than taming the hurtfulness of the
representations
        <xref ref-type="bibr" rid="ref18">(Schick et al., 2021)</xref>
        , we actively use
it to promote social good. In this first study, we
employ word representations derived from generic
textual corpora of English, in order to capture the
background knowledge needed to disambiguate
instances of microaggressions in the text.
Recently, however, there have been studies involving
word representations created from tailored
collections of social media content aimed at capturing
abusive phenomena like verbal aggression
        <xref ref-type="bibr" rid="ref7">(Dynel,
2021)</xref>
        and hate speech
        <xref ref-type="bibr" rid="ref6">(Caselli et al., 2020)</xref>
        .
      </p>
      <p>We devise a simple and effective method that
exploits existing bias in word embeddings and
identify words in a message that are related to
particular and distant semantic areas in the
embedding space. Messages are analysed in three
steps: first, for each token ti we compute its
relatedness to a list of manually curated seed words
s = s1, ..., sn denoting potential targets of
microaggressions; second, we consider only the
similarities of the pairs (ti, sj ) above an empirical
similarity threshold ST and compute their
variance vi; finally, we classify the token ti as a micro
aggression trigger, and consequently the message
as a micro aggression, if the vi is above an
empirically determined variance threshold V T .</p>
      <p>The intuitive idea behind this algorithm is that
some lexical elements in a verbal microaggression
are often (yet sometimes subtly) hinting at specific
features of the recipient of the message, in an
otherwise neutral lexical context.</p>
      <p>In this work, we choose to focus on
microaggressions related to race and gender, therefore the
seed words have to be chosen accordingly. The
seed word lists for race and gender are,
respectively, [white, black, asian, latino, hispanic, arab,
african, caucasian] and [girl, boy, man, woman,
male, female] for gender. There is also a
practical reasons to focus on gender and race, namely
the scarcity of data available for other categories
of microaggression and other idiosincrasies of the
available datasets — the religion class was
specific to different religions, therefore hard to
generalise, sexuality and gender presented a large
overlap, and so on.</p>
      <p>An example of how the proposed method works
is illustrated in Figure 1. In the example,
consider the word ”chopsticks” in the message ”Ford:
Built With Tools, Not With Chopsticks” (from the
SelfMA dataset, described in Section 3). The
target word exhibits a much higher relatedness to
the word asian (0.237) than any other seed words.
Even just considering the seed words with a
similarity above a fixed threshold ( white, asian and,
african), the variance of their similarity score with
respect to chopsticks is still higher than the
variance threshold, and therefore this target word, in
this context, triggers a microaggression
according to the algorithm. This process is repeated for
all the words in the message in order to detect
microaggressions. Some categories of words are
bound to exhibit a high relatedness to all the seed
words, e.g., “people” or “human”. This is the
reason to introduce the variance threshold in the
final step of our algorithm, to filter out these cases
when classifying a given message, and instead
focus on words that are related to different races (or
genders) unevenly, with a skewed distribution of
similarity scores.</p>
      <p>An important by-product of this algorithm is
that the output is one or more trigger words, in
addition to the microaggression label — in the
example, the trigger word is indeed chopsticks —
therefore enabling a more informative and interpretable
decision process.</p>
      <sec id="sec-2-1">
        <title>Source</title>
        <p>SelfMA Gender
SelfMA Racial
Tumblr</p>
      </sec>
      <sec id="sec-2-2">
        <title>Number of posts</title>
        <p>
          1,314
1,278
2,021
To test our method, we use two subsets of the
SelfMA: microaggressions.com dataset
          <xref ref-type="bibr" rid="ref5">(Breitfeller
et al., 2019)</xref>
          , comprised of 1,314 and 1,278 Tumblr
posts respectively1. The posts in SelfMA are all
instances of microaggressions, manually tagged
with one of four categories: race, gender,
sexuality and religion. These posts can be tagged with
more than one form of microaggressions,
meaning certain instances can appear in both subsets
of race and gender used for the purposes of this
study. The dataset consists of first and second
hand accounts of microaggressions, as well as
direct quotes of phrases or sentences said to the
person posting. In order to reduce linguistic
perturbation introduced by accounts of a situation, we
only take direct quotes found in the dataset as
instances of microaggressions that we can detect
with our unsupervised method. For training, we
pull out direct quotes from the gender (561) and
racial (519) dataset to test the algorithm. In order
to balance the dataset, we scraped 2,021 random
Tumblr posts, for a total of 4,612 instances.
Table 1 summarises the composition of our dataset.
        </p>
        <p>It is important to note that a microaggression
can have multiple tags, so there is an overlap of
1Tumblr is a popular American microblogging platform
https://www.tumblr.com
instances. However, the seed words used to detect
microaggression types in the method are different
for each target phenomenon (e.g., race, gender).</p>
        <p>
          We ran the algorithm on the SelfMA dataset,
empirically optimising the two thresholds on the
training split, for each word embedding type and
each microaggression category, filtering by the
seed words listed in Section 2. We test the
algorithm with three pre-trained word embedding
models for English, namely FastText
          <xref ref-type="bibr" rid="ref9">(Joulin
et al., 2016)</xref>
          (trained on Wikipedia and
Common Crawl), word2vec
          <xref ref-type="bibr" rid="ref16">(Mikolov et al., 2013)</xref>
          (trained on Google News), and GloVe
          <xref ref-type="bibr" rid="ref17">(Pennington et al., 2014)</xref>
          (trained on Wikipedia, GigaWord
corpus, and Common Crawl). The optimization is
performed by exhaustive grid search over the
hyperparamter space.
        </p>
        <p>The results, shown in Table 2, indicate that
FastText has a better F1 score on Racial
microaggressions while word2vec performs
better on Gender microaggressions. The
difference in performance between FastText and
word2vec is not major, and we attribute this
to the difference between the corpora on which
the two models were trained (i.e., web crawl
and Wikipedia for FastText vs. news data
for word2vec). The GloVe pretrained model,
trained on a combination of newswire texts,
encyclopedic entries and texts from the Web,
underperforms in both experiments. In general, the
absolute figures are encouraging, especially
considering the simplicity of this unsupervised approach.
4</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Discovering Microaggressions</title>
      <p>To better understand the performance of our
unsupervised model, we performed an additional
experiment. Our goal is to understand the false
positive results and the potential harm the model could
cause. To do so, we use our unsupervised model to
label unseen instances from another domain
(Twitter) than the SelfMA dataset (Tumblr) in order to
see how the model would perform in detecting
microaggressions.</p>
      <p>We begin by performing keyword searches on
Twitter (using Twitter’s official API) and collect
a new dataset of of 3M tweets with seven
keywords potentially containing race and gender
expressions.Next, we set the threshold values ST
and V T in our model in order to obtain the highest
Precision scores, rather than the highest F1 value.
This step is performed exactly like the
optimization described in Section 2 with the only difference
of the target metric. The aim of this step is to only
label tweets as microaggressions with the highest
possible degree of confidence. We set ST = 0.12
and V T = 0.014 for racial microaggressions
leading to Precision of .931 and ST = 0.13 and
V T = 0.019 for gender-based microaggressions
leading to a Precision of .912. Precision has been
measured on the original SelfMA dataset used as
a validation set.</p>
      <p>We then run the unsupervised model on the new
Twitter dataset by automatically labelling 256,843
tweets for gender and 373,631 tweets for race.
After the data is labeled, we manually explore the
positive instances in order to evaluate the
performance of the model. The algorithm tuned for
high precision found in this dataset 6,306
genderrelated microaggression candidates, 13,004
racerelated microaggression candidates.</p>
      <p>We find that while the model does detect actual
instances of microaggression, there is a
noticeable amount of false positive instances. These
tweets discuss race or gender in some manner.
However, they do not necessarily contain
microaggressions towards these groups. While the
model does learn to detect discussions of these
topics, it seems to sometimes confuse these
discussions with microaggressions towards the
aforementioned groups. Some examples follow,
paraphrased to avoid tracking the original messages.</p>
      <p>Saying ”Arrested Development isn’t
funny” in an office full of women just to
feel something
“Men have moustaches, women have
oversized bracelets”</p>
      <p>The humorous attempts in this tweets hinge on
gender stereotypes, and therefore in some contexts
it could be perceived as offensive by some
recipients. The high relatedness in the word
embedding space between some words (moustaches and
bracelets) and gender-related seed words (men and
women) triggers the detection algorithm.</p>
      <p>The automatic detection of racial
microaggressions “in the wild” is more challenging than
gender-based ones, according to our manual
exploration of this automatically labeled dataset.
This may be due to the difficulty of crafting a
list of seed words that is sufficiently race-related,
but at the same time avoids generating too many
false positives. We indeed found many of them,</p>
      <sec id="sec-3-1">
        <title>Target</title>
      </sec>
      <sec id="sec-3-2">
        <title>Model</title>
        <sec id="sec-3-2-1">
          <title>Gender</title>
          <p>GloVe
FastText</p>
        </sec>
        <sec id="sec-3-2-2">
          <title>Race</title>
          <p>GloVe
word2vec
FastText
word2vec</p>
        </sec>
      </sec>
      <sec id="sec-3-3">
        <title>Class</title>
        <p>not-MA
MA
macro avg.
not-MA
MA
macro avg.
not-MA
MA
macro avg.
not-MA
MA
macro avg.
not-MA
MA
macro avg.
not-MA
MA
macro avg.
.692
.603
mainly due to named entities and multi-word
expressions such as “White House”, or simply
because of the polysemy of color words, e.g. “black”
and “white”. We, however, still found instances
of messages containing different extent of racial
stereotyping.</p>
        <p>“why are you being so dramatic? just
say I’m not originally arab, you don’t
have to fight about it”
“I will need to explain that to the
chinese old lady who works at my school’s
administrative office”</p>
        <p>In summary, running the unsupervised
microaggression detection algorithm on unseen data seems
to represent a promising intermediate step towards
the semi-automatic creation of language resources
for this phenomenon. While the accuracy is not
ideal, and lists of seed words have to be
handcrafted carefully in order to avoid false positives,
these drawbacks are balanced by the fairly cheap
computational cost and the ease of application in a
multilingual scenario.
5</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Conclusion and Future Work</title>
      <p>In this paper we introduce a novel algorithm that
exploits the existing bias in pre-trained word
embeddings to detect subtly abusive language
phenomena such as microagressions. While
supervised methods of detection in the field of
natural language processing are plentiful, these
methods are only viable for languages and topics with
available labeled datasets. That is however not the
case for many languages. As a result, the
unsupervised method of detection introduced in this study
could help address the need for the moderation of
microaggressions in languages other than English.
This is further helped by the availability of
multilingual word-embeddings as they would allow the
method to be used in any of the languages
supported by the embedding.</p>
      <p>The method is unsupervised and only needs a
small list of seed words. Considering its
simplicity, the results obtained from an experiment on
a dataset of manually annotated microaggressions
are very promising. Further, the method is
transparent, explicitly identifying the words triggering
a microaggression, and thus paving the way for
explainable microaggression detection.</p>
      <p>Although the preliminary results are promising,
an experiment on unseen data from a different
domain shows that there is leeway for improvement.
Given that we are looking at the explicit words
used in each message, our method is not sensitive
to implicit expressions like “you people” or “your
kind”, often occurring in microaggressions. We
would have to add further steps to our algorithm
to catch expressions like these.</p>
      <p>
        Polysemy is another known issue, e.g., in words
like “black” and “white” whose relatedness to
certain identified trigger words could not necessarily
be due to race. While a careful composition of
the seed word lists helps to minimize this issue, a
systematic approach to polysemy would certainly
be desirable. The seed word list may also be
expanded, either manually or exploiting existing
lexicons such as HurtLex
        <xref ref-type="bibr" rid="ref2">(Bassignana et al., 2018)</xref>
        for offensive terms (including stereotypes for
several categories of individuals) or specialized lists
of identity-related terms2.
      </p>
      <p>
        In future work, we plan on improving our model
to account for lexical ambiguity, and the
complexity derived from the interference between
pragmatic phenomena and aggression, e.g., in
humorous and ironic messages, following the intuition
in recent literature
        <xref ref-type="bibr" rid="ref8">(Frenda, 2018)</xref>
        about the
interconnection between irony or sarcasm and abusive
language online. Our current plan is to apply the
algorithm presented in this paper to bootstrap the
creation of a multilingual resource of online
verbal microaggressions and release it to the research
community.
      </p>
    </sec>
    <sec id="sec-5">
      <title>Acknowledgements</title>
      <p>This work of Valerio Basile is partially funded
by the project “Be Positive!” (under the 2019
“Google.org Impact Challenge on Safety” call).
2See for instance this compendium of LGBTQIA+
terminology: https://www.umass.edu/stonewall/si
tes/default/files/documents/allyship ter
m handout.pdf</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <string-name>
            <given-names>Omar</given-names>
            <surname>Ali</surname>
          </string-name>
          , Nancy Scheidt, Alexander Gegov, Ella Haig, Mo Adda, and
          <string-name>
            <given-names>Benjamin</given-names>
            <surname>Aziz</surname>
          </string-name>
          .
          <year>2020</year>
          .
          <article-title>Automated detection of racial microaggressions using machine learning</article-title>
          .
          <source>In 2020 IEEE Symposium Series on Computational Intelligence (SSCI)</source>
          , pages
          <fpage>2477</fpage>
          -
          <lpage>2484</lpage>
          . IEEE.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <string-name>
            <given-names>Elisa</given-names>
            <surname>Bassignana</surname>
          </string-name>
          , Valerio Basile, and
          <string-name>
            <given-names>Viviana</given-names>
            <surname>Patti</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>Hurtlex: A multilingual lexicon of words to hurt</article-title>
          .
          <source>In 5th Italian Conference on Computational Linguistics</source>
          , CLiC-it
          <year>2018</year>
          , volume
          <volume>2253</volume>
          , pages
          <fpage>1</fpage>
          -
          <lpage>6</lpage>
          . CEUR-WS.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <string-name>
            <given-names>Rishabh</given-names>
            <surname>Bhardwaj</surname>
          </string-name>
          , Navonil Majumder, and
          <string-name>
            <given-names>Soujanya</given-names>
            <surname>Poria</surname>
          </string-name>
          .
          <year>2020</year>
          .
          <article-title>Investigating gender bias in bert</article-title>
          . arXiv preprint arXiv:
          <year>2009</year>
          .05021.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <string-name>
            <given-names>Tolga</given-names>
            <surname>Bolukbasi</surname>
          </string-name>
          ,
          <string-name>
            <surname>Kai-Wei</surname>
            <given-names>Chang</given-names>
          </string-name>
          , James Zou, Venkatesh Saligrama, and
          <string-name>
            <given-names>Adam</given-names>
            <surname>Kalai</surname>
          </string-name>
          .
          <year>2016</year>
          .
          <article-title>Man is to computer programmer as woman is to homemaker? debiasing word embeddings</article-title>
          .
          <source>arXiv preprint arXiv:1607</source>
          .
          <fpage>06520</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <string-name>
            <given-names>Luke</given-names>
            <surname>Breitfeller</surname>
          </string-name>
          , Emily Ahn, David Jurgens,
          <string-name>
            <given-names>and Yulia</given-names>
            <surname>Tsvetkov</surname>
          </string-name>
          .
          <year>2019</year>
          .
          <article-title>Finding microaggressions in the wild: A case for locating elusive phenomena in social media posts</article-title>
          .
          <source>In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLPIJCNLP)</source>
          , pages
          <fpage>1664</fpage>
          -
          <lpage>1674</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <string-name>
            <given-names>Tommaso</given-names>
            <surname>Caselli</surname>
          </string-name>
          , Valerio Basile, Jelena Mitrovic´, and
          <string-name>
            <given-names>Michael</given-names>
            <surname>Granitzer</surname>
          </string-name>
          .
          <year>2020</year>
          .
          <article-title>HateBERT: Retraining BERT for Abusive Language Detection in English</article-title>
          . arXiv preprint arXiv:
          <year>2010</year>
          .12472.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          <string-name>
            <given-names>Marta</given-names>
            <surname>Dynel</surname>
          </string-name>
          .
          <year>2021</year>
          .
          <article-title>Humour and (mock) aggression: Distinguishing cyberbullying from roasting</article-title>
          .
          <source>Language &amp; Communication</source>
          ,
          <volume>81</volume>
          :
          <fpage>17</fpage>
          -
          <lpage>36</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          <string-name>
            <given-names>Simona</given-names>
            <surname>Frenda</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>The role of sarcasm in hate speech. a multilingual perspective</article-title>
          .
          <source>In e Doctoral Symposium of the XXXIVInternational Conference of the Spanish Society for Natural Language Processing (SEPLN</source>
          <year>2018</year>
          ), pages
          <fpage>13</fpage>
          -
          <lpage>17</lpage>
          . Lloret, E.;
          <string-name>
            <surname>Saquete</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ;
          <article-title>Mart´ınez-</article-title>
          <string-name>
            <surname>Barco</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Moreno</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          <string-name>
            <given-names>Armand</given-names>
            <surname>Joulin</surname>
          </string-name>
          , Edouard Grave, Piotr Bojanowski, and
          <string-name>
            <given-names>Tomas</given-names>
            <surname>Mikolov</surname>
          </string-name>
          .
          <year>2016</year>
          .
          <article-title>Bag of tricks for efficient text classification</article-title>
          .
          <source>arXiv preprint arXiv:1607</source>
          .
          <fpage>01759</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          <string-name>
            <given-names>David</given-names>
            <surname>Jurgens</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Libby</given-names>
            <surname>Hemphill</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Eshwar</given-names>
            <surname>Chandrasekharan</surname>
          </string-name>
          .
          <year>2019</year>
          .
          <article-title>A just and comprehensive strategy for using NLP to address online abuse</article-title>
          .
          <source>In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics</source>
          , pages
          <fpage>3658</fpage>
          -
          <lpage>3666</lpage>
          , Florence, Italy, July. Association for Computational Linguistics.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          <string-name>
            <given-names>Anne</given-names>
            <surname>Lauscher and Goran Glavasˇ</surname>
          </string-name>
          .
          <year>2019</year>
          .
          <article-title>Are we consistently biased? multidimensional analysis of biases in distributional word vectors</article-title>
          . arXiv preprint arXiv:
          <year>1904</year>
          .11783.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          <string-name>
            <given-names>Thomas</given-names>
            <surname>Manzini</surname>
          </string-name>
          , Yao Chong Lim, Yulia Tsvetkov, and Alan W Black.
          <year>2019</year>
          .
          <article-title>Black is to criminal as caucasian is to police: Detecting and removing multiclass bias in word embeddings</article-title>
          . arXiv preprint arXiv:
          <year>1904</year>
          .04047.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          <string-name>
            <given-names>Binny</given-names>
            <surname>Mathew</surname>
          </string-name>
          , Ritam Dutt, Pawan Goyal, and
          <string-name>
            <given-names>Animesh</given-names>
            <surname>Mukherjee</surname>
          </string-name>
          . 2019a.
          <article-title>Spread of hate speech in online social media</article-title>
          .
          <source>In Proceedings of the 10th ACM conference on web science</source>
          , pages
          <fpage>173</fpage>
          -
          <lpage>182</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          <string-name>
            <given-names>Binny</given-names>
            <surname>Mathew</surname>
          </string-name>
          , Anurag Illendula, Punyajoy Saha, Soumya Sarkar, Pawan Goyal, and
          <string-name>
            <given-names>Animesh</given-names>
            <surname>Mukherjee</surname>
          </string-name>
          . 2019b.
          <article-title>Temporal effects of unmoderated hate speech in gab</article-title>
          . arXiv preprint arXiv:
          <year>1909</year>
          .10966.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          <string-name>
            <surname>Merriam-Webster</surname>
          </string-name>
          .
          <year>2021</year>
          .
          <article-title>Merriam-webster's definition of microaggression</article-title>
          . https://www.merriamwebster.com/dictionary/microaggres sion.
          <source>Accessed: 2021-03-08.</source>
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          <string-name>
            <given-names>Tomas</given-names>
            <surname>Mikolov</surname>
          </string-name>
          , Ilya Sutskever, Kai Chen, Greg S Corrado, and
          <string-name>
            <given-names>Jeff</given-names>
            <surname>Dean</surname>
          </string-name>
          .
          <year>2013</year>
          .
          <article-title>Distributed representations of words and phrases and their compositionality</article-title>
          . In C. J.
          <string-name>
            <surname>C. Burges</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Bottou</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Welling</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          <string-name>
            <surname>Ghahramani</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          K. Q. Weinberger, editors,
          <source>Advances in Neural Information Processing Systems</source>
          , volume
          <volume>26</volume>
          . Curran Associates, Inc.
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          <string-name>
            <given-names>Jeffrey</given-names>
            <surname>Pennington</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Socher</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Christopher D.</given-names>
            <surname>Manning</surname>
          </string-name>
          .
          <year>2014</year>
          .
          <article-title>Glove: Global vectors for word representation</article-title>
          .
          <source>In EMNLP.</source>
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          <string-name>
            <given-names>Timo</given-names>
            <surname>Schick</surname>
          </string-name>
          , Sahana Udupa, and Hinrich Schu¨tze.
          <year>2021</year>
          .
          <article-title>Self-diagnosis and self-debiasing: A proposal for reducing corpus-based bias in nlp</article-title>
          .
          <source>arXiv preprint arXiv:2103</source>
          .
          <fpage>00453</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          <string-name>
            <given-names>Derald</given-names>
            <surname>Sue</surname>
          </string-name>
          , Christina Capodilupo, Gina Torino, Jennifer Bucceri, be Aisha, Kevin Nadal, and
          <string-name>
            <given-names>Marta</given-names>
            <surname>Esquilin</surname>
          </string-name>
          .
          <year>2007</year>
          .
          <article-title>Racial microaggressions in everyday life: Implications for clinical practice</article-title>
          .
          <source>The American psychologist</source>
          ,
          <volume>62</volume>
          :
          <fpage>271</fpage>
          -
          <lpage>86</lpage>
          ,
          <fpage>05</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          <string-name>
            <given-names>Nathaniel</given-names>
            <surname>Swinger</surname>
          </string-name>
          , Maria
          <string-name>
            <surname>De-Arteaga</surname>
          </string-name>
          ,
          <article-title>Neil Thomas Heffernan IV, Mark DM Leiserson,</article-title>
          and Adam Tauman Kalai.
          <year>2019</year>
          .
          <article-title>What are the biases in my word embedding</article-title>
          ?
          <source>In Proceedings of the 2019 AAAI/ACM Conference on AI, Ethics, and Society</source>
          , pages
          <fpage>305</fpage>
          -
          <lpage>311</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          <string-name>
            <given-names>Zekun</given-names>
            <surname>Yang</surname>
          </string-name>
          and
          <string-name>
            <given-names>Juan</given-names>
            <surname>Feng</surname>
          </string-name>
          .
          <year>2020</year>
          .
          <article-title>A causal inference method for reducing gender bias in word embedding relations</article-title>
          .
          <source>In Proceedings of the AAAI Conference on Artificial Intelligence</source>
          , volume
          <volume>34</volume>
          , pages
          <fpage>9434</fpage>
          -
          <lpage>9441</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          <string-name>
            <given-names>Jieyu</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <surname>Yichao Zhou</surname>
            ,
            <given-names>Zeyu</given-names>
          </string-name>
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>Wei</given-names>
          </string-name>
          <string-name>
            <surname>Wang</surname>
            , and
            <given-names>KaiWei</given-names>
          </string-name>
          <string-name>
            <surname>Chang</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>Learning gender-neutral word embeddings</article-title>
          . arXiv preprint arXiv:
          <year>1809</year>
          .01496.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>