<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Mitigation of Biases in Deep Learning Language Models</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Ismael Garrido-Muñoz</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>CEATIC (Universidad de Jaén) Campus las lagunillas s/n. Jaén</institution>
          ,
          <addr-line>23071</addr-line>
          ,
          <country country="ES">Spain</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2022</year>
      </pub-date>
      <abstract>
        <p>Recent advances in artificial intelligence have made it possible to make our everyday lives better, from apps that translate with great accuracy, to search engines that understand your query, to virtual assistants that answer your questions. However, these models capture the biases present in society and incorporate them into their knowledge. These biases and prejudices appear in a multitude of applications and systems. Given the same context, the model will have vastly diferent results depending on attributes such as the subject's gender, race or religion. In this thesis we focus on natural language bias in neural networks. However, bias is present in many areas of artificial intelligence. This behaviour is pervasive, first the models capture these associations and then replicate them as a result of their application. Bias in AI is encompassed in study areas such as Fairness or Explainability 0000-0001-6656-9679 (I. Garrido-Muñoz)</p>
      </abstract>
      <kwd-group>
        <kwd>bias</kwd>
        <kwd>deep learning</kwd>
        <kwd>nlp</kwd>
        <kwd>fairness</kwd>
        <kwd>explicability</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>Artificial intelligence has come a long way in recent years, and a good part of this progress is
possible thanks to neural network models. These models that are trained with large amounts of
data have demonstrated a great capacity to capture reality. However, capturing reality with
such precision can sometimes be negative, since they also capture and replicate undesirable
stereotypes</p>
      <p>
        One example is the police COMPAS system in the United States. This system assigns detainees
a level of risk of recidivism. From an independent analysis, it was discovered that the system
failed for both white and black people[
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], but the type of error was diferent. In the case of
white people, the system would systematically provide a lower level of recidivism risk than the
actual level they should have, it was failing in their favor. While in the case of black people
the error was against them, the system assigned a higher level of risk. In this case, there is a
social problem in which an algorithm can be disruptive in people’s lives and simultaneously a
resource allocation problem in which a system whose malfunctioning causes resources not to
be allocated where they are really needed. A similar example can be found in a medical system
called Optum[
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], which would systematically allocate black patients fewer resources for their
Doctoral Symposium on Natural Language Processing from the PLN.net network 2022 (RED2018-102418-T), 21-23
https://isgarrido.github.io/ (I. Garrido-Muñoz)
treatment than white patients for the same level of need. This is a case of resource allocation
by a biased system that can negatively influence people’s health. There are several examples of
automated recruitment systems such as HireVue[
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] which uses artificial intelligence models
to evaluate candidates. However, the system disadvantaged candidates who deviated from the
model’s definition of normal. This behavior is quite frequent when the model was not trained
with examples that are suficiently diverse. In this case, it is intuited that HireVue malfunctioned
on candidates that were not English-native speakers, since their accents would confuse the
model. The most controversial part is that in those cases where the model was not properly
working, the candidate was automatically discarded and did not receive any information about
the reasons. This makes us think that the application of non-explainable models may be unfair
and even harmful in some situations. Amazon also discarded[
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] a similar tool for recruitment,
as it was found to be biased against women. Those are just some examples but those biases
have been found in systems where the model is applied to tasks such as computer vision[
        <xref ref-type="bibr" rid="ref5">5</xref>
        ],
audio processing[
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] or linguistic corpora[
        <xref ref-type="bibr" rid="ref7 ref8">7, 8</xref>
        ].
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. Bias in NLP with deep learning</title>
      <p>The study of bias in neural network language models can be divided into two tasks. On one
hand, the evaluation task aims to discern whether a model is biased. With a biased model, the
next step should be to characterize and quantify the bias. Characterization helps us understand
what form it takes and thus study what causes it and where it is encoded. Quantification helps
us to know to what extent its application to real-world problems can provoke harm to society.
Once the evaluation has been conducted, the next step is the mitigation task, also known as
correction of the bias, in this task the objective is to change the model to avoid the detrimental
efect that was found in previous phases.</p>
      <p>
        When quantifying or characterizing the bias present in a model, it is important to consider the
task to which the model is applied and which attributes should be protected for that task. For
example, in a personnel selection process, a protected attribute might be the candidate’s gender,
since it is a characteristic that should not influence the model’s perception of the candidate.
Other protected attributes could be age, race, religion, origin, etc. In the case of wanting to
study more than one protected attribute in the same model, one could consider applying a
method that automatically detects the bias for each protected attribute or applying an adapted
version of the method developed for each given protected attribute. For example, if we want to
check if there is a personnel selection bias, the system should score equally two candidates with
the same qualifications in which the only change is gender, nationality or race. If this is not the
case, we could study whether there is a systematic bias against each of these. In the following
section, I briefly present the study of bias in data models, for more detail see a previous paper[
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]
in which we conducted an in-depth review of the study of bias.
      </p>
      <sec id="sec-2-1">
        <title>2.1. Bias evaluation</title>
        <p>
          All kinds of models have been evaluated, from the most basic ones such as the study by Bolukbasi
et al. [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ] on Word Embeddings, in which it is already found that the models associate men and
women with diferent professions. While some analogies like  −   ≈  −  work
correctly and show us that  =  , other analogies like −  ≈     − ,
the model output is  = ℎ , as it does not associate computer programmer with woman,
moreover it finds a set of professions strongly associated with each gender.
        </p>
        <p>
          Later studies approach the bias issue from the point of view of Coreference Resolution, this
is the approach of Zhao et al. [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ] and Jieyu et al. [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ] with GloVe. Also for Word Embedding
models Caliskan et al. [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ] did create WEAT as an association test between a concept and a
protected attribute. Following the previous example, this test could be used to measure if any
of the genders is strongly associated with the profession while the other is not. This test is
used and extended later by Dev et al. [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ] or the work of Manzini et al. [15] that besides gender
also studies race and religion. Also, Lauscher et al. [16] has explored the problem of Word
Embeddings in more languages such as German, Spanish, Italian, Russian, Croatian and Turkish.
        </p>
        <p>More recent and complex models such as BERT or GPT-2 were also studied. In the work of
Vig [17] he introduces a tool to visualize the attention of the model to try to understand where
it is capturing unwanted biases. From the previously mentioned WEAT there are numerous
adaptations such as SEAT[18], that instead of testing against a word, does test the protected
attribute against a sentence for contextual models such as BERT. This work is also extended by
Tan and Celis [19] that adapts SEAT to work with the full context instead of just on a sentence
level. This latest evaluation method is applied to GPT-2, BERT, ELMo, among others.</p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Bias correction</title>
        <p>There are several techniques for bias correction, on the one hand, the correction of the vector
space of the model. In this technique the author search for the dimension or dimensions of
the vector space that encodes the bias and equalizes the diferences between, for example, the
distance between the protected attribute (man, woman) and the biased concept (the profession).
This is one of the approaches present on the work of Manzini et al. [15], Zhou et al. [20], Dev
and Phillips [21]. However the work of Gonen and Goldberg [22] suggests that this technique
does not eliminate bias but hides it, this work also shows how in the Word Embedding models
that had the bias removed, it can still be found and is therefore the current approaches are not
suficient.</p>
        <p>On the other hand, another line of work is helping the model acquire the ability to give
an unbiased result without changing its internals. Other studies[23] train or fine-tune the
model by doing some changes to its data like doing a data augmentation technique called
Counterfactual Data Augmentation where they balance the times that appear each instance
protected attribute (for example if in the data has ”he is a computer programmer”, this technique
would add ”she is a computer programmer”, this would be extensible to other protected attributes
with more dimensions such as race). The main advantage of fine-tuning already trained models
to reach similar results since training a model from scratch is costly and this would be an
impediment to its application given time or cost restrictions.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Relevance of the problem</title>
      <p>Day by day, these huge models are being integrated as part of products and systems in production.
On the one hand, we could speak of a purely economic disadvantage since a biased system
would not work properly for all potential users. On the other hand, these models have an impact
on people’s lives, there are certain contexts in which it would not be appropriate to use a model
that may contain some type of bias, such as systems for distributing public resources, systems
related to employment, or bank credit. Therefore, bias in data models should be studied, as well
as its causes, in order to either avoid using biased models or to mitigate harmful bias when
necessary.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Hypotheses and objectives</title>
      <p>The following hypothesis is assumed: Given a language model based on deep learning, it will
be possible to discern whether it contains biases, and characterize, measure, and mitigate them.</p>
      <p>The following objectives are established:
• Conduct an intensive study of the state of the art regarding detection, evaluation or
mitigation of biases in deep learning models.
• Analyze and characterize biases present in existing models.
• Development of techniques and algorithms for unsupervised or semi-supervised detection
and characterization of bias in existing models.
• Development of techniques and algorithms for the mitigation or correction of bias in
existing models.
• Development of techniques that allow the creation of bias-free models in a given context.
• Development of techniques that allow the creation of robust and explainable models.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Methodology and the proposed experiments</title>
      <p>
        To achieve these objectives, we first conducted an extensive study[
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] of the state of the art
regarding bias in data models. The next step was to run some experiments to look for bias in
20 deep learning models available for Spanish. We did adapt some of the existing association
tests. We did adapt the gender experiments conducted in StereoSet[24] . In Stereoset they
compare the result of the model by giving it a sentence that acts as a context and is followed
by either a stereotyped sentence or an anti-stereotyped sentence. This way you can test the
model to see if it is capturing the stereotype. We took this idea to design our experiment. We
created a set of sentences where the only change is the genre. For example Su profesor es muy
MASK , Su profesora es muy MASK . We ask the model to complete these sentences, expecting
to receive an adjective. This way can study how the models perceive men and women. The
initial results gave us some clues, you can see that the models perceive women by their physical
characteristics. In the table 1 we have an example, we can see how more words referring to
the physical appearance ( in italics ) appear in the female version of the sentence, in fact in the
female version the first 3 positions are occupied by words referring to the physical appearance
when we order by the probability that the model assigns to each token (score).
      </p>
      <p>El chico es &lt;mask&gt;.</p>
      <p>Index token_str
0 guapo
1 genial
2 buenísimo
3 encantador
4 precioso
5 perfecto
6 espectacular
7 impresionante
8 estupendo
9 bueno</p>
      <p>This experiment confirmed that Spanish language models capture a bias that in certain
contexts can be detrimental. The proposed adjectives are categorized according to the Supersenses
taxonomy[26]. The main used categories to group the adjectives are Body, Behaviour, Social,
Feel, Mind. By applying this task to a set of template sentences and aggregating the results,
we concluded that Spanish language models perceive women by their physical appearance
(adjectives on the body category) while men by their behavior-related characteristics. This
work is pending publication.</p>
      <p>In order to facilitate the interpretation of the results of this study, a tool was developed that
allows us to explore the results of the study in a visual way. In the tool we can:
• Visualize how the predictions of each model are distributed among the diferent categories,
immediately comparing the weight associated with the male sentences with respect to
the female sentences.
• Explore in detail the model values for each category, as well as visualize them using a
heatmap to see which models are the most biased and if they are biased towards one
gender or the other.
• To know the number of adjectives contributed by each model for male, female, and both.
• To explore each of the specific model responses for each of the templates.
. The tool can be found at https://ismael.codes/categoryviewer/</p>
      <sec id="sec-5-1">
        <title>5.1. Future work</title>
        <p>The next step is to find out what causes these behaviors within the model. We could pose it
as a problem of causal inference in which it is a confounder that causes the biased result, by
ifnding the confounder we could try to correct the model to eliminate its efect and thus obtain
unbiased results. A concrete approach following this idea would be to study the variation of
the model parameters between pairs of sentences, one biased and the other unbiased. This way
we could locate what the diference is in the model and therefore be able correct the model.
https://aclanthology.org/2021.emnlp-main.411. doi:1 0 . 1 8 6 5 3 / v 1 / 2 0 2 1 . e m n l p - m a i n . 4 1 1 .
[15] T. Manzini, L. Yao Chong, A. W. Black, Y. Tsvetkov, Black is to criminal as caucasian is
to police: Detecting and removing multiclass bias in word embeddings, in: Proceedings
of the 2019 Conference of the North American Chapter of the Association for
Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers),
Association for Computational Linguistics, Minneapolis, Minnesota, 2019, pp. 615–621.</p>
        <p>URL: https://aclanthology.org/N19-1062. doi:1 0 . 1 8 6 5 3 / v 1 / N 1 9 - 1 0 6 2 .
[16] A. Lauscher, G. Glavas, S. P. Ponzetto, I. Vulic, A general framework for implicit and
explicit debiasing of distributional word vector spaces, in: AAAI, 2020.
[17] J. Vig, A multiscale visualization of attention in the transformer model, 2019, pp. 37–42.</p>
        <p>doi:1 0 . 1 8 6 5 3 / v 1 / P 1 9 - 3 0 0 7 .
[18] C. May, A. Wang, S. Bordia, S. R. Bowman, R. Rudinger, On measuring social biases
in sentence encoders, in: Proceedings of the 2019 Conference of the North American
Chapter of the Association for Computational Linguistics: Human Language Technologies,
Volume 1 (Long and Short Papers), Association for Computational Linguistics, Minneapolis,
Minnesota, 2019, pp. 622–628. URL: https://aclanthology.org/N19-1063. doi:1 0 . 1 8 6 5 3 / v 1 /
N 1 9 - 1 0 6 3 .
[19] Y. C. Tan, L. E. Celis, Assessing social and intersectional biases in contextualized word
representations, in: NeurIPS, 2019.
[20] P. Zhou, W. Shi, J. Zhao, K.-H. Huang, M. Chen, K.-W. Chang, Analyzing and mitigating
gender bias in languages with grammatical gender and bilingual word embeddings, in:
ACL 2019, 2019.
[21] S. Dev, J. M. Phillips, Attenuating bias in word vectors, CoRR abs/1901.07656 (2019). URL:
http://arxiv.org/abs/1901.07656. a r X i v : 1 9 0 1 . 0 7 6 5 6 .
[22] H. Gonen, Y. Goldberg, Lipstick on a pig: Debiasing methods cover up systematic gender
biases in word embeddings but do not remove them, in: NAACL-HLT, 2019.
[23] R. H. Maudslay, H. Gonen, R. Cotterell, S. Teufel, It’s all in the name: Mitigating gender
bias with name-based counterfactual data substitution, CoRR abs/1909.00871 (2019). URL:
http://arxiv.org/abs/1909.00871. a r X i v : 1 9 0 9 . 0 0 8 7 1 .
[24] M. Nadeem, A. Bethke, S. Reddy, Stereoset: Measuring stereotypical bias in pretrained
language models, CoRR abs/2004.09456 (2020). URL: https://arxiv.org/abs/2004.09456.
a r X i v : 2 0 0 4 . 0 9 4 5 6 .
[25] A. Gutiérrez-Fandiño, J. Armengol-Estapé, M. Pàmies, J. Llop-Palao, J. Silveira-Ocampo,
C. P. Carrino, C. Armentano-Oller, C. Rodriguez-Penagos, A. Gonzalez-Agirre, M. Villegas,
Maria: Spanish language models, Procesamiento del Lenguaje Natural 68 (2022) 39–60.</p>
        <p>URL: http://journal.sepln.org/sepln/ojs/ojs/index.php/pln/article/view/6405.
[26] Y. Tsvetkov, N. Schneider, D. Hovy, A. Bhatia, M. Faruqui, C. Dyer, Augmenting
English Adjective Senses with Supersenses, in: Proceedings of the Ninth
International Conference on Language Resources and Evaluation (LREC’14), European
Language Resources Association (ELRA), Reykjavik, Iceland, 2014, pp. 4359–4365. URL:
http://www.lrec-conf.org/proceedings/lrec2014/pdf/1096_Paper.pdf.</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>J. L. Julia</given-names>
            <surname>Angwin</surname>
          </string-name>
          ,
          <article-title>Machine bias - there's software used across the country to predict future criminals. and it's biased against blacks</article-title>
          .,
          <year>2016</year>
          . URL: https://www.propublica.org/article/ machine-bias
          <article-title>-risk-assessments-in-criminal-sentencing.</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>Z. O. U.</given-names>
            <surname>Berkeley</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Obermeyer</surname>
          </string-name>
          , U. Berkeley, S. M. U. o. Chicago, S. Mullainathan, U. o. Chicago,
          <string-name>
            <given-names>O. M. A.</given-names>
            <surname>Metrics</surname>
          </string-name>
          ,
          <article-title>Dissecting racial bias in an algorithm that guides health decisions for 70 million people</article-title>
          :
          <source>Proceedings of the conference on fairness, accountability, and transparency</source>
          ,
          <year>2019</year>
          . URL: https://dl.acm.org/doi/10.1145/3287560.3287593.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>D.</given-names>
            <surname>Harwell</surname>
          </string-name>
          ,
          <article-title>A face-scanning algorithm increasingly decides whether you deserve the job</article-title>
          ,
          <year>2019</year>
          . URL: https://www.washingtonpost.com/technology/2019/10/22/ ai-hiring
          <article-title>-face-scanning-algorithm-increasingly-decides-whether-you-deserve-job/.</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>J.</given-names>
            <surname>Dastin</surname>
          </string-name>
          ,
          <article-title>Amazon scraps secret ai recruiting tool that showed bias against women</article-title>
          ,
          <year>2018</year>
          . URL: https://www.reuters.com/article/ us-amazon
          <article-title>-com-jobs-automation-insight-idUSKCN1MK08G.</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>A.</given-names>
            <surname>Howard</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Borensteion</surname>
          </string-name>
          , Trust and bias in robots,
          <year>2019</year>
          . URL: https://www. americanscientist.org/article/trust-and
          <article-title>-bias-in-robots.</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>J.</given-names>
            <surname>Rodger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Pendharkar</surname>
          </string-name>
          ,
          <article-title>A field study of the impact of gender and user's technical experience on the performance of voice-activated medical tracking application</article-title>
          ,
          <source>Int. J. Hum.-Comput. Stud</source>
          .
          <volume>60</volume>
          (
          <year>2004</year>
          )
          <fpage>529</fpage>
          -
          <lpage>544</lpage>
          .
          <source>doi:1 0 . 1 0 1 6 / j . i j h c s . 2 0 0 3 . 0 9 . 0 0 5 .</source>
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>J. A.</given-names>
            <surname>Bullinaria</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. P.</given-names>
            <surname>Levy</surname>
          </string-name>
          ,
          <article-title>Extracting semantic representations from word co-occurrence statistics: A computational study</article-title>
          ,
          <source>Behavior Research Methods</source>
          <volume>39</volume>
          (
          <year>2007</year>
          )
          <fpage>510</fpage>
          -
          <lpage>526</lpage>
          . URL: https://doi.org/10.3758/BF03193020.
          <source>doi:1 0 . 3 7 5 8 / B F 0 3</source>
          <volume>1 9 3 0 2 0 .</volume>
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>M.</given-names>
            <surname>Barlow</surname>
          </string-name>
          ,
          <article-title>Michael stubbs. text and corpus analysis: Computer-assisted studies of language and culture</article-title>
          ,
          <source>International Journal of Corpus Linguistics</source>
          <volume>3</volume>
          (
          <year>1998</year>
          )
          <fpage>319</fpage>
          -
          <lpage>327</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>I.</given-names>
            <surname>Garrido-Muñoz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Montejo-Ráez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Martínez-Santiago</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. A.</given-names>
            <surname>Ureña-López</surname>
          </string-name>
          ,
          <article-title>A survey on bias in deep nlp</article-title>
          ,
          <source>Applied Sciences</source>
          <volume>11</volume>
          (
          <year>2021</year>
          ). URL: https://www.mdpi.
          <source>com/2076-3417/11/7/ 3184. doi:1 0 . 3 3</source>
          <volume>9 0</volume>
          / a p p
          <volume>1 1 0 7 3 1 8 4 .</volume>
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>T.</given-names>
            <surname>Bolukbasi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. Y.</given-names>
            <surname>Zou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Saligrama</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Kalai</surname>
          </string-name>
          , Man is to Computer Programmer as Woman is to Homemaker? Debiasing Word Embeddings,
          <source>CoRR abs/1607</source>
          .06520 (
          <year>2016</year>
          ). URL: http://arxiv.org/abs/1607.06520.
          <article-title>a r X i v : 1 6 0 7 . 0 6 5 2 0</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>J.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Yatskar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Ordonez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.-W.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <article-title>Gender Bias in Coreference Resolution: Evaluation and Debiasing Methods</article-title>
          , arXiv e-prints (
          <year>2018</year>
          ) arXiv:
          <year>1804</year>
          .06876.
          <article-title>a r X i v : 1 8 0 4 . 0 6 8 7 6</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>Jieyu</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>Zhou</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>K.-W.</given-names>
          </string-name>
          <string-name>
            <surname>Chang</surname>
          </string-name>
          , Learning Gender-Neutral Word Embeddings, arXiv e-prints (
          <year>2018</year>
          ) arXiv:
          <year>1809</year>
          .01496.
          <article-title>a r X i v : 1 8 0 9 . 0 1 4 9 6</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>A.</given-names>
            <surname>Caliskan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. J.</given-names>
            <surname>Bryson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Narayanan</surname>
          </string-name>
          ,
          <article-title>Semantics derived automatically from language corpora contain human-like biases</article-title>
          ,
          <source>Science</source>
          <volume>356</volume>
          (
          <year>2017</year>
          )
          <fpage>183</fpage>
          -
          <lpage>186</lpage>
          . URL: https://www.science.org/doi/abs/10.1126/science.aal4230.
          <source>doi:1 0 . 1 1 2</source>
          <article-title>6 / s c i e n c e . a a l 4 2 3 0 . a r X i v</article-title>
          : h t t p s : / / w w w . s
          <article-title>c i e n c e . o r g / d o i / p d f / 1 0 . 1 1 2 6 / s c i e n c e . a a l 4 2 3 0</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>S.</given-names>
            <surname>Dev</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Phillips</surname>
          </string-name>
          , V. Srikumar,
          <article-title>OSCaR: Orthogonal subspace correction and rectification of biases in word embeddings</article-title>
          ,
          <source>in: Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing</source>
          , Association for Computational Linguistics, Online and
          <string-name>
            <given-names>Punta</given-names>
            <surname>Cana</surname>
          </string-name>
          , Dominican Republic,
          <year>2021</year>
          , pp.
          <fpage>5034</fpage>
          -
          <lpage>5050</lpage>
          . URL:
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>