<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <issn pub-type="ppub">1613-0073</issn>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Towards Fairness and Inclusivity</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Ismael Garrido-Muñoz</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Universidad de Jaén</institution>
          ,
          <addr-line>Campus Las Lagunillas s/n, 23071 Jaén</addr-line>
          ,
          <country>España</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>The emergence of large language models (LLMs) has revolutionized the field of natural language processing, facilitating remarkable progress across various domains. However, the inherent opaqueness of these models, functioning as black boxes, presents significant challenges. The lack of transparency obstructs our comprehension of their internal mechanisms and decision-making processes, raising concerns about their reliability and fairness. Various forms of biases have already been identified within these models. It is crucial to identify the location and encoding of these biases within LLMs to enable necessary modifications that ensure their safe and equitable application free of social biases in all kind of areas. Given the extensive deployment of LLMs in real-world applications, their impact on individuals' lives is magnified. Thus, the subsequent phase of this thesis will focus on efectively mitigating biases in deep learning models.</p>
      </abstract>
      <kwd-group>
        <kwd>bias</kwd>
        <kwd>deep learning</kwd>
        <kwd>nlp</kwd>
        <kwd>fairness</kwd>
        <kwd>mitigation</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>CEUR
ceur-ws.org</p>
    </sec>
    <sec id="sec-2">
      <title>1. Introduction</title>
      <p>
        The advent of GPT-3[
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] has sparked a massive adoption of this model, with predictions of its
profound impact on the labor market, as outlined by [2]. This remarkable influence stems from
the diverse range of capabilities that these models possess, including question answering, text
generation, translation, summarization, information retrieval, act as a conversational agent,
programming assistance, educational support, story telling and more.
      </p>
      <p>However, despite the tremendous utility of LLMs, they also pose an emerging challenge:
their tendency to operate as black boxes. While they exhibit impressive performance, their
internal mechanisms and decision-making processes often remain opaque, making them dificult
to comprehend and explain. This lack of transparency gives rise to concerns regarding their
trustworthiness, fairness, and the potential biases that may be embedded within their models.</p>
      <p>The concept of a black box refers to a system or model where the inputs and outputs are
known, but the inner mechanisms and algorithms that generate those outputs remain concealed
or poorly understood. LLMs, with their complex neural networks and millions, or even
billions, of parameters, are intricate black boxes that often surpass human comprehension. This
opaqueness hampers our ability to fully grasp the decision-making processes of these models,
https://ismael.codes/ (I. Garrido-Muñoz)
CEUR
Workshop
Proceedings</p>
      <p>© 2023 Copyright for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).
making it dificult to tackle biases, recognize potential vulnerabilities, and guarantee ethical
and responsible utilization. Consequently, there is a pressing need to enhance transparency
and develop techniques that shed light on the inner workings of LLMs.</p>
      <p>In recent years, artificial intelligence has made significant advances, and a substantial
portion of this progress can be attributed to neural network models. These models, trained on
extensive datasets, have showcased remarkable capabilities in capturing various aspects of
reality. However, while their ability to capture reality with precision is commendable, it can
also have negative implications. One such concern arises from their propensity to inadvertently
perpetuate and replicate undesirable stereotypes.</p>
      <p>These models are already being used in multiple production systems such as medical
systems[3], legal systems[4], hiring[5], content moderation[6], CRM[7], marketing[8], virtual
assistants, harmful content detection[9], chatbots, etc.</p>
      <p>These systems are used in products despite having proven to be unsafe. It is well known
that sometimes these black boxes cause unintended harm. One example is the police COMPAS
system, which assigned an unreal recidivism value to both white and black people. For white
individuals, the assigned value was lower than the actual value, while for black individuals, it
was higher than the actual value[10]. Another example is the medical system called Optum[11],
which systematically allocated fewer resources for the treatment of black patients compared to
white patients with the same level of need.</p>
      <p>
        This realization raises concerns about the fairness and potential harm that may arise from the
application of non-explainable models in certain situations. For instance, Amazon discontinued
the use of a recruitment tool [12] when it was discovered to be biased against women. These
examples highlight the presence of biases not only in language models but also in systems
employing computer vision [13], audio processing [14], and linguistic corpora [15], [
        <xref ref-type="bibr" rid="ref2">16</xref>
        ]. It is
crucial to address these biases as they can perpetuate inequality and have real-world consequences.
Understanding and mitigating biases in such systems is a pressing concern.
      </p>
      <p>
        In the case of GPT-3[
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] (or its frontends like Chat-GPT or Bing GPT) or Google’s alternative,
Bard[
        <xref ref-type="bibr" rid="ref3">17</xref>
        ], studying these models is not feasible because they are provided as services through
APIs or web interfaces. However, there have been releases of models with similar numbers of
parameters and capabilities as the aforementioned ones. For instance, models like Llama[
        <xref ref-type="bibr" rid="ref4">18</xref>
        ],
Vicuna[
        <xref ref-type="bibr" rid="ref5">19</xref>
        ], Bloom[
        <xref ref-type="bibr" rid="ref6">20</xref>
        ], OPT[
        <xref ref-type="bibr" rid="ref7">21</xref>
        ], XGLM[
        <xref ref-type="bibr" rid="ref8">22</xref>
        ], and the recent Falcon[
        <xref ref-type="bibr" rid="ref9">23</xref>
        ] do provide access to
the trained models weights. This access enables us to review, correct, or mitigate any biases
present in them.
      </p>
      <p>This will be the next step of the thesis. Next, we provide a brief overview of evaluation
techniques, followed by a collection of the most relevant techniques for bias mitigation. In
previous works, such as the one mentioned in , a broader summary of the state-of-the-art in
studying bias in language models can be found.</p>
      <p>
        This will be the next step of the thesis. In the following section, we provide a brief overview
of evaluation techniques, followed by a collection of the most relevant techniques for bias
mitigation. In previous works, a broader summary of the state-of-the-art in studying bias in
language models can be found [
        <xref ref-type="bibr" rid="ref10">24</xref>
        ].
      </p>
    </sec>
    <sec id="sec-3">
      <title>2. Bias in NLP with deep learning</title>
      <p>
        When we talk about bias in language models, we can approach it as a representational
problem[
        <xref ref-type="bibr" rid="ref11">25</xref>
        ]. This refers to the bias that certain demographic groups face in terms of
misrepresentation, including negative associations or even their absence in the data and consequently in the
model. On the other hand, we can approach it as an allocation problem, which refers to issues of
opportunities or resource distribution for individuals belonging to specific demographic groups.
      </p>
      <sec id="sec-3-1">
        <title>2.1. Bias evaluation</title>
        <p>
          There is extensive work when it comes to evaluating language models for bias, starting with
the work of Bolukbasi et al. [
          <xref ref-type="bibr" rid="ref12">26</xref>
          ] on simple word embeddings. Later studies approached the
bias issue from the perspective of coreference resolution, such as [
          <xref ref-type="bibr" rid="ref13">27</xref>
          ] with GloVe embeddings.
Bias is also examined by measuring the association between concepts and protected attributes.
Caliskan et al. [
          <xref ref-type="bibr" rid="ref14">28</xref>
          ] created the Word Embedding Association Test (WEAT) for this purpose. This
test was extended by Dev et al. [
          <xref ref-type="bibr" rid="ref15">29</xref>
          ] and Manzini et al. [30]. Also it was extended by Lauscher
et al. [31] by adding more protected attributes and applying it to languages other than English .
It was later on adapted to more complex models like BERT, under the name SEAT, by May et al.
[32] and Tan and Celis [33].
        </p>
        <p>There are other approaches for more complex models like BERT or GPT-2.Vig [34] introduced
visualization tools to understand where these models capture unwanted biases by examining
their attention. Additionally, adaptations of WEAT, such as SEAT, have been developed. SEAT
tests the protected attribute against a sentence instead of a word, specifically designed for
contextual models like BERT. This work was further extended to consider the full context
instead of just the sentence level. The latest evaluation method is applied to models like GPT-2,
BERT, ELMo, among others.</p>
        <p>More complex models also make serious errors. A compendium of errors discovered in
ChatGPT is presented in the work of Borji [35]. The paper explains that this model is unable to
successfully complete tasks that require spatial, temporal, or physical reasoning unless it has
been specifically trained for those tasks.</p>
      </sec>
      <sec id="sec-3-2">
        <title>2.2. Bias correction</title>
        <p>
          The main approaches to address bias in language models consist of the following: fine-tuning
the model [36], data augmentation to balance categories and avoid distortions towards one
category[
          <xref ref-type="bibr" rid="ref13">27</xref>
          ], protecting the attribute during model training to prevent bias capture[37], or
correcting the vector space of the model as presented in the works of Manzini et al. [30], Zhou
et al. [38], Dev and Phillips [37]. Among these techniques, fine tuning and model editing are
considered the most realistic, especially in the case of large-scale models, since retraining a
model from scratch would be very costly in terms of time, hardware resources, money and the
efort required to perform the pre-processing and tuning of training data.
        </p>
        <p>One of the most promising techniques for model editing involves identifying how the model
encodes certain knowledge and then making edits accordingly. The proposal of Meng et al.
[39] focuses on editing factual knowledge and serves as a foundation for further adaptation.
This technique first identifies the model’s influential parts that contribute the most weight in
choosing the last token by using causal mediation analysis. From there, the model’s weights
are edited to guide it towards the desired token. For example, if the model answers Obama
to the question ”What is the surname of the U.S. president?”, the weights can be located and
corrected to select the desired token Biden since this would be the updated and accurate answer.
Similarly, this method can be generalized to make broader corrections. In fact, in a subsequent
work[40] they adapt this method to perform mass corrections across the model weights. Then
they evaluate whether the model only edits knowledge for the specific context given in the
prompt or if it can generalize by asking questions about the same fact using diferent questions
or contexts.</p>
        <p>These techniques hold great potential in tackling bias, enhancing the accuracy, and bolstering
the reliability of language models. By facilitating targeted edits that align with desired outcomes,
these approaches enable the mitigation of unwanted biases in the models’ responses. As a
result, they contribute to an improved understanding of fairness and ensure more reliable and
unbiased outputs from language models.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>3. Relevance of the problem</title>
      <p>Every day, these enormous models are increasingly integrated into various products and
production systems. However, this integration comes with its own set of challenges. From an
economic standpoint, utilizing a biased system can lead to significant disadvantages, as it may
not function efectively for all users. On the other hand, the impact of these models on people’s
lives cannot be overlooked. There are specific contexts, such as systems for resource distribution,
employment, or bank credit, where it is crucial to avoid using models that may contain any
form of bias. Therefore, it is imperative to thoroughly study bias in data models and understand
its underlying causes. This knowledge will enable us to either avoid deploying biased models
altogether or develop strategies to mitigate harmful biases when they arise.</p>
      <p>Furthermore, when a language model is identified as not performing adequately in a
production system, such as a commercial product, companies face important decisions. Given
the immense size and cost associated with training these models, some proposed solutions
may be dificult to justify from an economic perspective. For instance, training the model
from scratch with revised, filtered, or corrected training data would entail significant expenses.
Another option, albeit costly, could involve discontinuing the use of the model, as a poorly
performing model is unsuitable for deployment in production systems. This proposition gains
some relevance considering the potential non-compliance of such models with new European
AI regulations[41]. Alternatively, more practical approaches could involve retraining the model
or leveraging state-of-the-art bias mitigation techniques to address the identified issues.</p>
      <p>The choice of approach will depend on various factors, including the severity of the bias,
the feasibility of retraining or mitigating the model, and the legal and ethical obligations that
must be met. Regardless of the chosen course of action, it is essential to proactively address and
rectify bias issues to ensure responsible and fair deployment of language models in real-world
applications. By doing so, we can foster inclusivity, promote equitable outcomes, and uphold
the principles of fairness and ethical AI.</p>
    </sec>
    <sec id="sec-5">
      <title>4. Hypotheses and objectives</title>
      <p>The following hypothesis is assumed: Given a language model based on deep learning, it will
be possible to discern whether it contains biases, and characterize, measure, and mitigate them.</p>
      <p>The following objectives are established:
• Conduct an intensive study of the state of the art regarding detection, evaluation, or
mitigation of biases in deep learning models.
• Analyze and characterize biases present in existing models.
• Development of techniques and algorithms for unsupervised or semi-supervised detection
and characterization of bias in existing models.
• Development of techniques and algorithms for the mitigation or correction of bias in
existing models.</p>
      <p>At this phase of the thesis, our primary focus is on the last point.</p>
    </sec>
    <sec id="sec-6">
      <title>5. Methodology and the proposed experiments</title>
      <p>As we move forward with the use of large language models, our next step will involve adapting
and evaluating the previous work[42] in the context of LLMs. Specifically, the previous study
shed light on how models tend to perceive women based on their physical appearance, while
men are assessed primarily based on their behavior. This pattern was observed across the
majority of the models investigated.</p>
      <p>To proceed, we will replicate the aforementioned experiment using large language models
(LLMs) and analyze to what extent increasing the model size afects bias, whether it exacerbates
or reduces it. Once this evaluation is completed, our focus will shift towards bias mitigation
strategies.</p>
      <p>To mitigate bias, we will construct a corpus of prompts that elicit biased responses from the
models. This corpus will serve as a foundation for our work in two main areas. First, we will
develop methods to detect and identify biased terms produced by the model in its responses.
Second, we will explore the previously discussed fact editing techniques to edit the behavior
of the model for the detected biases in order to reduce or eliminate them. This will require
adapting the causal mediation analysis mechanism to our problem, since editing a specific fact
is not the same as making an edit that causes a trade-of between diferent classes of a protected
attribute. After the editing process, we will evaluate the performance of the model with the
same set of prompts to check the efectiveness of the mitigation method.</p>
      <p>By undertaking these steps, we aim to gain insights into the behavior of large language
models regarding bias and work towards developing efective strategies for bias mitigation.
S. Gray, B. Chess, J. Clark, C. Berner, S. McCandlish, A. Radford, I. Sutskever, D. Amodei,
Language models are few-shot learners, CoRR abs/2005.14165 (2020). URL: https://arxiv.or
g/abs/2005.14165. arXiv:2005.14165.
[2] T. Eloundou, S. Manning, P. Mishkin, D. Rock, Gpts are gpts: An early look at the labor
market impact potential of large language models, ArXiv abs/2303.10130 (2023).
[3] S. Velupillai, H. Suominen, M. Liakata, A. Roberts, A. D. Shah, K. Morley, D. Osborn,
J. Hayes, R. Stewart, J. Downs, W. Chapman, R. Dutta, Using clinical natural language
processing for health outcomes research: Overview and actionable suggestions for future
advances, J Biomed Inform 88 (2018) 11–19.
[4] R. DALE, Law and word order: Nlp in legal tech, Natural Language Engineering 25 (2019)
211–217. doi:10.1017/S1351324918000475.
[5] M. Bogen, A. Rieke, Help wanted: an examination of hiring algorithms, equity, and bias,
2018.
[6] T. Gillespie, Custodians of the Internet: Platforms, Content Moderation, and the Hidden</p>
      <p>Decisions That Shape Social Media, 2018. doi:10.12987/9780300235029.
[7] Salesforce, Salesforce Announces AI Cloud – Bringing Trusted Generative AI to the
Enterprise — investor.salesforce.com, https://investor.salesforce.com/press-releases/pres
s-release-details/2023/Salesforce-Announces-AI-Cloud--Bringing-Trusted-Generative-A
I-to-the-Enterprise/default.aspx, 2023. [Accessed 18-Jun-2023].
[8] Adobe, Adobe Announces New Sensei GenAI Services to Reimagine End-to-End Marketing
Workflows — news.adobe.com, https://news.adobe.com/news/news-details/2023/Adobe-A
nnounces-New-Sensei-GenAI-Services-to-Reimagine-End-to-End-Marketing-Workflows
/default.aspx, 2023. [Accessed 18-Jun-2023].
[9] S. Tabahriti, Twitter is now relying more on AI to identify harmful content, says its new
trust and safety chief — businessinsider.com, https://www.businessinsider.com/twitter-n
ow-relying-more-ai-identify-harmful-content-2022-12, 2022. [Accessed 18-Jun-2023].
[10] J. L. Julia Angwin, Machine bias - there’s software used across the country to predict future
criminals. and it’s biased against blacks., 2016. URL: https://www.propublica.org/article/m
achine-bias-risk-assessments-in-criminal-sentencing.
[11] Z. O. U. Berkeley, Z. Obermeyer, U. Berkeley, S. M. U. o. Chicago, S. Mullainathan, U. o.</p>
      <p>Chicago, O. M. A. Metrics, Dissecting racial bias in an algorithm that guides health decisions
for 70 million people: Proceedings of the conference on fairness, accountability, and
transparency, 2019. URL: https://dl.acm.org/doi/10.1145/3287560.3287593.
[12] J. Dastin, Amazon scraps secret ai recruiting tool that showed bias against women, 2018.</p>
      <p>URL: https://www.reuters.com/article/us-amazon-com-jobs-automation-insight-idUSK
CN1MK08G.
[13] A. Howard, J. Borensteion, Trust and bias in robots, 2019. URL: https://www.americanscie
ntist.org/article/trust-and-bias-in-robots.
[14] J. Rodger, P. Pendharkar, A field study of the impact of gender and user’s technical
experience on the performance of voice-activated medical tracking application, Int. J.</p>
      <p>Hum.-Comput. Stud. 60 (2004) 529–544. doi:10.1016/j.ijhcs.2003.09.005.
[15] J. A. Bullinaria, J. P. Levy, Extracting semantic representations from word co-occurrence
statistics: A computational study, Behavior Research Methods 39 (2007) 510–526. URL:
https://doi.org/10.3758/BF03193020. doi:10.3758/BF03193020.
on Empirical Methods in Natural Language Processing, Association for Computational
Linguistics, Online and Punta Cana, Dominican Republic, 2021, pp. 5034–5050. URL: https:
//aclanthology.org/2021.emnlp-main.411. doi:10.18653/v1/2021.emnlp- main.411.
[30] T. Manzini, L. Yao Chong, A. W. Black, Y. Tsvetkov, Black is to criminal as caucasian is
to police: Detecting and removing multiclass bias in word embeddings, in: Proceedings
of the 2019 Conference of the North American Chapter of the Association for
Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers),
Association for Computational Linguistics, Minneapolis, Minnesota, 2019, pp. 615–621.</p>
      <p>URL: https://aclanthology.org/N19-1062. doi:10.18653/v1/N19- 1062.
[31] A. Lauscher, G. Glavas, S. P. Ponzetto, I. Vulic, A general framework for implicit and
explicit debiasing of distributional word vector spaces, in: AAAI, 2020.
[32] C. May, A. Wang, S. Bordia, S. R. Bowman, R. Rudinger, On measuring social biases
in sentence encoders, in: Proceedings of the 2019 Conference of the North American
Chapter of the Association for Computational Linguistics: Human Language Technologies,
Volume 1 (Long and Short Papers), Association for Computational Linguistics, Minneapolis,
Minnesota, 2019, pp. 622–628. URL: https://aclanthology.org/N19-1063. doi:10.18653/v1/
N19- 1063.
[33] Y. C. Tan, L. E. Celis, Assessing social and intersectional biases in contextualized word
representations, in: NeurIPS, 2019.
[34] J. Vig, A multiscale visualization of attention in the transformer model, 2019, pp. 37–42.</p>
      <p>doi:10.18653/v1/P19- 3007.
[35] A. Borji, A categorical archive of chatgpt failures, 2023. doi:10.21203/rs.3.rs- 2895792
/v1.
[36] R. H. Maudslay, H. Gonen, R. Cotterell, S. Teufel, It’s all in the name: Mitigating gender
bias with name-based counterfactual data substitution, CoRR abs/1909.00871 (2019). URL:
http://arxiv.org/abs/1909.00871. arXiv:1909.00871.
[37] S. Dev, J. M. Phillips, Attenuating bias in word vectors, CoRR abs/1901.07656 (2019). URL:
http://arxiv.org/abs/1901.07656. arXiv:1901.07656.
[38] P. Zhou, W. Shi, J. Zhao, K.-H. Huang, M. Chen, K.-W. Chang, Analyzing and mitigating
gender bias in languages with grammatical gender and bilingual word embeddings, in:
ACL 2019, 2019.
[39] K. Meng, D. Bau, A. Andonian, Y. Belinkov, Locating and editing factual associations in</p>
      <p>GPT, Advances in Neural Information Processing Systems 36 (2022).
[40] K. Meng, A. Sen Sharma, A. Andonian, Y. Belinkov, D. Bau, Mass editing memory in a
transformer, arXiv preprint arXiv:2210.07229 (2022).
[41] H. Ziady, Europe is leading the race to regulate AI. Here’s what you need to know | CNN
Business — edition.cnn.com, https://edition.cnn.com/2023/06/15/tech/ai-act-europe-key-t
akeaways/index.html, 2023. [Accessed 18-Jun-2023].
[42] I. Garrido, A. Montejo Raéz, F. Martínez Santiago, Maria and beto are sexist: evaluating
gender bias in large language models for spanish, Language Resources and Evaluation
(2022).</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>T. B. Brown</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Mann</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Ryder</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Subbiah</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Kaplan</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Dhariwal</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Neelakantan</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Shyam</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          <string-name>
            <surname>Sastry</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Askell</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Agarwal</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Herbert-Voss</surname>
            , G. Krueger,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Henighan</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Child</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Ramesh</surname>
            ,
            <given-names>D. M.</given-names>
          </string-name>
          <string-name>
            <surname>Ziegler</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Wu</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Winter</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Hesse</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Chen</surname>
          </string-name>
          , E. Sigler, M. Litwin,
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>M.</given-names>
            <surname>Barlow</surname>
          </string-name>
          ,
          <article-title>Michael stubbs. text and corpus analysis: Computer-assisted studies of language and culture</article-title>
          ,
          <source>International Journal of Corpus Linguistics</source>
          <volume>3</volume>
          (
          <year>1998</year>
          )
          <fpage>319</fpage>
          -
          <lpage>327</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>R.</given-names>
            <surname>Anil</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. M.</given-names>
            <surname>Dai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Firat</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Johnson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Lepikhin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. T.</given-names>
            <surname>Passos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Shakeri</surname>
          </string-name>
          , E. Taropa,
          <string-name>
            <given-names>P.</given-names>
            <surname>Bailey</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Chen</surname>
          </string-name>
          , et al.,
          <source>Palm 2 technical report, ArXiv abs/2305</source>
          .10403 (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>H.</given-names>
            <surname>Touvron</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Lavril</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Izacard</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Martinet</surname>
          </string-name>
          , M.
          <article-title>-</article-title>
          <string-name>
            <surname>A. Lachaux</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Lacroix</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Rozière</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Goyal</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          <string-name>
            <surname>Hambro</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Azhar</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Rodriguez</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Joulin</surname>
          </string-name>
          , E. Grave, G. Lample, LLaMA: Open and Eficient Foundation Language Models,
          <source>ArXiv abs/2302</source>
          .13971 (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [19]
          <string-name>
            <surname>W.-L. Chiang</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          <string-name>
            <surname>Lin</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>Sheng</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          <string-name>
            <surname>Wu</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Zhang</surname>
            , L. Zheng,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Zhuang</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>Zhuang</surname>
            ,
            <given-names>J. E.</given-names>
          </string-name>
          <string-name>
            <surname>Gonzalez</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          <string-name>
            <surname>Stoica</surname>
            ,
            <given-names>E. P.</given-names>
          </string-name>
          <string-name>
            <surname>Xing</surname>
            ,
            <given-names>Vicuna:</given-names>
          </string-name>
          <article-title>An open-source chatbot impressing gpt-4 with 90%* chatgpt quality</article-title>
          ,
          <year>2023</year>
          . URL: https://lmsys.org/blog/2023-03-30-vicuna/.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>T. L.</given-names>
            <surname>Scao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Fan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Akiki</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.-J.</given-names>
            <surname>Pavlick</surname>
          </string-name>
          , S. Ili'c, D. Hesslow, R. Castagn'e,
          <string-name>
            <given-names>A. S.</given-names>
            <surname>Luccioni</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Yvon</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Gallé</surname>
          </string-name>
          , et al.,
          <article-title>Bloom: A 176b-parameter open-access multilingual language model</article-title>
          ,
          <source>ArXiv abs/2211</source>
          .05100 (
          <year>2022</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>S.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , S. Roller,
          <string-name>
            <given-names>N.</given-names>
            <surname>Goyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Artetxe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Dewan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Diab</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X. V.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Mihaylov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ott</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Shleifer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Shuster</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Simig</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. S.</given-names>
            <surname>Koura</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Sridhar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Zettlemoyer</surname>
          </string-name>
          , Opt:
          <article-title>Open pre-trained transformer language models</article-title>
          ,
          <source>ArXiv abs/2205</source>
          .01068 (
          <year>2022</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>X. V.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Mihaylov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Artetxe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Simig</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ott</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Goyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Bhosale</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Du</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Pasunuru</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Shleifer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. S.</given-names>
            <surname>Koura</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Chaudhary</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B. O</given-names>
            <surname>'Horo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Zettlemoyer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Kozareva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. T.</given-names>
            <surname>Diab</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Stoyanov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <article-title>Few-shot learning with multilingual language models</article-title>
          ,
          <source>CoRR abs/2112</source>
          .10668 (
          <year>2021</year>
          ). URL: https://arxiv.org/abs/2112.10668. arXiv:
          <volume>2112</volume>
          .
          <fpage>10668</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>E.</given-names>
            <surname>Almazrouei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Alobeidli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Alshamsi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Cappelli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Cojocaru</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Debbah</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Gofinet</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Heslow</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Launay</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Malartic</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Noune</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Pannier</surname>
          </string-name>
          , G. Penedo, Falcon-40B:
          <article-title>an open large language model with state-of-the-art performance (</article-title>
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>I.</given-names>
            <surname>Garrido-Muñoz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Montejo-Ráez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Martínez-Santiago</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. A.</given-names>
            <surname>Ureña-López</surname>
          </string-name>
          ,
          <article-title>A survey on bias in deep nlp</article-title>
          ,
          <source>Applied Sciences</source>
          <volume>11</volume>
          (
          <year>2021</year>
          ). URL: https://www.mdpi.com/2076-3417/11/7 /3184. doi:
          <volume>10</volume>
          .3390/app11073184.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>K.</given-names>
            <surname>Ramesh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Sitaram</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Choudhury</surname>
          </string-name>
          ,
          <article-title>Fairness in Language Models Beyond English: Gaps and Challenges, in: Findings of the Association for Computational Linguistics: EACL 2023, Association for Computational Linguistics</article-title>
          , Dubrovnik, Croatia,
          <year>2023</year>
          , pp.
          <fpage>2106</fpage>
          -
          <lpage>2119</lpage>
          . URL: https://aclanthology.org/
          <year>2023</year>
          .findings-eacl.
          <volume>157</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>T.</given-names>
            <surname>Bolukbasi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. Y.</given-names>
            <surname>Zou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Saligrama</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Kalai</surname>
          </string-name>
          , Man is to Computer Programmer as Woman is to Homemaker? Debiasing Word Embeddings,
          <source>CoRR abs/1607</source>
          .06520 (
          <year>2016</year>
          ). URL: http://arxiv.org/abs/1607.06520. arXiv:
          <volume>1607</volume>
          .
          <fpage>06520</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [27]
          <string-name>
            <given-names>J.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Yatskar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Ordonez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.-W.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <article-title>Gender Bias in Coreference Resolution: Evaluation and Debiasing Methods</article-title>
          , arXiv e-prints (
          <year>2018</year>
          ) arXiv:
          <year>1804</year>
          .06876. arXiv:
          <year>1804</year>
          .06876.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [28]
          <string-name>
            <given-names>A.</given-names>
            <surname>Caliskan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. J.</given-names>
            <surname>Bryson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Narayanan</surname>
          </string-name>
          ,
          <article-title>Semantics derived automatically from language corpora contain human-like biases</article-title>
          ,
          <source>Science</source>
          <volume>356</volume>
          (
          <year>2017</year>
          )
          <fpage>183</fpage>
          -
          <lpage>186</lpage>
          . URL: https: //www.science.org/doi/abs/10.1126/science.aal4230. doi:
          <volume>10</volume>
          .1126/science.aal4230. arXiv:https://www.science.org/doi/pdf/10.1126/science.aal4230.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [29]
          <string-name>
            <given-names>S.</given-names>
            <surname>Dev</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Phillips</surname>
          </string-name>
          , V. Srikumar,
          <article-title>OSCaR: Orthogonal subspace correction and rectification of biases in word embeddings</article-title>
          ,
          <source>in: Proceedings of the 2021 Conference</source>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>