<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Looks to Essence: A Shift in Perspective with Physical Appearance Debiasing</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Shivatmica Murgai</string-name>
          <email>shivatmica@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="editor">
          <string-name>Gender, WEAT Score, De-bias</string-name>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Artificial Intelligence</institution>
          ,
          <addr-line>Machine Learning, Natural Language Processing, Bias, Physical appearance</addr-line>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>The International School of Bangalore (TISB)</institution>
          ,
          <country country="IN">India</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2023</year>
      </pub-date>
      <fpage>6</fpage>
      <lpage>7</lpage>
      <abstract>
        <p>As Natural Language Processing (NLP) gains popularity due to its powerful nature, it is crucial to consider its role in promoting stereotypes and society's implicit bias. While previous research addresses mitigating bias for gender or race, this study considers bias due to physical appearance in current text corpora, where algorithms mimic human prejudice from trained data. The methods to detect implicit bias in word embeddings include cosine similarity and analogies to evaluate a hard de-biasing approach. The objective of this research is to extend the existing bias detection and mitigation methods to physical appearance in text corpora to promote fairness, inclusivity, and reduce perpetuating stereotypes. To quantify the de-biasing approach, the Word Embedding Association Test (WEAT) score compares the original and de-biased word embeddings on text8 corpus. The proposed alternatives improve mitigating physical appearance bias on text8 corpora by 7.14%. The paper aims to make valuable contributions to the continuous eforts to promote ethical AI applications.</p>
      </abstract>
      <kwd-group>
        <kwd>Debiasing</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>Artificial Intelligence (AI) and Machine Learning (ML) have the potential to revolutionize the
world and have proven to do so in the past. Its algorithms can serve a wide variety of purposes
given enough data, but they are ingrained with a level of bias from training of of biased or
skewed data which does not represent the population accurately.</p>
      <p>Bias in natural language processing (NLP) models has raised significant ethical concerns
and challenges the fairness of generated text. Biased language models not only reflect but
also reinforce societal biases, leading to biased outcomes in various applications. They tend to
discriminate certain groups, specifically minorities, leading to unfairness in key areas such as
employment, healthcare and education.</p>
      <p>
        Stereotypes associated with gender are deemed dangerous when they restrict people to certain
professions, decisions, or afect their overall well-being. In children’s literature, stereotypes with
respect to gender [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], race [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ], [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ], [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ], body image [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ], and such are easily accepted and,
extensively, internalized throughout society. The stereotypes or general messages conveyed
through children’s books are integral to a developing child’s mindset.
      </p>
      <p>
        Discrimination against physical appearance can profoundly impact individuals’ mental health
and well-being, as well as their employment prospects [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ], [15]. Derogatory or negative
judgments lead to low self-esteem, body image issues, or even depression. The relentless
pressure to conform to societal beauty standards often results in eating disorders or risky
cosmetic procedures. This causes marginalization, perpetuating a spiral of inequality and
reduced economic prospects for those targeted. To combat this bias, promoting an inclusive
society to recognize the value of diversity, while prioritizing character over physical appearance,
is of vital importance.
      </p>
      <p>
        The Implicit Association Test (IAT) [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] measures the strength of associations between concepts
and evaluations or stereotypes to reveal individuals’ hidden or subconscious biases. The test
involves categorizing stimuli, such as words or images, into diferent categories such as “good”
or “bad,” for example. The IAT test records the speed and accuracy of participants’ responses to
determine their implicit associations and subconscious bias.
      </p>
      <p>
        Often, AI algorithms amplify prejudice and reinforce particular stereotypes, rather than
correcting bias. This can be extremely harmful where AI is prominent in highly autonomous
ifelds, such as applications in interview scanning, facial recognition, and the Correctional
Ofender Management Profiling for Alternative Sanctions (COMPAS) algorithm [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ].
      </p>
      <p>
        The main contributions of this research paper are:
1. We explore the use of cosine similarity to uncover implicit biases related to physical
appearance and socioeconomic status.
2. We extend the existing work of de-biasing techniques with gender and race bias into the
realm of physical appearance bias and derive their respective mathematical relations.
3. We use WEAT score [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] for bias quantification. WEAT score builds upon cosine similarity
by providing a structured framework for analyzing and quantifying associations and biases
in word embeddings, especially in the context of social attributes. The WEAT [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] score
mimicks the structure of IAT [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] by comparing associations of opposing characteristics
with sets of words like professions or adjectives with positive or negative connotations.
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. Literature Review</title>
      <p>
        Bolukbasi et al. [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] introduced cosine similarity to quantify gender bias, indicating biases in the
embeddings, which are illustrated through diferent cosine similarity values from the word pairs.
It compares the similarity between diferent word pairs to measure the association between
nouns like occupations and a specific gender. Our research paper applies the cosine similarity
method to physical appearance bias and will be discussed in Section III of this paper.
      </p>
      <p>
        Kim et al [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] asserts that previous work on debiasing word embeddings has largely focused
on individual and independent social categories, such as solely gender or race, and de-biases
embeddings for this specific singular subspace. This is highly ineficient though since
realworld corpora often presents multiple social categories, intersecting with each other. This
paper proposes techniques to debias word embeddings for multiple social categories, where
individual biases intersect non-trivially, known as intersectional bias. People part of several
possibly marginalized social categories may experience bias unique to their intersection of social
identities; the experiences and discrimination an African American woman would face vary
vastly from that of an African American man or a white woman. An intersectional subspace is
constructed to debias embeddings for multiple social categories using nonlinear geometry for
individual biases, additionally supported by empirical evaluations. In our research, we extend
the debias technique for physical appearance bias and discuss further details in Section III.
      </p>
      <p>
        The authors in [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] provide a comprehensive survey of more than 300 papers on gender bias in
NLP and describe the limitations of current approaches. Most of the studies have used English,
Chinese and Spanish languages, which lacks a holistic view of bias in NLP. Lack of bias testing
in various NLP models is another issue, which leads to ethical and societal detriment.
      </p>
      <p>In this research paper, we start by discussing crucial aspects of using text corpus datasets and
word embeddings. The paper then delves into physical appearance detection and proposes a
de-bias method, extended from previous research regarding gender or race. The paper concludes
with empirical evidence of these findings and outlines potential directions for future research.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Methodology</title>
      <sec id="sec-3-1">
        <title>3.1. Word Embeddings &amp; Dataset (Text Corpus)</title>
        <p>
          Word embeddings, with common algorithms like GloVe [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ] or Word2Vec [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ], are vector
representations of words where similar words will be mapped closer together in an embedding space.
GloVe considers words’ co-occurrences over an entire corpus and the embeddings relate to the
probability that two words are used together in diferent contexts. However, Word2Vec uses
the meaning of words in local context (and also has faster computation), which is much more
useful for recognizing bias. Word2vec embeddings, for example, create these vectors based of
of previous text’s data.
        </p>
        <p>
          The author experimented with Text8 [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ] to train the word embeddings. Text8 is extensively
used in NLP to train language models. One of the key advantages of the Text8 dataset is that it
has gone through significant text preprocessing to remove unnecessary formatting, punctuation,
and other noise, making it suitable for training word embedding models like Word2Vec or GloVe.
These models learn to represent words as vectors based on the co-occurrence patterns of words
in the dataset. It provides highly compact word vectors while maintaining a wide range of
words and contexts.
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Bias Detection Metrics</title>
        <sec id="sec-3-2-1">
          <title>3.2.1. Cosine similarity: Metric to determine two words’ relation</title>
          <p>
            Cosine similarity is a metric to determine how closely related or similar two words are, whose
range is [
            <xref ref-type="bibr" rid="ref1">−1, 1</xref>
            ] according to the range of the cosine function. Cosine similarity, between the
words  and  with  and  as their respective word embeddings, is defined as
Cosine Similarity(, ) =
 • 
‖‖‖ ‖
(1)
where  •  is the dot product of the two vector word embeddings and ‖‖ is the norm or the
length of ‖‖ .
          </p>
          <p>The similarity between two vectors is defined by the magnitude of the angle between them,  .
Additionally, the cosine of an angle is closest to 1 for the smallest angles, which implies greater
similarity, and decreases as the magnitude of the angle increases, signifying smaller correlation.
Conventionally, we’ll consider that diferences in cosine similarity scores are reflective of word
associations in the trained text corpus, which therefore portrays society’s bias. This metric
can be employed to detect implicit bias relevant to gender, repulsion towards certain physical
appearance, or even socioeconomic backgrounds.</p>
          <p>This strengthens the findings of Harvard’s Implicit Association Test (IAT), where people were
more hesitant to match positive descriptives with negatively assumed characteristics.</p>
          <p>Several words are plotted in Figure 1 using cosine similarity score between two words’
respective vector embeddings. It depicts society’s perceptions of positive and negative physical
characteristics as well as unfair stereotypes or assumptions associated with these words. The
lengths of the edges connecting diferent nodes are inversely proportional to the words’
similarity. The distance between two nodes represents the correlation between two words, where a
smaller distance connotes a greater cosine similarity score.</p>
          <p>For example, there are evident instances of bias with these correlations:
• “attractive”, “athletic”, “healthy”, and “confident” are both closer to “fit” than “fat”
• “confident” is closer to “beautiful” than “ugly”
• “lazy” is closer to “overweight” than “underweight”
• “acne” is closer to “ugly” than “attractive”, “confident”, or “beautiful”
• “tall” is closer to “beautiful” than “short.”
For the above observations, if a word is relatively closer to a word than some other, it’s considered
to have a higher cosine similarity score and the two words therefore are more associated
according to society’s perceptions.</p>
        </sec>
        <sec id="sec-3-2-2">
          <title>3.2.2. Analogies</title>
          <p>
            In addition to the cosine similarity metric, analogies ensure that the relationship [
            <xref ref-type="bibr" rid="ref3">3</xref>
            ]
between two pairs of words is nearly equivalent. A typical example of an analogy would be
man : woman :: king : queen. For example, in the analogy  1 ∶  2 ∶∶  1 ∶  2, the relationship
between  1 and  2 that is the same as the relationship between  1 and  2. The relationship
between two words is defined by subtracting the words’ respective word embeddings,
word_to_vec_map[ 1] − word_to_vec_map[ 2].
          </p>
          <p>With this definition, the relationship between two words can be examined in contrast with
the relationship of another two words. An “inverse” relationship would be where  1 −  2 is
equivalent to  2 −  1 since the order is reversed of one side of the analogy.</p>
          <p>Efectively, the relationship between  1 and  2 is examined and a word  2 is found such that
the relationship between  2 and  1 is identical to the former. The initial brute force method, by
considering the word that provides the closest relationship with the other pair of words, did
not hold to be as efective if the three provided words are not related.</p>
          <p>
            For instance, in our experiments on Text8 corpus [
            <xref ref-type="bibr" rid="ref7">7</xref>
            ], analogies(“black”, “white”, “up”) outputs
an entirely irrelevant word. Consequently, the “most_similar” function from word2vec more
efectively comprehended the relationship between  1 and  2.
          </p>
          <p>The output of analogies (“black”, “white”, “up”) was “down,” as desired. Although this seems
to be more accurate than the brute force method, it still includes a large amount of bias in the
context of gender. The model succumbs to the famous example of gender bias,
woman : man :: homemaker : ___
and outputs “machinist.”</p>
          <p>For bias related to physical appearance, these embeddings are dominated by stereotypes
associating negative or positive adjectives with society’s assumptions of these characteristics.
For instance, given the first three words of the following analogy, it provides the word “weak”,
‘big : strong :: thin : weak’.</p>
          <p>
            It is evident from our empirical findings that these vectors contain an ingrained notion of
bias. The paper now discusses ways to reduce bias with gender or physical appearance by hard
de-biasing the word embeddings. We would leverage the approach suggested by Kim et al [
            <xref ref-type="bibr" rid="ref4">4</xref>
            ]
to the physical appearance realm. This includes neutralizing or equalizing the vectors to fit an
idea of gender equality or removing discrimination based on physical appearance for words
that, in an ideal world, should not be correlated with specific stereotypical characteristics.
          </p>
        </sec>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Generalizing de-biasing technique for an arbitrary subspace</title>
        <p>
          A subspace can be defined as the relationship between two opposing adjectives, similar to
how the gender subspace is defined as ”man” − ”woman” in [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ]. The diference between the
vectors that map to these antonyms defines the subspace, upon which the vectors are projected
onto. The original word’s vector is the word embedding and the “biased” component is the
projection of the original word’s embedding onto the biased axis of gender. The projection from
the original word can be subtracted to obtain the unbiased vector.
        </p>
        <p>This biased axis is defined as the diference between the vectors corresponding to antonyms
representing the subspace. The subspace encapsulates the idea or direction of a specific category
such as gender, race, or a physical characteristic.</p>
        <p>The projection of the word’s vector w onto the biased axis can be calculated as
w</p>
        <p>w • subspace
= ‖subspace‖2 ⋅ subspace
where  is the word’s embedding, subspace is the biased axis, and w is the word’s
biased component. Since hard de-biasing entirely removes this biased component, in a way, the
definitions of the word embeddings can be updated to obtain
(2)
(3)
Algorithm 1 Debias (   ,  ,   
1:  ←    _ _  _[  ]
2:  _ ← (Dot product of  &amp;  ) ⋅  /(
3:  _ ←  −  _
4: return  _
w</p>
        <p>= w − w
_ _  _
)
norm of gender)2</p>
        <p>The amount of bias in a particular subspace can be compared by using the cosine similarity
metric with the vector representing the subspace (i.e. defined as  −  ). A negative score implies
a word’s greater association with  , and if it is positive, then  .</p>
        <p>Additionally, the magnitude of the similarity with this subspace is irrelevant to understanding
which word is more closely related with  instead of  (very small positive values do not
necessarily indicate they are skewed towards the word  , but rather comment upon correlation
with the subspace vector, since the diference between  and  is considered). In other words, 1
does not entirely represent associated with  , and neither does −1 with  .</p>
      </sec>
      <sec id="sec-3-4">
        <title>3.4. Equalization for an arbitrary subspace</title>
        <p>Equalization applies to words equivalent in meaning but difering with respect to the
corresponding subspace. Both these words should be equidistant to each category, to alter how a
word is sometimes ingrained with bias relevant to physical appearance or gender.</p>
        <p>
          Debiased vectors remove any notion of the subspace from them and the resulting vector
word embedding from these words solely exists on the axis with the other  − 1 dimensions,
excluding this biased axis. Equalization, whose techniques are discussed in [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ], ensures that
any word specific to a subspace’s category, such as a gender-specific word, is equidistant from
this axis of  − 1 dimensions.
        </p>
        <p>After equalizing, the only diference between the two words is their direction on the biased
axis. The stem of these words is the same since excluding the projection onto the biased axis
(or the biased component) of each vector should result in the same vector component. Since the
words’ biased components are in opposite directions, the vectors’ sum consists of no biased
components because the addition causes it to cancel out. By taking the average of the two
opposite vectors, the unbiased component is obtained.</p>
        <p>Algorithm 2 Equalize( _ _   ,  _ ,    _ _  _ )
1:  1,  2 ←    _ _  _[ _ _  ]
2:  ← Average of word vectors  1 and  2
3:  _ ← Biased component of  (using projection)
4:  _ ←  −  _
5:  1 _,  2 _ ← Biased component of</p>
        <p>(using projection of the vectors onto the biased axis)
6:  _ 1 _,  _ 2 _ ← Biased components
used to equalize both vectors according to equations for
eq_w1_biased and eq_w2_biased
7: return  _ 1 _ +  _ ,  _ 2 _ +  _
// combines equalized components with the unbiased component to get
equalized word vectors for  1 and  2 below</p>
        <p>Let  1 and  2 be two words opposite in the biased axis. The biased component of their
average,  =  1 +2 2 , can be obtained, and then subtracted from  to obtain the unbiased vector.
The unbiased and biased components of  are calculated by replacing the subspace projected
onto earlier (while de-biasing) with the biased axis.</p>
        <p>Both words,  1 and  2 are now projected onto the biased axis, once again subspace with the
biased axis. Finally, the biased components of the two words’ vector embeddings are obtained
to be the following:
eq_w1_biased
eq_w2_biased
= √|1 − ||x
= √|1 − ||x
||2| ⋅
||2| ⋅</p>
        <p>eq1 
||(eq_ 1 − x</p>
        <p>eq2
||(eq_ 2 − x
− x
− x
) − x
) − x
||
||
,
.</p>
        <p>
          We have extended the concept provided for gender bias in Bolukbasi at el. [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ] to physical
appearance bias in the above equations. To find the equalized word embeddings, the above
biased components are added to x . Therefore, the following is true
eq_w1 = eq_w1_biased + x
eq_w2 = eq_w2_biased + x
,
.
        </p>
        <p>By equalizing, both words,  and  are become equidistant to the vector representing the
subspace, which means that 1 and −1 would represent the two words that make up the subspace.
Neutralization prevents words skewing to one category of the subspace specifically, while
ensuring that the embeddings retain their original meaning.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Experiments &amp; Results</title>
      <sec id="sec-4-1">
        <title>4.1. Equalization example for gender bias</title>
        <p>This is necessary since gender-specific words, such as “hero” and “heroine,” are not equidistant
from the biased axis of gender.</p>
        <p>The subspace is defined as</p>
        <p>gender = word_to_vec_map["she"] − word_to_vec_map["he"].
(The word “man” was not used since it is more commonly used to connote a person such as in
“mankind” instead of a male.)</p>
        <p>Before equalizing
Before equalizing
After equalizing
After equalizing</p>
        <p>Gender
“he”
“she”
“he”
“she”</p>
        <p>Cosine Similarity score with gender
−0.23826271 ≈ −0.24</p>
        <p>0.62127906 ≈ 0.62
−0.62998897 ≈ −0.63
0.62998885 ≈ 0.63</p>
        <p>The subspace is defined as the gender vector to equalize upon. As mentioned earlier, −1 does
not entirely correlate to men or 1 to women, as shown in Table I.
(4)
(5)
(6)
(7)</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. De-biasing Word Embeddings for Physical Appearance</title>
        <p>Physical appearance is a broad category which may include various definitions of the
subspace to debias upon. For instance, bias due to weight, appearance, and attractiveness, with
their associated societal stereotypes can be explored. Word 1 and Word 2 are antonyms to
describe a single subspace, defining a category of physical appearance such as height, weight,
or complexion, for example.</p>
        <p>In essence, these distinct subspaces could de-bias word embeddings individually, and cosine
similarity is calculated again using Word 2’s de-biased word embedding. This updated cosine
similarity, between the Word 1 and the de-biased word embedding of Word 2, is calculated and
these results are shown in the rightmost column of Table 2, where de-biasing projects the vector
onto each subspace and subtracts the resulting biased component. The table depicts a subset of
the words considered for this research paper.</p>
        <p>The subspace in each scenario is defined as the vector corresponding to the diference between
the two antonyms like Word 2 − Word 1. For instance, the cosine similarity score of the word
“actor” with the subspace weight (defined as “tall” − “short”) is 0.227, which is reduced to
2.71 ⋅ 10−8. This decrease elucidates the efectiveness of de-biasing.</p>
        <p>However, this would be an extremely targeted, and consequently limited, approach, which
is not scalable. This includes several pairs of positive and negative characteristics based on
society’s inherently flawed standards and expectations, as well as common stereotypes, jobs,
and assumptions for these characteristics. A single subspace is instead constructed from
individual subspaces (defined as the diference of antonyms) by averaging the vectors representing
subspaces for diferent categories of physical characteristics.</p>
      </sec>
      <sec id="sec-4-3">
        <title>4.3. Identification of biased stereotypes and characteristics</title>
        <p>Cosine similarity scores and analogies were especially beneficial to highlight words that
presented profound associations with certain attributes of physical appearance, such as in Table 2.
If a word such as “beautiful” has a positive correlation with the vector representing height, or
more generally, physical appearance, it is evident this conveys some idea of bias. The examples
shown in Table 2 exemplify humans’ pyschological biases towards people’s inborne physical
characteristics; this prejudice inevitably leaks into algorithms trained of of this biased data.
Some words may exhibit strong positive correlations with favorable or less favorable physical
characteristics, allowing for their categorization.</p>
      </sec>
      <sec id="sec-4-4">
        <title>4.4. WEAT Score</title>
        <p>
          The Word Embedding Association Test (WEAT) score [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] is a measure of the association between
sets of words based on their word embeddings. It is commonly used to evaluate the presence of
biased associations in word embeddings. A positive WEAT score between two sets of words
indicates greater association, while a negative score indicates weaker association.
        </p>
        <p>The original WEAT score on the text8 corpus was 0.35, indicating the existence of physical
appearance bias in word embeddings. After employing debiasing techniques, WEAT score
decreased to 0.325, improving by 7.14%, quantifying average eficacy of this approach in mitigating
physical appearance bias in word embeddings.</p>
        <p>Word 1</p>
        <p>Fat</p>
        <p>Fit
Overweight
Underweight</p>
        <p>Tall
Short</p>
        <p>Pimples
Unblemished</p>
        <p>Ugly
Pretty</p>
        <p>Villager
Metropolitan</p>
        <p>Bald</p>
        <p>Word 2</p>
        <p>Lazy</p>
        <p>Lazy
Healthy
Healthy
Beautiful
Beautiful
Attractive
Attractive</p>
        <p>Kind</p>
        <p>Kind
Uneducated
Uneducated
Attractive</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion</title>
      <p>The research successfully identified instances of physical appearance bias in word embeddings
using cosine similarity and analogy methods. The de-biasing approach showed promising
results in mitigating physical appearance bias. Maintaining a fine balance between retaining
contextual information while mitigating biases is a challenging but crucial aspect in achieving
efective debiasing results. Leveraging WEAT score, the magnitude of such biases was quantified,
providing a meaningful metric to assess the level of bias reduction achieved through our
debiasing eforts. After debiasing certain biased words from the corpus, WEAT score improved
by 7.14%. The proposed de-biasing approach showed promising results in reducing bias but
there are challenges in achieving complete de-biased dataset. As AI continues to evolve, future
research has great potential to efectively transform the world into a socially responsible
AIpowered society.</p>
    </sec>
    <sec id="sec-6">
      <title>6. Future Work</title>
      <p>This study only focused on debiasing certain identified biased words. This could be extended to
identify and mitigate bias in the entire text8 or Google News corpora. Instead of considering
individual subspaces, the diferent subspaces defined within the category of physical appearance
are correlated in potentially a non-linear relation. Futher research could examine the
relationships between these words to further improve debiasing methodology and create a singular
subspace encapsulating society’s standards of physical appearance.</p>
    </sec>
    <sec id="sec-7">
      <title>7. Acknowledgements</title>
      <p>The author would like to sincerely thank Mr. Catalin Voss, CTO and Co-founder of Ello
(helloello.com), and Prof. Nick Haber, Assistant Professor at Stanford University, for their
review feedback, continuous support, and guidance throughout the research study.
[15] Perceived Discrimination and Physical, Cognitive, and Emotional Health in Older
Adulthood; Sutin, Angelina, Stephan, Yannick Carretta, Henry, Terracciano, Antonio 2014/03/01;
American Journal of Geriatric Psychiatry
[16] Elisa Bassignana, Dominique Brunato, Marco Polignano, Alan Ramponi, Preface to the
Seventh Workshop on Natural Language for Artificial Intelligence (NL4AI), in: Proceedings
of the Seventh Workshop on Natural Language for Artificial Intelligence (NL4AI 2023)
co-located with 22th International Conference of the Italian Association for Artificial
Intelligence (AI* IA 2023), 2023.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1] Loyola Marymount University. “
          <string-name>
            <surname>Test Your Implicit Bias - Implicit Association Test (IAT) - Loyola Marymount</surname>
          </string-name>
          University.” Lmu.edu,
          <year>2023</year>
          , resources.lmu.edu/dei/initiativesprograms/implicitbiasinitiative, Accessed 17
          <string-name>
            <surname>June</surname>
          </string-name>
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>Machine</given-names>
            <surname>Bias</surname>
          </string-name>
          :
          <article-title>There's software used across the country to predict future criminals. And it's biased against blacks; Angwin, Julia and Larson, Jef and Mattu, Surya</article-title>
          and Kirchner, Lauren,
          <year>2016</year>
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <article-title>[3] Man is to Computer Programmer as Woman is to Homemaker, Debiasing Word Embeddings, Tolga Bolukbasi and Kai-Wei Chang and James Zou and Venkatesh Saligrama</article-title>
          and
          <string-name>
            <given-names>Adam</given-names>
            <surname>Kalai</surname>
          </string-name>
          ,
          <year>2016</year>
          , https://doi.org/10.48550/arXiv.1607.06520
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>Nayoung</given-names>
            <surname>Kim</surname>
          </string-name>
          ,
          <string-name>
            <surname>Huan</surname>
            <given-names>L</given-names>
          </string-name>
          ,
          <source>Proceedings of the 29th International Conference on Computational Linguistics, October 12-17</source>
          ,
          <year>2022</year>
          https://aclanthology.org/
          <year>2022</year>
          .coling-
          <volume>1</volume>
          .110.pdf
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>A</given-names>
            <surname>Survey on</surname>
          </string-name>
          <article-title>Gender Bias in Natural Language Processing</article-title>
          KAROLINA STAŃCZAK, University of Copenhagen ISABELLE AUGENSTEIN, University of Copenhagen https://arxiv.org/pdf/2112.14168.pdf
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>Caliskan</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bryson</surname>
            ,
            <given-names>J. J.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Narayanan</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          (
          <year>2017</year>
          ).
          <article-title>Semantics derived automatically from language corpora contain human-like biases</article-title>
          .
          <source>Science</source>
          ,
          <volume>356</volume>
          (
          <issue>6334</issue>
          ),
          <fpage>183</fpage>
          -
          <lpage>186</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>Mahoney</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <article-title>Text8</article-title>
          <string-name>
            <surname>Corpus</surname>
          </string-name>
          ,
          <year>2006</year>
          . https://mattmahoney.net/dc/text8.zip
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <surname>Mikolov</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Corrado</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Dean</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <article-title>Distributed Representations of Words and Phrases and their</article-title>
          <string-name>
            <surname>Compositionality</surname>
          </string-name>
          ,
          <year>2013</year>
          .
          <source>Proceedings of the 26th International Conference on Neural Information Processing Systems (NIPS</source>
          <year>2013</year>
          ), pp.
          <fpage>3111</fpage>
          -
          <lpage>3119</lpage>
          . https://- code.google.com/archive/p/word2vec/
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          <article-title>[9] GloVe: Global Vectors for Word Representation</article-title>
          , Pennington, Jefrey and Socher, Richard and Manning,
          <string-name>
            <surname>Christopher</surname>
            <given-names>D</given-names>
          </string-name>
          ,
          <source>Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP)</source>
          ,
          <year>2014</year>
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <article-title>Distributed Representations of Words and Phrases and their Compositionality; Mikolov, Tomas and Sutskever, Ilya and Chen, Kai and Corrado, Greg and Dean</article-title>
          , Jef,
          <year>2013</year>
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>Media</surname>
          </string-name>
          <article-title>'s Influence on Body Image Disturbance and Eating Disorders: We've Reviled Them</article-title>
          ,
          <string-name>
            <surname>Now Can We Rehabilitate Them? J. Kevin</surname>
            <given-names>Thompson</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>Leslie J.</given-names>
            <surname>Heinberg</surname>
          </string-name>
          ,
          <string-name>
            <surname>Martina M. Altabe</surname>
          </string-name>
          ,
          <string-name>
            <surname>Madeline M. Tantlef-Dunn</surname>
          </string-name>
          ,
          <source>Journal of Social Issues</source>
          ,
          <year>1999</year>
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>Gender</surname>
            <given-names>Stereotypes</given-names>
          </string-name>
          ,
          <source>Naomi Ellemers Annual Review of Psychology</source>
          <year>2018</year>
          69:
          <issue>1</issue>
          ,
          <fpage>275</fpage>
          -
          <lpage>298</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <article-title>The Consequences of Race for Police Oficers' Responses to Criminal Suspects, Plant</article-title>
          , Ashby, Peruche, Michelle,
          <volume>10</volume>
          .1111/j.0956-
          <fpage>7976</fpage>
          .
          <year>2005</year>
          .
          <volume>00800</volume>
          .x, Psychological science,
          <year>2005</year>
          /04/01
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>A</given-names>
            <surname>Race-Based Size</surname>
          </string-name>
          Bias for Black Adolescent Boys: Size, Innocence, and Threat; Freiburger, Erin, Sim, Mattea, Halberstadt, Amy, Hugenberg, Kurt;
          <volume>10</volume>
          .1177/01461672231167978; Personality &amp; social psychology bulletin;
          <year>2005</year>
          /04/01
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>