<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>E. Amengual-Alcover);</journal-title>
      </journal-title-group>
      <issn pub-type="ppub">1613-0073</issn>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>- a dataset of facial expressions generated by AI: development and validation</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Pablo Núñez-Pérez</string-name>
          <email>pablo200055@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Esperança Amengual-Alcover</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Maria Francesca Roig-Maimó</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ramon Mas-Sansó</string-name>
          <email>ramon.mas@uib.es</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Miquel Mascaró-Oliver</string-name>
          <email>miquel.mascaro@uib.cat</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Workshop</string-name>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>University of the Balearic Islands</institution>
          ,
          <addr-line>Carretera de Valldemossa, km 7.5, Palma de Mallorca</addr-line>
          ,
          <country country="ES">Spain</country>
        </aff>
      </contrib-group>
      <volume>000</volume>
      <fpage>0</fpage>
      <lpage>0001</lpage>
      <abstract>
        <p>This paper introduces UIBAIFED, a novel facial expression dataset designed to enhance Facial Expression Recognition (FER) by providing high-quality, realistic images labeled with detailed demographic attributes, including age group, gender, and ethnicity. Unlike existing datasets, UIBAIFED incorporates a fine-grained classification of 22 micro-expressions, based on the universal facial expressions defined by Ekman and the micro-expression taxonomy proposed by Gary Faigin. The dataset was generated using advanced difusion models and validated through a convolutional neural network (CNN), achieving an accuracy of 82% in expression classification. The results highlight the dataset's reliability and potential to improve FER systems. UIBAIFED iflls a critical gap in the field by ofering a more comprehensive labeling system, enabling future research on expression recognition across diferent demographic groups and advancing the robustness of FER models in diverse applications.</p>
      </abstract>
      <kwd-group>
        <kwd>HCI</kwd>
        <kwd>machine learning</kwd>
        <kwd>facial expression dataset</kwd>
        <kwd>FER</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Facial Expression Recognition (FER) has experienced significant advances in recent years, largely driven
by improvements in deep learning techniques [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] and the increasing availability of high-quality datasets
[
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. These datasets play a crucial role in training models that can accurately interpret facial expressions
in various contexts. However, existing datasets still present challenges related to demographic diversity,
class imbalances, and ethical concerns such as bias in representation [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. Addressing these issues is
essential for developing more robust and generalizable FER models.
      </p>
      <p>
        Despite the increasing availability of FER datasets, widely used collections such as Fer2013 [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ],
CK+ [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], RAF-DB [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], and AfectNet [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] have limitations. These include class imbalances where some
emotions, like happiness, are overrepresented, while others, such as fear or disgust, remain
underrepresented [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. Additionally, many datasets primarily feature young and Western populations, limiting the
generalization of FER models to underrepresented groups [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. To mitigate these shortcomings, previous
research has explored alternative approaches such as data augmentation techniques and synthetic facial
expressions, to improve data diversity and model performance [
        <xref ref-type="bibr" rid="ref10 ref11">10, 11</xref>
        ]. However, to the best of our
knowledge, no publicly available FER dataset has been entirely generated using AI.
      </p>
      <p>In this work, we introduce UIBAIFED (UIB Artificial Intelligence Facial Expression Dataset), the
ifrst AI-generated dataset designed to improve FER model training and evaluation. Unlike traditional
datasets, UIBAIFED uses generative AI techniques to create a diverse and balanced dataset of facial
expressions. This approach ensures a more representative training corpus for modern FER systems,
reducing demographic biases and improving overall model robustness.</p>
      <p>CEUR</p>
      <p>ceur-ws.org</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related work</title>
      <sec id="sec-2-1">
        <title>2.1. Traditional FER datasets</title>
        <p>
          Several datasets have been widely used in FER research, including FER2013, CK+, RAF-DB, and AfectNet.
These datasets have significantly contributed to the advancement of deep learning models for emotion
classification. However, they often sufer from limitations such as:
• Demographic Imbalances: Many datasets focus on younger and Western populations, resulting
in models that generalize poorly to underrepresented groups [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ].
• Class Imbalances: Some emotions, such as happiness and neutrality, are more frequently
represented than others, such as fear or disgust, which can lead to biased model performance [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ].
• Labelling Inconsistencies: Diferences in how emotions are annotated across datasets can
introduce noise and hinder model generalization [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ].
        </p>
        <p>These limitations have motivated researchers to develop new datasets that ofer more balanced and
diverse samples, ensuring better generalizability of FER models.</p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Micro-expression recognition</title>
        <p>
          Micro-expressions are brief, involuntary facial expressions that reveal suppressed emotions. Their
lfeeting nature makes them dificult to capture and classify, yet they are crucial in fields such as
psychology, security, and Human-Computer Interaction (HCI) [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ].
        </p>
        <p>
          One of the major gaps in current FER datasets is the absence of systematic labelling for
microexpressions. Unlike standard datasets that focus on broader emotional categories, micro-expressions
require finer granularity and precise annotation. This limitation hinders the development of models
capable of detecting subtle emotional cues in real-time applications [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ].
        </p>
        <p>
          Faigin’s categorization of facial expressions provides a comprehensive framework for understanding
facial dynamics beyond the traditional seven emotional categories [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ]. This taxonomy emphasizes the
complexity of expressions, capturing subtle variations that are often overlooked in conventional FER
studies. However, existing datasets rarely incorporate this level of detail, limiting the ability of current
models to recognize nuanced emotional states. Bridging this gap requires datasets explicitly designed
to align with Faigin’s categorization.
        </p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Methods</title>
      <p>To address the aforementioned challenges, in this work we introduce UIBAIFED, an AI-generated dataset
designed to provide a more balanced and diverse representation of facial expressions. By leveraging
generative models, we ensure controlled variations in age, gender, and ethnicity while maintaining
realistic diferences in pose, lighting, and expression intensity. This approach aims to mitigate biases in
traditional datasets and enhance the robustness of FER models.</p>
      <sec id="sec-3-1">
        <title>3.1. Facial models</title>
        <p>
          To ensure the quality of the dataset, the generated images adhere to the following criteria: the face
must be centred and occupy between 40% and 70% of the image area; lighting should be suficient
to clearly highlight facial expression details, while the background remains uniform and neutral to
prevent potential classification interference. Additionally, facial expressions must accurately replicate
the descriptions proposed by Gary Faigin [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ]. Furthermore, visual artefacts should be minimal and
should not compromise the expressiveness of the face.
        </p>
        <p>The UIBAIFED dataset ensures a balanced distribution across sex, five distinct age groups (see
Figure 1) and three body composition categories (see Figure 2).</p>
        <p>(a) 15
(d) 65
(b) 25
(e) 85
(a) Underweight
(b) Normal weight
(c) Overweight</p>
        <p>
          Ethnic diversity is considered based on the classification provided by the Ofice of Management
and Budget (OMB) [
          <xref ref-type="bibr" rid="ref17">17</xref>
          ], which includes groups such as Native Americans, Asians, Black individuals,
Hispanics, Native Hawaiians or other Pacific Islanders, and White individuals of European, North
African, or Middle Eastern descent (see Figure 3).
        </p>
        <p>(a) White
(b) Black
(a) Alaskan-Native
(b) Hispanic
(c) Hawaiian</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Image creation and filtering process</title>
        <p>
          For the generation of facial expression images in the UIBAIFED dataset, the Stable Difusion model [
          <xref ref-type="bibr" rid="ref18">18</xref>
          ]
was utilized. This open-source technology can be run locally, ofering the advantage of generating an
unlimited number of images. Its flexible nature and the continuous contributions from the community
have enabled the development of improved versions, enhancing the variety and quality of the results,
ensuring that the images meet the criteria established for facial expression analysis.
        </p>
        <p>
          The Stable Difusion checkpoints are pre-trained models designed to generate images from textual
descriptions. Large datasets are used to learn the correlations between words and visual elements. The
selection of a checkpoint requires considering the ability to generate a variety of images with all the
required characteristics, while also minimizing the generation time. Based on empirical findings, it has
been found that the Realistic Vision checkpoint [
          <xref ref-type="bibr" rid="ref19">19</xref>
          ] best meets the needs of the dataset.
        </p>
        <p>
          To optimize the model for facial expression generation, Low-Rank Adaptation (LoRA) [
          <xref ref-type="bibr" rid="ref20">20</xref>
          ], [
          <xref ref-type="bibr" rid="ref21">21</xref>
          ]
was employed. LoRA allows the adaptation of machine learning models to new contexts quickly by
adding lightweight components to the original model rather than modifying the entire structure. In the
case of Stable Difusion, LoRAs specifically tailored for facial expression generation were sourced from
CivitAI [
          <xref ref-type="bibr" rid="ref22">22</xref>
          ]. Table 1 depicts the LoRAs used for the generation of the UIBAIFED dataset.
        </p>
        <p>Additionally, the necessary prompts for generating the micro-expressions that make up the dataset
were developed. Out of the 33 micro-expressions described by Gary Faigin, a subset of only 22 was
successfully reproduced due to the dificulty in describing certain subtleties for generative models.</p>
        <p>An example of the generated (positive and negative) prompts is as follows:
--prompt "White Man, 15y.o, (AngryShouting:0),(angry!!), (((shouting!!!))),
&lt;lora:l\_ang\_ae\_sd\_64\_32:0.9&gt;, Underweight, ((looking at the camera)),
hyperrealistic, professional photo, studio lighting, sharp focus,
centered on the image, vertical alignment, face, plain grey background"
--negative\_prompt "((Deformed)), disfigured, hat,(artifacts in eyes, bad iris),
((artifacts in face)), hawaiian clothes, worse quality, low quality, jpeg,
pixelated, anime, ((poorly illuminated face)), red eyes, ((bad teeth)),
((body, arms, hands, legs, naked))"</p>
        <p>The prompt described above generates the image shown in Figure 4, which represents a
15-yearold male of lean build with the Anger expression, specifically the micro-expression AngryShouting,
according to Gary Faigin’s taxonomy.</p>
        <p>The structure of the diferent prompts is consistently maintained, following this format:
"Ethnicity, gender, age,&lt;description of the expression&gt;"</p>
        <p>Within the description of the expression, the reference to the LoRA is included using the following
nomenclature:</p>
        <p>&lt;lora: (LoRA name):(weight)&gt;</p>
        <p>In this structure, “weight” refers to the intensity of the expression. Certain micro-expressions are
generated using the same LoRA but with diferent descriptors. For example, the micro-expressions
NearlyCrying and Sad, both representing sadness, are generated with the following two prompts,
producing the images shown in Figure 5, while utilizing the same LoRA.</p>
        <p>--prompt "Black Woman, 25y.o, (NearlyCrying:0), ((sad mouth)), miserable face,
(sad:1.2), &lt;lora:l\_sad\_se\_sd\_64\_32:1&gt;, Overweight, ((looking at the camera)),
hyperrealistic, professional photo, studio lightning, sharp focus,
centered on the image, vertical alignment, face, plain grey background"
--prompt "Black Woman, 25y.o, (Sad:0), (sad), (melancholic face), closed lips,
small mouth, &lt;lora:l\_sad\_se\_sd\_64\_32:1&gt;, Overweight, ((looking at the camera)),
hyperrealistic, professional photo, studio lightning, sharp focus,
centered on the image, vertical alignment, face, plain grey background"
Parentheses and numerical values are used to emphasize specific words or phrases.</p>
        <p>
          Figure 6 displays the 22 expressions generated for a 15-year-old White male. An automated script
was developed to generate 3960 prompts, resulting from the combination of 2 genders, 6 ethnicities,
5 age groups, and 3 body types, all organized according to the six universal expressions according to
Ekman’s classification [
          <xref ref-type="bibr" rid="ref23">23</xref>
          ].
        </p>
        <p>
          The images corresponding to the generated prompts were produced using the Automatic111
application [
          <xref ref-type="bibr" rid="ref24">24</xref>
          ]. Due to the random nature of the image generation process, not all images are expected to be
accurate on the first attempt. Therefore, for each micro-expression, between 15 and 30 images were
generated to ensure the desired quality and consistency.
        </p>
        <p>
          The images generated using the specified prompts were manually selected based on their alignment
with the descriptions and graphical representations provided by Gary Faigin [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ]. Figure 7 illustrates
the manual matching process for the micro-expression SlySmile. The left image represents the generated
expression from the UIBAIFED dataset, while the right image corresponds to the reference illustration
from Gary Faingin’s work. The selection process ensured that each image accurately represented the
intended facial expression and adhered to the established criteria.
        </p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. The UIBAIFED dataset</title>
      <p>The total number of images in the dataset is 2948. Images generated for prompts with diferent body
types were removed due to the minimal diferences observed between those labelled as Normal Weight
and Underweight. A greater number of representations were retained for more complex expressions,
leading to the distribution of images per micro-expression shown in Table 2.</p>
      <p>
        The database is organized into folders, each containing images with a resolution of 512×512 pixels.
There is one folder for each of the six universal expressions according to Ekman’s classification [
        <xref ref-type="bibr" rid="ref23">23</xref>
        ]. It
is important to note that the seventh expression, Contempt, which Ekman later added to his original
classification, is labelled in our dataset as the micro-expression Disdain. This expression is included in
the Disgust folder, following Gary Faigin’s classification approach.
      </p>
      <p>Within each folder, the images are named according to the following format:</p>
      <p>Num\_ethnicity\_gender\_age\_microexpression.png
“Num” represents the generation number assigned by Stable Difusion and indicates the order of the
images within each folder. The images are organized first by micro-expression, followed by ethnicity,
gender, and age.</p>
    </sec>
    <sec id="sec-5">
      <title>5. UIBAIFED validation</title>
      <p>To initially validate the UIBAIFED dataset, a simple Convolutional Neural Network (CNN) model was
employed for facial expression classification. The model takes grayscale images of size 128 ×128 pixels as
input, which are processed through three convolutional layers. These layers are followed by a Rectified
Linear Unit (ReLU) layer and a max-pooling layer to extract key features. The architecture also includes
four Fully Connected (FC) layers, which are used to classify the facial expressions into one of the 22
target micro-expressions described in the dataset. The overall network structure is shown in Figure 8.</p>
      <p>To enhance generalization and prevent overfitting, a Dropout layer is applied between the fully
connected layers. This dropout technique helps the network learn more robust features by randomly
dropping units during training, which improves the model’s ability to generalize to unseen data.</p>
      <p>The dataset is split into training and test sets, with 67% of the data used for training and 33% for
testing. The data distribution is balanced based solely on micro-expression types, and other factors
such as gender, body type, ethnicity, and age are not considered in this validation step. These factors
will be explored in future studies.</p>
    </sec>
    <sec id="sec-6">
      <title>6. Results</title>
      <p>After completing the training process, a Loss value close to 0.5 and an overall Accuracy of 82% were
achieved. These results were obtained using 67% of the images for training, 1975 images in total. Figure 9
shows the evolution of these values as a function of the training epochs.</p>
      <p>The trained CNN model was tested with the test dataset (formed by 1975 images), achieving an
overall accuracy of 85,71%. The resulting confusion matrix is presented in Figure 10, while Table 3
details the performance metrics for each of the 22 micro-expressions. Additionally, Table 4 provides a
summary of the overall classification metrics.</p>
      <p>The test results indicate that the CNN model has successfully learned and generalized most facial
expressions in a validation set of over 900 images that were not used during training.</p>
      <p>Most facial expressions achieve an accuracy above 75%. However, expressions related to Joy present
greater classification challenges. Specifically, AbashedSmile is sometimes misclassified as Sad or Worried,
while FalseLaughter1 is frequently confused with CryingOpenMouth. This misclassification likely occurs
because some samples of FalseLaughter1 include eyes that are suficiently closed, making them visually
similar to CryingOpenMouth.</p>
      <p>A recurring pattern observed across all training-test cycles is the confusion between FalseLaughter2
and UproariousLaughter. The primary dificulty in distinguishing these expressions lies in their strong
resemblance. Both feature a wide, open mouth and eyes that are either closed or nearly closed. This
issue was already anticipated during the image filtering process, where it was noted that the visual
diferences between these expressions were minimal (see Figure 11).</p>
    </sec>
    <sec id="sec-7">
      <title>7. Conclusion and future work</title>
      <p>In this study, we have introduced and tested UIBAIFED, a novel facial expression dataset that features
high-quality, realistic color images labeled according to age group, gender, ethnicity, and facial
expression. The labeling follows the universal expressions, encompassing a total of 22 micro-expressions
based on the terminology proposed by Gary Faigin.</p>
      <p>To validate the dataset, a Convolutional Neural Network (CNN) was employed, achieving an accuracy
of 80%, with strong performance across most expressions.</p>
      <p>Compared to existing facial expression datasets, UIBAIFED introduces a key innovation by providing
a more detailed level of labeling. To the best of our knowledge, no other database currently ofers this
granularity in annotation.</p>
      <p>Moving forward, the dataset enables new research opportunities, particularly in analyzing FER
performance across diferent age and ethnic groups. Addressing these challenges will contribute to
further advancements in the field of facial expression recognition.</p>
    </sec>
    <sec id="sec-8">
      <title>Acknowledgments</title>
      <p>This work is part of the Project PID2022-136779OB-C32 (PLEISAR) funded by
MICIU/AEI/10.13039/501100011033/ and FEDER, EU. The authors thank the University of the
Balearic Islands and the Department of Mathematics and Computer Science for their support.</p>
    </sec>
    <sec id="sec-9">
      <title>Declaration on generative AI</title>
      <p>During the preparation of this work, the authors used Stable Difusion 1.5 and Realistic Vision V6.0 B1
for the generation of the images that comprise the UIBAIFED dataset. Additionally, ChatGPT (GPT-4,
March 2025 version) was employed to assist with grammar, spelling, and language refinement. After
using these tool(s)/service(s), the author(s) reviewed and edited the content as needed and take(s) full
responsibility for the publication’s content.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>S.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Deng</surname>
          </string-name>
          ,
          <article-title>Deep facial expression recognition: A survey</article-title>
          ,
          <source>IEEE Transactions on Afective Computing</source>
          <volume>13</volume>
          (
          <year>2022</year>
          )
          <fpage>1195</fpage>
          -
          <lpage>1215</lpage>
          . doi:
          <volume>10</volume>
          .1109/TAFFC.
          <year>2020</year>
          .
          <volume>2981446</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>D.</given-names>
            <surname>Kollias</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Schulc</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Hajiyev</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Zafeiriou</surname>
          </string-name>
          ,
          <article-title>Analysing afective behavior in the first abaw 2020 competition</article-title>
          , in: 2020
          <source>15th IEEE International Conference on Automatic Face and Gesture Recognition (FG</source>
          <year>2020</year>
          ), IEEE,
          <year>2020</year>
          , pp.
          <fpage>637</fpage>
          -
          <lpage>643</lpage>
          . doi:
          <volume>10</volume>
          .1109/FG47880.
          <year>2020</year>
          .
          <volume>00126</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>J.</given-names>
            <surname>Buolamwini</surname>
          </string-name>
          , T. Gebru,
          <article-title>Gender shades: intersectional accuracy disparities in commercial gender classification, in: S. A</article-title>
          .
          <string-name>
            <surname>Friedler</surname>
          </string-name>
          , C. Wilson (Eds.),
          <source>Proceedings of the 1st Conference on Fairness, Accountability and Transparency</source>
          , volume
          <volume>81</volume>
          <source>of Proceedings of Machine Learning Research, PMLR</source>
          ,
          <year>2018</year>
          , pp.
          <fpage>77</fpage>
          -
          <lpage>91</lpage>
          . URL: https://proceedings.mlr.press/v81/buolamwini18a.html.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>I. J.</given-names>
            <surname>Goodfellow</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Erhan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. L.</given-names>
            <surname>Carrier</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Courville</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Mirza</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Hamner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Cukierski</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Tang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Thaler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.-H.</given-names>
            <surname>Lee</surname>
          </string-name>
          , et al.,
          <article-title>Challenges in representation learning: a report on three machine learning contests</article-title>
          ,
          <source>in: Neural information processing:</source>
          20th international conference,
          <source>ICONIP</source>
          <year>2013</year>
          , daegu, korea,
          <source>november 3-7</source>
          ,
          <year>2013</year>
          . Proceedings,
          <source>Part III 20</source>
          , Springer,
          <year>2013</year>
          , pp.
          <fpage>117</fpage>
          -
          <lpage>124</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>P.</given-names>
            <surname>Lucey</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. F.</given-names>
            <surname>Cohn</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Kanade</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Saragih</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Ambadar</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Matthews</surname>
          </string-name>
          ,
          <article-title>The extended cohn-kanade dataset (CK+): a complete dataset for action unit and emotion-specified expression</article-title>
          , in
          <article-title>: 2010 IEEE computer society conference on computer vision and pattern recognition-workshops,</article-title>
          <string-name>
            <surname>IEEE</surname>
          </string-name>
          ,
          <year>2010</year>
          , pp.
          <fpage>94</fpage>
          -
          <lpage>101</lpage>
          . doi:
          <volume>10</volume>
          .1109/CVPRW.
          <year>2010</year>
          .
          <volume>5543262</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>S.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Deng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Du</surname>
          </string-name>
          ,
          <article-title>Reliable crowdsourcing and deep locality-preserving learning for expression recognition in the wild</article-title>
          ,
          <source>in: Proceedings of the IEEE conference on computer vision and pattern recognition</source>
          ,
          <year>2017</year>
          , pp.
          <fpage>2852</fpage>
          -
          <lpage>2861</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>A.</given-names>
            <surname>Mollahosseini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Hasani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. H.</given-names>
            <surname>Mahoor</surname>
          </string-name>
          ,
          <article-title>AfectNet: A database for facial expression, valence, and arousal computing in the wild</article-title>
          ,
          <source>IEEE Transactions on Afective Computing</source>
          <volume>10</volume>
          (
          <year>2017</year>
          )
          <fpage>18</fpage>
          -
          <lpage>31</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>J.</given-names>
            <surname>Fan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Deng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Tao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H. K.</given-names>
            <surname>Kwan</surname>
          </string-name>
          ,
          <article-title>Combating uncertainty and class imbalance in facial expression recognition</article-title>
          ,
          <source>in: TENCON 2022 - 2022 IEEE Region 10 Conference (TENCON)</source>
          , IEEE,
          <year>2022</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>4</lpage>
          . doi:
          <volume>10</volume>
          .1109/TENCON55691.
          <year>2022</year>
          .
          <volume>9977693</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>I.</given-names>
            <surname>Dominguez-Catena</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Paternain</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Galar</surname>
          </string-name>
          ,
          <article-title>Metrics for dataset demographic bias: a case study on facial expression recognition</article-title>
          ,
          <source>IEEE Transactions on Pattern Analysis and Machine Intelligence</source>
          <volume>46</volume>
          (
          <year>2024</year>
          )
          <fpage>5209</fpage>
          -
          <lpage>5226</lpage>
          . doi:
          <volume>10</volume>
          .1109/TPAMI.
          <year>2024</year>
          .
          <volume>3361979</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>A.</given-names>
            <surname>Psaroudakis</surname>
          </string-name>
          , D. Kollias,
          <string-name>
            <surname>MixAugment</surname>
          </string-name>
          &amp;
          <article-title>Mixup: augmentation methods for facial expression recognition</article-title>
          ,
          <source>in: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition</source>
          ,
          <year>2022</year>
          , pp.
          <fpage>2367</fpage>
          -
          <lpage>2375</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>J.</given-names>
            <surname>Yu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Fan</surname>
          </string-name>
          , G. Sun,
          <article-title>Mixcut:a data augmentation method for facial expression recognition</article-title>
          ,
          <year>2024</year>
          . URL: https://arxiv.org/abs/2405.10489. arXiv:
          <volume>2405</volume>
          .
          <fpage>10489</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , Y. Li, l. Qin,
          <string-name>
            <given-names>X.</given-names>
            <surname>Liu</surname>
          </string-name>
          , W. Deng,
          <article-title>Leave no stone unturned: mine extra knowledge for imbalanced facial expression recognition</article-title>
          , in: A.
          <string-name>
            <surname>Oh</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Naumann</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Globerson</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Saenko</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Hardt</surname>
          </string-name>
          , S. Levine (Eds.),
          <source>Advances in Neural Information Processing Systems</source>
          , volume
          <volume>36</volume>
          ,
          <string-name>
            <surname>Curran</surname>
            <given-names>Associates</given-names>
          </string-name>
          , Inc.,
          <year>2023</year>
          , pp.
          <fpage>14414</fpage>
          -
          <lpage>14426</lpage>
          . URL: https://proceedings.neurips.cc/paper_files/paper/ 2023/file/2e6744370a8616c90d1e3b7a41993b7c-Paper-Conference.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Joo</surname>
          </string-name>
          ,
          <article-title>Understanding and mitigating annotation bias in facial expression recognition</article-title>
          ,
          <source>in: Proceedings of the IEEE/CVF International Conference on Computer Vision</source>
          (ICCV),
          <year>2021</year>
          , pp.
          <fpage>14980</fpage>
          -
          <lpage>14991</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>I. P.</given-names>
            <surname>Adegun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H. B.</given-names>
            <surname>Vadapalli</surname>
          </string-name>
          ,
          <article-title>Facial micro-expression recognition: a machine learning approach</article-title>
          ,
          <source>Scientific African</source>
          <volume>8</volume>
          (
          <year>2020</year>
          )
          <article-title>e00465</article-title>
          . URL: https://www.sciencedirect.com/science/article/pii/ S2468227620302039. doi:https://doi.org/10.1016/j.sciaf.
          <year>2020</year>
          .e00465.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>H.</given-names>
            <surname>Guerdelli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Ferrari</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Barhoumi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Ghazouani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Berretti</surname>
          </string-name>
          ,
          <article-title>Macro- and micro-expressions facial datasets: a survey</article-title>
          ,
          <source>Sensors</source>
          <volume>22</volume>
          (
          <year>2022</year>
          ). doi:
          <volume>10</volume>
          .3390/s22041524.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <surname>G. Faigin,</surname>
          </string-name>
          <article-title>The artist's complete guide to facial expression</article-title>
          , Watson-Guptill, New York,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17] Ofice of Institutional Research, Race/ethnicity FAQs,
          <year>2025</year>
          . URL: https://provost.tufts.edu/ institutionalresearch/race-ethnicity-faq/.
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <surname>Stable</surname>
            <given-names>Difussion</given-names>
          </string-name>
          , Stable Difusion online,
          <year>2025</year>
          . URL: https://stabledifffusion.com.
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <surname>Stable</surname>
            <given-names>Difussion</given-names>
          </string-name>
          ,
          <source>Realistic vision V6</source>
          .0,
          <year>2025</year>
          . URL: https://civitai.com/models/4201/ realistic-vision-v60-
          <fpage>b1</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>E. J.</given-names>
            <surname>Hu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Shen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Wallis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Allen-Zhu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Chen</surname>
          </string-name>
          , et al.,
          <article-title>LoRA: low-rank adaptation of large language models</article-title>
          ,
          <source>ICLR</source>
          <volume>1</volume>
          (
          <year>2022</year>
          )
          <article-title>3</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zeng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <article-title>The expressive power of low-rank adaptation</article-title>
          ,
          <source>arXiv preprint arXiv:2310.17513</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <surname>Civitai</surname>
          </string-name>
          ,
          <source>LoRA Stable Difusion &amp; Flux AI models</source>
          ,
          <year>2025</year>
          . URL: https://civitai.com/tag/lora.
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>P.</given-names>
            <surname>Ekman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Dalgleish</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Power</surname>
          </string-name>
          ,
          <article-title>Handbook of cognition and emotion</article-title>
          , Wiley Online Library,
          <year>1999</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <fpage>AUTOMATIC1111</fpage>
          ,
          <string-name>
            <surname>Stable</surname>
            <given-names>Difusion web UI</given-names>
          </string-name>
          ,
          <year>2022</year>
          . URL: https://github.com/AUTOMATIC1111/ stable-diffusion-webui.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>