<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>LACELL at SatiSPeech-IberLEF 2025: Multimodal Linguistic Features plus Embeddings for Satire Identification from YouTube</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Ángela Almela</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Pascual Cantos-Gómez</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Daniel Granados-Meroño</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Gema Alcaraz-Mármol</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Facultad de Educación, Universidad de Castilla-La Mancha</institution>
          ,
          <addr-line>45004, Toledo</addr-line>
          ,
          <country country="ES">Spain</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Facultad de Letras, Universidad de Murcia, Campus de La Merced</institution>
          ,
          <addr-line>30001, Murcia</addr-line>
          ,
          <country country="ES">Spain</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <abstract>
        <p>These working notes summarize the LACELL team's participation in the SatiSPeech 2025 shared task, which focuses on multimodal satire detection, integrating textual and acoustic features to better capture the subtleties of humorous and satirical communication. Its application in Spanish is particularly relevant given the linguistic and cultural diversity of Spanish-speaking communities, where satire often relies on both content and prosody. We participated in the multimodal task using an ensemble approach combining linguistic and prosodic features with sentence embeddings extracted from several fine-tuned LLMs, and achieved the 8th place with a macro F1 score of 81.4695%. This result demonstrates the potential of multimodal strategies for capturing complex communicative intentions such as satire.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Linguistic Features</kwd>
        <kwd>Sentence Embeddings</kwd>
        <kwd>Multimodal Classification</kwd>
        <kwd>Satire Classification</kwd>
        <kwd>Natural Language Processing</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Satire detection is an increasingly important task in Natural Language Processing (NLP) and multimodal
analysis, especially in contexts where irony, humor, and exaggeration play a central role in public
discourse. Unlike emotion recognition, satire detection poses additional challenges due to its frequent
reliance on implicit meaning, ambiguity, and culturally grounded references, making it dificult to
identify using surface-level features alone [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. These challenges are compounded when satire is conveyed
through multiple modalities, such as tone of voice and textual cues.
      </p>
      <p>
        The automatic detection of satire is a challenging task due to its reliance on figurative language,
cultural references, and implicit meaning, which are often dificult to capture through surface-level
features alone. Prior studies highlight the importance of cultural and contextual understanding in
humor and satire perception [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], as well as the role of figurative language in social media discourse [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ].
Recent approaches have explored multimodal strategies that combine textual, visual, or acoustic cues to
improve satire recognition [
        <xref ref-type="bibr" rid="ref4 ref5">4, 5</xref>
        ], and even the application of large language models in low-resource
scenarios.
      </p>
      <p>
        The SatiSPeech 2025 shared task [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], organized within IberLEF 2025 [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], aims to advance the field of
multimodal satire detection by providing a benchmark dataset in Spanish that combines transcribed
speech with audio recordings. This combination allows the exploration of prosodic and linguistic
cues to better distinguish satirical from non-satirical content. Identifying satire in spoken language
is particularly relevant in Spanish given its widespread use in culturally diverse regions, where satire
often relies on local nuances and vocal delivery.
      </p>
      <p>
        Multimodal satire detection faces similar challenges to those of Automatic Emotion Recognition
from a similar shared task [
        <xref ref-type="bibr" rid="ref8 ref9">8, 9</xref>
        ], such as the integration of heterogeneous features and the scarcity of
authentic, annotated datasets. While many previous works have focused on textual satire, incorporating
prosodic features such as pitch, rhythm, or intonation ofers new avenues for capturing the subtle
communicative signals that characterize satirical speech.
      </p>
      <p>In this edition of the task, our team participated exclusively in the multimodal track, designing a
system based on ensemble learning that integrates linguistic features from UMUTextStats, sentence
embeddings from various large language models (LLMs), and prosodic audio features. Our approach
achieved the 8th position, with a macro F1-score of 81.4695%, showing the efectiveness of combining
diverse modalities for tackling the complexity of satirical communication. Additionally, this result
outperformed the provided baseline, which reached a macro F1-score of 79.9243%.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Dataset</title>
      <p>According to the organizers, the SatiSPeech 2025 dataset consists of audio segments in Spanish extracted
from a variety of YouTube videos containing either satirical or non-satirical content. The dataset was
compiled with the aim of capturing the multimodal nature of satire, which often relies on vocal delivery,
prosody, and contextual cues in addition to textual content. The satirical segments were collected
from well-known Spanish-language satirical sources such as El Mundo Today and Polònia, while the
non-satirical examples were taken from informative or neutral media.</p>
      <p>Each audio segment is accompanied by its corresponding manual transcription (generated
semiautomatically), and both modalities - text and audio - are provided for system development. The training
set contains a total of 2,000 labeled segments, balanced between satirical and non-satirical classes.
Additionally, a separate test set was provided for final evaluation.</p>
      <p>In our experiments, we did not use the initial development set released prior to the competition.
Instead, we created a custom development split by reserving 25% of the oficial training set for validation
and hyperparameter tuning. Table 1 summarizes the dataset statistics. Although the dataset is roughly
balanced in terms of class distribution, the diversity in speaker style, audio quality, and prosodic
variation adds significant complexity to the classification task.</p>
      <p>
        For the analysis of the dataset, we used the UMUTextStats tool [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] to extract linguistic features
and compare their distribution across satirical and non-satirical texts (see Figure 1). We found that
morphosyntactic features such as nouns, articles, pronouns, and adverbs are among the most informative
categories. Satirical texts tend to contain a higher frequency of afix-related features (e.g., sufixes
in nominative nouns), longer words, and a higher syllable-per-word ratio, possibly due to a more
elaborate or playful language style. In contrast, non-satirical texts present a higher average sentence
length and a more consistent use of punctuation symbols such as commas and periods. Interestingly,
personal pronouns (especially first-person forms) and question marks appear more frequently in satirical
documents, which may reflect a rhetorical strategy to engage the audience or simulate a conversational
tone. These observations suggest that satirical speech often relies on distinct structural and stylistic
markers, making linguistic profiling a valuable component in multimodal satire detection.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. System description</title>
      <p>
        We evaluated UMUTextStats [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] for linguistic feature extraction tool. This tool is similar to LIWC,
the de-facto standard for psychological and linguistic profiling. While LIWC2022 is the latest English
version available [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ], the last available version for Spanish dates back to 2007 [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ], and lacks support
for many contemporary linguistic constructs. In contrast, UMUTextStats is especifically designed for
Spanish, capturing linguistic phenomena that LIWC does not address, such as grammatical gender, verb
tense variations, and morphosyntactic patterns. The tool has been successfully applied in previous
studies, including hate speech detection [13] and satire analysis [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], demonstrating its relevance for
complex classification tasks in Spanish.
      </p>
      <p>Prior to extracting the linguistic features, we applied a preprocessing pipeline to the transcriptions.
This process involved cleaning the text by removing elements such as hyperlinks, hashtags, mentions,
digits, and percentages, which were either eliminated or replaced by standardized tokens. In addition,
we corrected expressive lengthenings and spelling errors using the ASPELL tool1 to ensure consistency
in lexical analysis. Importantly, we retained the original, unprocessed transcriptions for the extraction of
features related to orthographic correctness and stylistic variation, as these aspects could be informative
for satire detection.</p>
      <p>Regarding sentence embeddings, we evaluated a diverse set of Spanish and multilingual Large
Language Models (LLMs) to represent the transcribed speech content. Specifically, we employed eight
Transformer-based models: BETO, BERTIN, MarIA, ALBETO, DistilBETO, mDeBERTa, Twhin, and
XLM-R Twitter. These models cover a range of pretraining strategies, including formal corpora, literary
texts, multilingual resources, and social media content. Sentence embeddings were extracted using the
approach from [14].</p>
      <p>Table 2 summarizes the training settings for each model. A closer look at the training configurations
of the evaluated LLMs reveals that three models –MarIA, DistilBETO, and mDeBERTa– were trained
without any warm-up steps, suggesting that in these cases the models adapted to the task without
requiring a gradual increase in learning rate. This may indicate their robustness or compatibility with</p>
      <p>satire
the goal of satire detection. Regarding the batch size, some models were tuned with a batch size of 8,
while others achieved their better results with 16, reflecting diferent trade-ofs in terms of memory
consumption and training stability. In addition, most models were trained on a relatively high number
of epochs (3 to 5), with 5 being the upper limit, suggesting that satire detection requires extended
exposure to training data to capture nuanced patterns in both structure and meaning.</p>
      <p>Among the diferent types of acoustic features available, we opted for the use of MFCC (Mel-Frequency
Cepstral Coeficients), following the baseline system provided by the organizers. MFCCs are widely used
in speech processing tasks because they approximate the way humans perceive sound by emphasizing
perceptually relevant frequency bands. They capture essential aspects of speech prosody and articulation
that are often indicative of expressive intentions, such as emphasis, exaggeration, or irony—elements
frequently present in satirical speech. The MFCC vectors were extracted from each audio segment and
used as input for a dedicated model in our ensemble.</p>
      <p>Since all feature sets -—linguistic features (LFs), sentence embeddings, and acoustic features-— are
represented as vectors, we explored various strategies to combine them and improve classification
performance. In particular, we adopted an ensemble learning approach, where individual models were
trained using each feature set independently. For the acoustic modality, we included MFCC features,
following the strategy used in the oficial baseline system.</p>
      <p>We then evaluated multiple fusion strategies to combine the predictions from the unimodal models.
These included majority voting (mode), probability averaging, and max-confidence selection, where the
predicted label corresponds to the category with the highest probability across models. This multimodal
ensemble setup allowed us to leverage the complementary strengths of textual, linguistic, and prosodic
cues in the satire detection task.</p>
      <p>In order to adjust the LLMs for this task, we first fine-tuned the models with the training dataset
using hyperparameter tuning. For each LLM, we evaluate 10 configurations that include variations on
the learning rate, the warm-up steps, the weight decay, the number of epochs, and the batch size. Table
2 depicts the results for both models resulting in a larger number of epochs (4 for BETO, 5 for Maria)
and little or no warm-up steps.</p>
      <p>To integrate the outputs of the LLMs, MFCC features and the linguistic features extracted with
UMUTextStats, we train a deep neural network on the concatenated feature vectors. After performing
hyperparameter tuning, the resulting architecture consisted of a light funnel-shaped network with seven
layers, starting from an initial layer with 90 neurons, and progressively reducing the dimensionality.
The network used ELU (Exponential Linear Unit) as the activation function and did not include dropout,
as regularization was not required during training. This simple architecture proved efective, likely
because the input features—particularly the sentence embeddings—already captured rich semantic
information, while the linguistic features complemented them with structural and stylistic cues relevant
to satire.</p>
      <p>First, we present the experiments conducted using the custom validation split in Table 4.1. The results
are grouped into three main blocks: linguistic and acoustic features, sentence embeddings from various
LLMs, and feature integration strategies, including ensemble learning (EL) and Knowledge Integration
(KI).</p>
      <p>The first block shows the performance of models trained with individual feature sets. The linguistic
features extracted with UMUTextStats achieved strong results on their own, with an F1-score of 89.30%,
highlighting the relevance of surface-level and structural features such as part-of-speech distributions,
punctuation, and stylistic cues for detecting satirical content. In contrast, the MFCC-based audio model
obtained significantly lower results with an F1-score of only 48.86%. This suggests that, although
prosodic cues are informative, they are less reliable when used in isolation, possibly due to variability
in speaker style or recording conditions. This confirms that acoustic features alone are not suficient
for robust satire detection and need to be complemented with textual information.</p>
      <p>In the second group, we evaluated several LLMs by extracting sentence embeddings from the
transcribed speech. All models outperformed the linguistic and acoustic baselines. Among them, MarIA
achieved the highest individual F1-score (94.90%), followed closely by Twhin and BERTIN, with 94.30%
and 94.22%, respectively. These results suggest that domain-specific pretraining can provide embeddings
well-suited for capturing the subtle semantics and discourse patterns characteristic of satirical texts.
While models like DistilBETO and mDeBERTa performed slightly worse, they still exceeded 93%
F1score, confirming that even compact or multilingual architectures contribute valuable representations
to the task.</p>
      <p>The last block reports results from the integration strategies. As expected, combining diferent
sources of information consistently improved performance. The weighted ensemble (EL WEIGHTED)
achieved the highest scores across all metrics, with a macro F1-score of 95.82%, indicating that giving
more weight to stronger models benefits the ensemble. Interestingly, the KI approach achieved the
exact same performance, demonstrating that even a simple fusion architecture can compete with more
complex ensemble methods. Other strategies such as mean probability and mode voting also performed
well, but slightly below the best. These results highlight that multimodal integration, when carefully
designed, leads to robust and high-performing satire detection systems.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Results</title>
      <sec id="sec-4-1">
        <title>4.1. Validation</title>
        <p>In this section, we report the results with our custom validation split (see Section 4.1), the oficial leader
board (see Section 4.2, and an error analysis of the custom validation split (see Section 4.3).
Next, we present the detailed classification report of the Knowledge Integration (KI) strategy, where
all features—sentence embeddings, MFCC audio features, and linguistic features—were combined into
a single shallow neural network. Table 4 shows the results on the custom validation split, including
precision, recall, and F1-score for both classes, as well as macro and weighted averages. The model
achieved very balanced performance across classes, with macro and weighted F1-scores around 96%,
indicating that it generalizes well to both satirical and non-satirical content. Notably, the F1-score
for satire reached 95.73%, showing that the integrated feature representation efectively captures the
nuanced characteristics of satirical speech.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Oficial results</title>
        <p>unified neural network. This result outperformed the oficial baseline provided by the organizers
(79.9243%) and placed us within a competitive range, only 6.8708 points below the top-ranked team,
which achieved a macro F1-score of 88.3403%. These results highlight the challenge of multimodal
satire detection and the potential of integrated architectures even in highly competitive settings.
l
a
u
t
c
A
no satire
satire
Predicted</p>
      </sec>
      <sec id="sec-4-3">
        <title>4.3. Error Analysis</title>
        <p>To conduct the error analysis, we examined the confusion matrix of our best system, which used a
Knowledge Integration strategy with the custom validation split (see Figure 2). The results show a
strong performance across both classes, with most instances being correctly classified. However, we
observed an asymmetry in the errors: the system misclassified 29 satirical segments as non-satirical,
compared to only 19 non-satirical instances misclassified as satire. This suggests a slight bias toward
the "no satire" class, possibly due to its more regular prosodic or linguistic patterns. Despite these
misclassifications, the overall balance in predictions and the low error rates confirm the robustness of
the integrated approach.</p>
      </sec>
      <sec id="sec-4-4">
        <title>4.4. Conclusions and further work</title>
        <p>In this working notes, we have described the participation of the LACELL team in Task 2 of the SatiSPeech
2025 competition, focused on multimodal satire detection in Spanish. Our system integrated sentence
embeddings from several LLMs, linguistic features extracted with UMUTextStats, and acoustic features
based on MFCCs, combined through a shallow neural network following a Knowledge Integration
strategy. We achieved the 8th position in the oficial ranking with a macro F1-score of 81.4695%,
outperforming the baseline by 1.5452 points. While our approach did not reach the top-tier performance,
it demonstrated the viability of lightweight fusion strategies and confirmed the contribution of each
modality individually.</p>
        <p>As further work, we plan to explore the incorporation of pretrained acoustic language models
such as Wav2Vec 2.0 or UniSpeech, which could enhance the representation of prosodic and phonetic
cues in satire detection, as suggested in [15]. Additionally, we aim to investigate alternative fusion
strategies between modalities, including early, late, and hybrid approaches, to more efectively exploit the
complementary information of text and audio. We are also considering incorporating feature selection
or attention-based mechanisms to identify the most relevant acoustic or lexical cues contributing to the
satirical tone.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Acknowledgments</title>
      <p>This work is part of the research projects LaTe4PoliticES (PID2022- 138099OB-I00) funded by
MICIU/AEI/10.13039/501100011033 and the European Regional Development Fund (ERDF)-a way to make
Europe and LT-SWM (TED2021-131167B-I00) funded by MICIU/AEI/10.13039/ 501100011033 and by the
European Union Next Generation EU/PRTR. This work is also part of the research project "Services based
on language technologies for political microtargeting" (22252/PDC/23) funded by the Autonomous
Community of the Region of Murcia through the Regional Support Program for the Transfer and
Valorization of Knowledge and Scientific Entrepreneurship of the Seneca Foundation, Science and
Technology Agency of the Region of Murcia.</p>
    </sec>
    <sec id="sec-6">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the author(s) used DeepL in order to Grammar and spelling check.
[13] J. A. García-Díaz, S. M. Jiménez-Zafra, M. A. García-Cumbreras, R. Valencia-García, Evaluating
feature combination strategies for hate-speech detection in spanish using linguistic features and
transformers, Complex &amp; Intelligent Systems 9 (2023) 2893–2914.
[14] N. Reimers, I. Gurevych, Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks,
in: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing
and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP),
2019, pp. 3982–3992.
[15] L. Pepino, P. Riera, L. Ferrer, Emotion Recognition from Speech Using wav2vec 2.0 Embeddings,
Proc. Interspeech 2021 (2021) 3400–3404.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>J. A.</given-names>
            <surname>García-Díaz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Valencia-García</surname>
          </string-name>
          ,
          <article-title>Compilation and Evaluation of the Spanish Saticorpus 2021 for Satire Identification using Linguistic Features and Transformers</article-title>
          ,
          <source>Complex &amp; Intelligent Systems</source>
          <volume>8</volume>
          (
          <year>2022</year>
          )
          <fpage>1723</fpage>
          -
          <lpage>1736</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>T.</given-names>
            <surname>Jiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Hou</surname>
          </string-name>
          , Cultural Diferences in Humor Perception, Usage, and
          <string-name>
            <surname>Implications</surname>
          </string-name>
          , Frontiers in psychology 10 (
          <year>2019</year>
          )
          <fpage>123</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>M.</given-names>
            <surname>del Pilar</surname>
          </string-name>
          Salas-Zárate,
          <string-name>
            <given-names>G.</given-names>
            <surname>Alor-Hernández</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. L.</given-names>
            <surname>Sánchez-Cervantes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Paredes-Valverde</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. L.</given-names>
            <surname>García-Alcaraz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Valencia-García</surname>
          </string-name>
          ,
          <article-title>Review of English Literature on Figurative Language applied to Social Networks</article-title>
          ,
          <source>Knowledge and Information Systems</source>
          <volume>62</volume>
          (
          <year>2020</year>
          )
          <fpage>2105</fpage>
          -
          <lpage>2137</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>L.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Levi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Hosseini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Broniatowski</surname>
          </string-name>
          ,
          <article-title>A Multi-Modal Method for Satire Detection using Textual and Visual Cues</article-title>
          ,
          <source>in: Proceedings of the 3rd NLP4IF Workshop on NLP for Internet Freedom: Censorship</source>
          , Disinformation, and
          <string-name>
            <surname>Propaganda</surname>
          </string-name>
          ,
          <year>2020</year>
          , pp.
          <fpage>33</fpage>
          -
          <lpage>38</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>R.</given-names>
            <surname>Ortega-Bueno</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Rosso</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. E. M.</given-names>
            <surname>Pagola</surname>
          </string-name>
          <article-title>, Multi-view informed attention-based model for Irony and Satire detection in Spanish variants</article-title>
          ,
          <source>Knowledge-Based Systems</source>
          <volume>235</volume>
          (
          <year>2022</year>
          )
          <fpage>107597</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>R.</given-names>
            <surname>Pan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. A.</given-names>
            <surname>García-Díaz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Bernal-Beltrán</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>García-Sánchez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Valencia-García</surname>
          </string-name>
          , Overview of SatiSPeech at IberLEF 2025:
          <article-title>Multimodal Audio-Text Satire Classification in Spanish</article-title>
          ,
          <source>Procesamiento del Lenguaje Natural</source>
          <volume>75</volume>
          (
          <year>2025</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>J.</given-names>
            <surname>Á</surname>
          </string-name>
          .
          <string-name>
            <surname>González-Barba</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Chiruzzo</surname>
            ,
            <given-names>S. M.</given-names>
          </string-name>
          <string-name>
            <surname>Jiménez-Zafra</surname>
          </string-name>
          ,
          <article-title>Overview of IberLEF 2025: Natural Language Processing Challenges for Spanish and other Iberian Languages, in: Proceedings of the Iberian Languages Evaluation Forum (IberLEF 2025), co-located with the 41st Conference of the Spanish Society for Natural Language Processing (SEPLN 2025), CEUR-WS</article-title>
          . org,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>R.</given-names>
            <surname>Pan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. A.</given-names>
            <surname>García-Díaz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Á</surname>
          </string-name>
          .
          <string-name>
            <surname>Rodríguez-García</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>García-Sanchez</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Valencia-García</surname>
          </string-name>
          ,
          <article-title>Overview of EmoSPeech 2024IberLEF: Multimodal Speech-text Emotion Recognition in Spanish</article-title>
          ,
          <source>Procesamiento del Lenguaje Natural</source>
          <volume>73</volume>
          (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>R.</given-names>
            <surname>Pan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. A.</given-names>
            <surname>García-Díaz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Á</surname>
          </string-name>
          .
          <string-name>
            <surname>Rodríguez-García</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Valencia-García</surname>
          </string-name>
          ,
          <source>Spanish MEACorpus</source>
          <year>2023</year>
          :
          <article-title>A multimodal speech-text corpus for emotion analysis in Spanish from natural environments</article-title>
          ,
          <source>Computer Standards &amp; Interfaces</source>
          (
          <year>2024</year>
          )
          <fpage>103856</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>J. A.</given-names>
            <surname>García-Díaz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. J.</given-names>
            <surname>Vivancos-Vicente</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Almela</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Valencia-García</surname>
          </string-name>
          ,
          <article-title>Umutextstats: A linguistic feature extraction tool for Spanish</article-title>
          ,
          <source>in: Proceedings of the Thirteenth Language Resources and Evaluation Conference</source>
          ,
          <year>2022</year>
          , pp.
          <fpage>6035</fpage>
          -
          <lpage>6044</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>R. L.</given-names>
            <surname>Boyd</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ashokkumar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Seraj</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. W.</given-names>
            <surname>Pennebaker</surname>
          </string-name>
          ,
          <article-title>The development and psychometric properties of LIWC-22</article-title>
          , Austin, TX: University of Texas at Austin (
          <year>2022</year>
          )
          <fpage>1</fpage>
          -
          <lpage>47</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>N.</given-names>
            <surname>Ramírez-Esparza</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. W.</given-names>
            <surname>Pennebaker</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F. A.</given-names>
            <surname>García</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Suriá</surname>
          </string-name>
          ,
          <article-title>La psicología del uso de las palabras: Un programa de computadora que analiza textos en español</article-title>
          , Revista mexicana de psicología (
          <year>2007</year>
          )
          <fpage>85</fpage>
          -
          <lpage>99</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>