<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Conference and Labs of the Evaluation Forum, September</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>ELiRF-VRAIN at eRisk 2024: Using LongFormers for Early Detection of Signs of Anorexia</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Andreu Casamayor</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Vicent Ahuir</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Antonio Molina</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Lluís-Felip Hurtado</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Valencian Research Institute for Artificial Intelligence (VRAIN), Universitat Politècnica de València</institution>
          ,
          <addr-line>Camino de Vera s/n, 46022 Valencia.</addr-line>
          <country country="ES">Spain</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2024</year>
      </pub-date>
      <volume>0</volume>
      <fpage>9</fpage>
      <lpage>12</lpage>
      <abstract>
        <p>This paper describes the approaches taken by the ELiRF-VRAIN team at the Task 2 of eRisk at CLEF 2024 focused on the early detection of signs of anorexia on English-language social media. Our work involved three distinct approaches: one using a Support Vector Machine (SVM) and the other two based on pre-trained Transformer models. Among the Transformer models, one approach employed BERT-like models, while the other used LongFormer models. To fine-tune our models, we implemented a data augmentation process on the dataset provided by the organization. In the validation phase, the models trained on the augmented dataset improved the F1 score results. In particular, F1 increased from 0.89 to 0.94 for the LongFormer model. During the testing phase the SVM model and LongFormer with data augmentation obtained the best results. LongFormer improved BERT-like model performance due to its ability to handle large contexts. Seeing the results achieved in the validation phase, we can say that the overall performance was not as good as expected. A detailed analysis of the results would be necessary to find out the reasons.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Longformers</kwd>
        <kwd>Transformers</kwd>
        <kwd>Support Vector Machine</kwd>
        <kwd>Anorexia</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Anorexia nervosa is the formal term for anorexia, and it’s a complex, really multi-structural eating
disorder. This is a disorder characterized by a fear of gaining weight and by the maintenance of a
distorted body image through severe food restriction and excessive weight loss. It is hazardous for both
males and females, but is most common among young women. Women account for 90-95% of those
afected; the age range is usually between 12 and 25 years, and it is most common between 12 and 17
years of age. [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]
      </p>
      <p>The impacts of anorexia extend to all aspects of one’s health and functioning, extending far beyond
malnutrition to nearly every organ system in the body, even when comorbid with other mental health
issues like depression and anxiety. Little is done, anorexia is often dificult to detect and treat due to its
insidious onset and the societal stigma surrounding mental health and eating disorders.</p>
      <p>For this reason, the analysis of social interactions to detect risks of anorexia has recently become
one of the most important ways of detection. This type of problem, anorexia detection, is complicated
due to some reasons, such as the amount and quality of the data. CLEF eRisk created diferent tasks, to
provide quality data and promote the creation of models for this early detection.</p>
      <p>
        In 2024’s edition, eRisk proposed three shared tasks [
        <xref ref-type="bibr" rid="ref2 ref3">2, 3</xref>
        ]: (1) Search for symptoms of depression, (2)
Early Detection of Signs of Anorexia, and (3) Measuring the severity of the signs of Eating Disorders.
      </p>
      <p>
        We focused our participation on the second shared task, where we used three diferent approaches to
tackle the problem posed by the task:
1. The initial approach employs a traditional machine learning algorithm, Support Vector Machines
(SVM). SVMs have shown meaningful performance in classifying lengthy texts, similar to this
case. We use this approach to evaluate the efectiveness of classical models.
2. The second approach utilizes Transformers [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] by leveraging a pre-trained RoBERTa model [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] as
a foundation, followed by a fine-tuning process to adapt it to the downstream task. We performed
ifne-tuning using two distinct datasets: one provided by the organization and the other created
through data augmentation.
3. The final approach is similar to the second one but aims to capture more context by using a
pre-trained LongFormer model [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. This model accommodates larger input sizes, allowing it to
grasp more contextual information. We fine-tuned the LongFormer model using the same dataset
as in the previous approach.
      </p>
      <p>We submitted four runs for Task 2, one for approaches 1 and 2, and two for approach 3. Before
selecting the best model for each approach, we put them through a validation phase, where we tested
diferent configurations and datasets used.</p>
      <p>
        We have done this kind of experimentation before and had excellent results, proving how reliable
and efective our approach is. In related topic works, we used similar methods and achieved substantial
outcomes [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ].
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. Description of Dataset and Task</title>
      <p>Task 2 involves the early detection of anorexia risk by sequentially analyzing pieces of evidence to
identify early signs of the disorder as promptly as possible. This task primarily focuses on evaluating
natural language processing solutions, particularly those that analyze texts from social media. Texts
must be processed in the chronological order in which they were created. This simulates better what
the system would do: monitor real-time user interactions on blogs, social networks, or other online
platforms.</p>
      <p>
        The dataset in Task 2 consisted of a writing (post or comments) collection from a set of Social Media
users formed from the datasets from previous editions of the task in 2018 and 2019. This collection has
the same format as the one delivered in [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], where there are two diferent classes: users who sufer from
anorexia and a control group (non-anorexia). Every user has a chronological collection of messages or
writings.
      </p>
      <p>Table 1 shows the distribution among the diferent labels in the dataset</p>
      <p>As mentioned, the primary goal of this competition is to predict signs of anorexia as promptly as
possible. To simulate realistic conditions, the organizers set up a server that sequentially delivers data
packets, each containing a message from a user. The system must predict the user’s signs of anorexia, if
any, by considering both the current message and all previous messages before receiving the next data
packet.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Systems and Architecture and Techniques</title>
      <p>In this type of task, a relevant factor to consider is the amount of context required for accurate detection.
Since each user can have numerous messages, the size of the input to the system becomes a crucial
consideration. One of our team’s objectives was to examine the impact of context in these tasks.
Specifically, we aimed to evaluate the performance of diferent systems based on their ability to handle
varying amounts of context. We selected three diferent systems to achieve this goal: the first based
on Support Vector Machines (SVM), the second based on a RoBERTa model, and the third based on
LongFormer model. Each system evaluated has a diferent size for context:
• Support Vector Machines (SVM) do not have a fixed limit on input size; they construct a vector
with a length corresponding to the vocabulary size. This flexibility allows SVMs to handle a large
and variable amount of data, as they can create feature vectors based on the entirety of the input
text’s vocabulary, accommodating diverse and extensive datasets.
• The selected RoBERTa model has a limit of 512 tokens in the input.</p>
      <p>• The selected LongFormer model has a limit of 4096 tokens in the input.</p>
      <p>Additionally, we developed two distinct datasets to train and evaluate the performance of the
transformer-based systems.</p>
      <p>Dataset 1. We created only one sample per user by aggregating all their messages, both for positive
and negative labeled users. This approach ensures that the dataset efectively captures the overall
context and messaging patterns of every user, facilitating a more accurate evaluation of the models’
performance in distinguishing between positive and negative cases.</p>
      <p>Dataset 2. If we had some a priori evidence of in which message a user begins to present symptoms
of mental illness risk, we could label the samples from previous messages as negative, and the samples
containing that message and subsequent ones as positive. In this way, we could increase the number of
positive samples to achieve a more precise model. This data augmentation process is explained in the
next section.</p>
      <p>To conduct our experimentation, we split the original dataset into two partitions: training (80% of
users) and development (20% of users). We ensured that both partitions maintained the same proportions
of positive and negative samples to preserve the dataset’s balance and integrity. Table 2 shows the
distribution of samples in Dataset 1.</p>
      <sec id="sec-3-1">
        <title>3.1. Data Augmentation</title>
        <p>The data augmentation process aims to generate additional samples for each positive user. As mentioned
earlier, we need evidence of when a user begins to exhibit signs of anorexia in their messages. To
identify this, we relied on predictions from the SVM-based classifier. We assume that all messages
preceding the SVM decision point do not express signs of anorexia. To implement this, we followed
these steps:
1. For positive users, we calculated how many messages the SVM needs to classify the user as
positive. Each user has a diferent trigger value.
2. For false negatives, we used the mean of the true positive trigger values as the trigger value.
3. For each positive user in the original data set, let  be the number of messages that the SVM
model needs to determine this user’s mental disorder risk,   be the maximum number of
messages the model supports as input, and  the ith message from the user.</p>
        <p>a) we created  − 1 negative samples as follows:</p>
        <p>(1), (12), (123), ..., (1...− 1)
b) and   −  + 1 positive samples:</p>
        <p>(1...), (1...+1), ..., (1...... )
4. Note that the value of   depends on which model was used and the number of tokens in the
messages. That is, we discard messages from an accumulated history of more than 512 tokens for
RoBERTa and 4096 for LongFormer. So, if  &gt;   only negative samples are generated.
5. For negative users, we created new samples accumulating the history as before, stopping when
the MAX was reached.</p>
        <p>The result of this technique is a new dataset with a higher number of positive samples for the training.
In the development partition, we held a sample per user, as in Dataset 1.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Classical Machine Learning Classifier Approach</title>
        <p>To evaluate the significance of the context, we aimed to use a classical machine learning classifier that
is capable of handling all the available context. One of the major issues with Transformer-based models
is that their ability to handle large texts is limited by the input size. This greatly afects performance
because the input cannot contain the length of the sample, whereby crucial information may be lost.
We would use such a classical machine learning model as SVM to create a vector as long as the size of
the vocabulary to show the model’s performance when it has no such restriction.</p>
        <p>
          First, we experimented to compare diferent types of classical machine learning classifiers. We
utilized the Scikit-learn library [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ] for this purpose, employing its default classifiers to identify the
best-performing model. The results, presented in Table 4, indicate that the Linear SVM emerged as the
top performer among the classifiers tested.
• Preprocess of Data:
1. First approach: Transform the text into tokens using TweetTokenizer and then eliminate
stop words.
2. Second Approach: Same as the first approach with the addition of methods to clean the text,
eliminate non-alphanumerical characters and others, and lemmatize tokens.
• Sentimental Analysis: We used the model
"lxyuan/distilbert-base-multilingual-cased-sentimentsstudent" [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ] to perform sentiment analysis on every user message. This process yielded three
results: positive messages, negative messages, and neutral messages. These results were
normalized and subsequently added as a new feature to the TF-IDF representation. This enhancement
allowed us to incorporate sentiment-based insights into our analysis, potentially improving the
performance and accuracy of our classification models.
• TF-IDF: We used the class TfidfVectorizer from Scikit-learn to vectorize the data. We
experimented with diferent configurations for the analyzer and ngram_range parameters, while
using the default values for other features. This approach allowed us to identify the optimal
configuration for the task.
        </p>
        <p>To find the best models for every approach, we did an exhaustive grid search over some specific
parameters, such as regularization parameter C, diferent tols, and diferent loss.</p>
        <p>We obtained 8 diferent approaches. Table 5 shows the diferent configurations used in the
experimentation, the column TF-IDF refers to the type of analyzers (word or char) used and the number of
n-grams. The last column refers to the best model found in the search grid.</p>
        <p>The result shows in Table 6 the best configuration is the SVM-1, using the first preprocess for the
data, without sentimental analysis, "char_wb" as the analyzer and (4-5) as ngram_range. This model
was used for Run0 in Task 2.</p>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. BERT-like Model Approach</title>
        <p>It is well known that state-of-the-art models in NLP are based on Transformers. Models like BERT
and RoBERTa typically ofer excellent versatility for classification tasks. However, these models are
often limited to handling a maximum of 512 tokens, which can be problematic for tasks requiring the
processing of long contexts, such as the one at hand. To address this issue, we used one of these models
as a baseline to compare against other models with a better capacity for managing large contexts. This
comparison allows us to evaluate the performance trade-ofs and benefits of diferent approaches in
handling extended textual data.</p>
        <p>
          We conducted research to find a base model trained in domains related to eating disorders; however,
we did not find any pre-trained model specialized in eating disorders. While we were doing the research,
we found the following: between 50% to 75% of those who struggle with an eating disorder will also
experience symptoms of depression or anxiety [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ]. Therefore, we used a pre-trained model related to
mental disorders instead.
        </p>
        <p>
          Research by Alireza Pourkeyvan [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ] indicates that the state-of-the-art model in mental disorder
detection is MentalRoBERTa [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ]. MentalRoBERTa is a variant of the RoBERTa model that is specialized
for mental health applications. It is pre-trained on a specialized corpus that includes texts from mental
health forums, clinical notes, and general language corpus. This pre-training enables MentalRoBERTa
to better understand and process language related to mental health, enhancing its applicability and
efectiveness in this domain.
        </p>
        <p>The model selected was AIMH/mental-roberta-large [14], a RoBERTa variant trained specifically
on mental health-related posts from Reddit. This model is available on the HuggingFace [15]
public hub (https://huggingface.co/AIMH/mental-roberta-large) and provides specialized capabilities for
understanding mental health discourse.</p>
        <p>We obtained two models by fine-tuning the base pre-trained model with two datasets: one using
Dataset 1 (RoBERTa-1) and the other using Dataset 2 (RoBERTa-2), with the second incorporating data
augmentation. Table 7 shows the configuration used in the fine-tuning process.</p>
        <p>Table 8 displays the results of each model on the development partition. The results indicate that
RoBERTa-2 obtained the best performance, a fine-tuned model with data augmentation. Consequently,
we used this model for Run1 in Task 2 of our participation.</p>
      </sec>
      <sec id="sec-3-4">
        <title>3.4. LongFormer Approach</title>
        <p>
          As previously mentioned, one of the major drawbacks of BERT-like or RoBERTa-like models based
on Transformers is their limited capacity to handle large contexts. However, there is a variant of
Transformers called LongFormer, which can process longer texts efectively [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ]
        </p>
        <p>LongFormer, which stands for “Long-Document Transformer,” is designed to process long contexts
more eficiently than traditional Transformer models such as BERT or RoBERTa. The LongFormer
architecture exhibits the following characteristics:
• New attention mechanism: An eficient attention mechanism that uses a sliding window, where
each token only attends to a fixed number of neighborhood tokens, reducing the complexity.
• Global attention selection: The architecture can select which tokens are globally attended and
which are just attended locally.</p>
        <p>The pre-trained model chosen was AIMH/mental-longformer-base-4096 [16] a pre-trained
LongFormer for the mental health domain. This model can be found in https://huggingface.co/AIMH/
mental-longformer-base-4096.</p>
        <p>As in with the RoBERTa model, we fine-tuned the LongFormer with the two datasets: Dataset 1
without data augmentation (LongFormer-1), and Dataset 2 with data augmentation (LongFormer-2). We
used the same fine-tuning parameters as in RoBERTa’s experimentation; the configuration is in Table 7.</p>
        <p>Table 9 shows the results of the experimentation, where LongFormer-2 (fine-tuned with data
augmentation) achieves better performance than LongFormer-1 (fine-tuned without data augmentation). We
used the two models in our participation, as Run2 and Run3</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Runs</title>
      <p>The rationale for selecting these models was to evaluate the significance of context in predicting
anorexia. Each model varies in its capacity to handle input length, allowing for the processing of
diferent context sizes. By comparing models with varying context-handling capabilities, we aim to
determine how the extent of context afects the accuracy and efectiveness of mental illness prediction.</p>
      <p>The results demonstrate that the SVM model, despite being less powerful in general, achieved
performance comparable to MentalRoBERTa. This can be attributed to the SVM’s ability to handle
large texts, leveraging the full context provided by the input data. On the other hand, LongFormer
models outperformed both BERT-like models and the SVM in this task. The performance of LongFormer
can be credited to its capability to process larger contexts while maintaining the powerful features
of Transformer-based models. This combination allows LongFormer to capture more comprehensive
contextual information, leading to more accurate predictions in mental illness detection tasks.</p>
      <sec id="sec-4-1">
        <title>4.1. Run Configuration</title>
        <p>Besides, to select the model for each run, the classification systems contained additional parameters
that needed to be set:
• For every round in the competition, we used as the input classifier a new sample created combining
the new message of the user with the previous ones.
• Each system has an initial context, in other words, we made our systems wait until the initial
context was suficiently large. This context was diferent in each system:
– SVM: An initial context of 50 tokens after the pre-process.</p>
        <p>– RoBERTa and LongFormer: An initial context of 100 tokens.
• The RoBERTa and LongFormer system has a limit of tokens, when the system was full we just
returned the last prediction made.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Results</title>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusion</title>
      <p>In this paper, we have presented the participation of the ELiRF-VRAIN team in Task 2 of eRisk at
CLEF 2024: early detection of signs of anorexia. In addition to testing classic classification models and
state-of-the-art Transformer models, we used LongFormers models to expand the context when making
the decision. In addition, a proposal for data augmentation was presented with successful results during
the training process.</p>
      <p>For future work, two lines of improvement are identified. On the one hand, try to improve early
detection so that the system does not need as much context to make the right decision; on the other
hand, use Explainable Artificial Intelligence (XAI) techniques to understand the system’s behavior
better.</p>
    </sec>
    <sec id="sec-7">
      <title>Acknowledgments</title>
      <p>This work is partially supported by MCIN/AEI/10.13039/501100011033, by the "European Union" and
“NextGenerationEU/MRR”, and by “ERDF A way of making Europe” under grants
PDC2021-120846C44 and PID2021-126061OB-C41. Partially supported by the Vicerrectorado de Investigación de la
Universitat Politècnica de València PAID-01-23. It is also partially supported by the Spanish Ministerio
de Universidades under the grant FPU21/05288 for university teacher training.
[14] AIMH, Mentalroberta: A robustly optimized bert pretraining approach for mental health, 2024.</p>
      <p>URL: https://huggingface.co/AIMH/mental-roberta-large, accessed: 2024-05-15.
[15] T. Wolf, L. Debut, V. Sanh, J. Chaumond, C. Delangue, A. Moi, P. Cistac, T. Rault, R. Louf, M.
Funtowicz, J. Davison, S. Shleifer, P. von Platen, C. Ma, Y. Jernite, J. Plu, C. Xu, T. L. Scao, S. Gugger,
M. Drame, Q. Lhoest, A. M. Rush, Transformers: State-of-the-art natural language processing,
Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System
Demonstrations (2020). URL: https://www.aclweb.org/anthology/2020.emnlp-demos.6.
[16] AIMH, Mentallongformer: A long-document transformer model for mental health, 2024. URL:
https://huggingface.co/AIMH/mental-longformer-base-4096, accessed: 2024-05-15.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>FEACAB</given-names>
            ,
            <surname>Anorexia</surname>
          </string-name>
          ,
          <year>2015</year>
          . URL: https://feacab.org/anorexia/, accessed:
          <fpage>2024</fpage>
          -05-28.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>J.</given-names>
            <surname>Parapar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Martín Rodilla</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. E.</given-names>
            <surname>Losada</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Crestani</surname>
          </string-name>
          , Overview of erisk 2024:
          <article-title>Early risk prediction on the internet</article-title>
          ,
          <source>in: Experimental IR Meets Multilinguality, Multimodality, and Interaction. 15th International Conference of the CLEF Association, CLEF 2024</source>
          , Springer International, Grenoble, France,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>J.</given-names>
            <surname>Parapar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Martín Rodilla</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. E.</given-names>
            <surname>Losada</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Crestani</surname>
          </string-name>
          , Overview of erisk 2024:
          <article-title>Early risk prediction on the internet (extended overview)</article-title>
          ,
          <source>in: Working Notes of the Conference and Labs of the Evaluation Forum CLEF</source>
          <year>2024</year>
          , CEUR Workshop Proceedings, Grenoble, France,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>A.</given-names>
            <surname>Vaswani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Shazeer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Parmar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Uszkoreit</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Jones</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. N.</given-names>
            <surname>Gomez</surname>
          </string-name>
          , Ł. Kaiser,
          <string-name>
            <surname>I. Polosukhin</surname>
          </string-name>
          ,
          <article-title>Attention is all you need</article-title>
          ,
          <source>Advances in Neural Information Processing Systems</source>
          <volume>30</volume>
          (
          <year>2017</year>
          ). URL: https://arxiv.org/abs/1706.03762, accessed:
          <fpage>2024</fpage>
          -05-15.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ott</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Goyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Du</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Joshi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Levy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Lewis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Zettlemoyer</surname>
          </string-name>
          , V. Stoyanov,
          <article-title>RoBERTa: A robustly optimized BERT pretraining approach</article-title>
          , arXiv preprint arXiv:
          <year>1907</year>
          .
          <volume>11692</volume>
          (
          <year>2019</year>
          ). URL: https://arxiv.org/abs/
          <year>1907</year>
          .11692.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>I.</given-names>
            <surname>Beltagy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. E.</given-names>
            <surname>Peters</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Cohan</surname>
          </string-name>
          ,
          <article-title>Longformer: The long-document transformer</article-title>
          , arXiv preprint arXiv:
          <year>2004</year>
          .
          <volume>05150</volume>
          (
          <year>2020</year>
          ). URL: https://arxiv.org/abs/
          <year>2004</year>
          .05150.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>A. M.</given-names>
            <surname>Mármol-Romero</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Moreno-Muñoz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F. M. P.</given-names>
            <surname>del Arco</surname>
          </string-name>
          ,
          <string-name>
            <surname>M. D.</surname>
          </string-name>
          Molina-González, M.-T. MartínValdivia, L. A.
          <string-name>
            <surname>Ureña-López</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Montejo-Ráez</surname>
          </string-name>
          , Overview of mentalriskes at iberlef 2024:
          <article-title>Early detection of mental disorders risk in spanish</article-title>
          ,
          <source>Procesamiento del Lenguaje Natural</source>
          <volume>73</volume>
          (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>D. E.</given-names>
            <surname>Losada</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Crestani</surname>
          </string-name>
          ,
          <article-title>A test collection for research on depression and language use, in: Experimental IR Meets Multilinguality, Multimodality, and Interaction</article-title>
          .
          <source>Proceedings of the 7th International Conference of the CLEF Association (CLEF</source>
          <year>2016</year>
          ),
          <year>2016</year>
          , pp.
          <fpage>28</fpage>
          -
          <lpage>39</lpage>
          . URL: https: //doi.org/10.1007/978-3-
          <fpage>319</fpage>
          -44564-
          <issue>9</issue>
          _3. doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>319</fpage>
          -44564-
          <issue>9</issue>
          _
          <fpage>3</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>F.</given-names>
            <surname>Pedregosa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Varoquaux</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Gramfort</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Michel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Thirion</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Grisel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Blondel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Prettenhofer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Weiss</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Dubourg</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Vanderplas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Passos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Cournapeau</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Brucher</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Perrot</surname>
          </string-name>
          , É. Duchesnay,
          <article-title>Scikit-learn: Machine learning in python</article-title>
          ,
          <source>Journal of Machine Learning Research</source>
          <volume>12</volume>
          (
          <year>2011</year>
          )
          <fpage>2825</fpage>
          -
          <lpage>2830</lpage>
          . URL: https://jmlr.org/papers/v12/pedregosa11a.html.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>L. X.</given-names>
            <surname>Yuan</surname>
          </string-name>
          , distilbert
          <article-title>-base-multilingual-cased-sentiments-student (</article-title>
          <source>revision 2e33845)</source>
          ,
          <year>2023</year>
          . URL: https://huggingface.co/lxyuan/distilbert-base
          <article-title>-multilingual-cased-sentiments-student</article-title>
          .
          <source>doi:10</source>
          .57967/hf/1422.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>N. I.</surname>
          </string-name>
          <article-title>of Mental Health, Eating disorders</article-title>
          , n.d. URL: https://www.nimh.nih.gov/health/statistics/ eating-disorders, accessed:
          <fpage>2024</fpage>
          -05-30.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>A.</given-names>
            <surname>Pourkeyvan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Safa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Sorourkhah</surname>
          </string-name>
          ,
          <article-title>Harnessing the power of hugging face transformers for predicting mental health disorders in social networks</article-title>
          ,
          <source>IEEE Access 12</source>
          (
          <year>2024</year>
          )
          <fpage>28025</fpage>
          -
          <lpage>28035</lpage>
          . URL: http://dx.doi.org/10.1109/ACCESS.
          <year>2024</year>
          .
          <volume>3366653</volume>
          . doi:
          <volume>10</volume>
          .1109/access.
          <year>2024</year>
          .
          <volume>3366653</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>S.</given-names>
            <surname>Ji</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , L. Ansari,
          <string-name>
            <given-names>J.</given-names>
            <surname>Fu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Tiwari</surname>
          </string-name>
          , E. Cambria,
          <article-title>MentalBERT: Publicly available pretrained language models for mental healthcare</article-title>
          , in: N.
          <string-name>
            <surname>Calzolari</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Béchet</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Blache</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Choukri</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Cieri</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Declerck</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Goggi</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Isahara</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Maegaard</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Mariani</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Mazo</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Odijk</surname>
          </string-name>
          , S. Piperidis (Eds.),
          <source>Proceedings of the Thirteenth Language Resources and Evaluation Conference</source>
          , European Language Resources Association, Marseille, France,
          <year>2022</year>
          , pp.
          <fpage>7184</fpage>
          -
          <lpage>7190</lpage>
          . URL: https://aclanthology.org/
          <year>2022</year>
          . lrec-
          <volume>1</volume>
          .
          <fpage>778</fpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>