<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Early Detection of Gambling Addiction Risk in Spanish: purple_john at MentalRiskES2025</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Cristina Damov</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Adina Ion</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Crystal Schlupek</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>University of Bucharest</institution>
          ,
          <addr-line>St. Academiei 14, Bucharest</addr-line>
          ,
          <country country="RO">Romania</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <abstract>
        <p>This paper presents our submission to the MentalRiskES 2025 shared task, focused on the early detection of gambling addiction risk from Spanish-language social media posts. Our approach explored a range of deep learning and traditional machine learning models, including CNN, MobileNet, Inception, and transformer-based architectures such as RoBERTa and BETO, combined with pre-trained embeddings like GloVe and FastText. Despite extensive experimentation and model diversity, classification performance remained modest, with our best accuracy reaching only 51% on final tests. These results suggest potential limitations not only in model capacity but also in the dataset and task formulation, where user-level annotations and subtle language diferences may be insuficient to reliably distinguish risk levels. Further improvements are needed to better detect behavioral risk in informal and multilingual data.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;mental health</kwd>
        <kwd>NLP</kwd>
        <kwd>Spanish</kwd>
        <kwd>social media</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Mental health concerns have become increasingly visible in recent years, with global estimates from
the World Health Organization indicating that one in eight people is afected by a mental disorder [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
The COVID-19 pandemic further exacerbated many of these issues, including anxiety, depression, and
various forms of addiction [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. As people increasingly turn to social media to share their thoughts and
experiences, these platforms have become a valuable source of information for identifying early signs
of mental health risks.
      </p>
      <p>
        While much existing work in the field of computational mental health has focused on conditions like
depression or suicidal ideation, gambling addiction remains relatively underexamined, particularly
in non-English languages. Gambling disorder is often dificult to detect, as its linguistic markers are
subtler and more context-dependent compared to other mental health conditions.
This paper presents our submission to the third edition of a shared task, focused on early detection
of mental health risks in Spanish social media, organized as part of IberLEF 2025 [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. This year’s
task introduced two new subtasks: one centered specifically on detecting gambling disorder risk, and
another on identifying types of addiction. Unlike previous editions, which focused more broadly on
psychological risk, this year’s theme required participants to address the nuances of addictive behaviors
in a Spanish-language context.
      </p>
      <p>Our goal was to develop an NLP system capable of identifying high risk of developing a gambling
addiction from a stream of user posts. In doing so, we aimed not only to contribute to the technical
challenge but also to support the broader goal of enabling earlier and more targeted mental health
interventions, especially in underrepresented languages and conditions.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Task</title>
      <p>
        2.1. Data
Our work addressed Subtask 1: Risk Detection of Gambling Disorders [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] , which involved classifying
messages as suggesting either a high risk or a low risk of the user developing a gambling addiction. The
challenge was framed as a binary classification problem, where systems had to make this determination
based on a user’s timeline of messages. Evaluation metrics included not just prediction scores, but also
computational resources (duration, emissions, energy consumed etc.)
The data provided consisted of Spanish-language social media messages, sent on Telegram groups or
Twitch chats. These annotations posed additional challenges due to the informal nature of the text,
potential noise, and the subtle linguistic cues that may signal gambling-related behavior. This subtask
aimed to push forward the use of NLP tools for behavioral risk assessment, particularly in less-resourced
languages and nuanced domains like gambling addiction.
      </p>
      <sec id="sec-2-1">
        <title>2.2. Submission Procedure</title>
        <p>The predictions were evaluated in an online, round-based manner. The program setup included server
connection and API calls to receive a batch of messages from users and submit a prediction for each
one before proceeding to the next round. This workflow simulated real-time detection and increased
the challenge, as any message could be the user’s last.</p>
        <p>Notably, true labels were set at user level, however predictions were made at message level. This meant
models were also challenged by the fact that some messages were not relevant to the user’s risk label
and there was no user message history to predict contextually.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Methodology</title>
      <sec id="sec-3-1">
        <title>3.1. Preprocessing</title>
        <p>The quality of the input data has a significant impact on the efectiveness and accuracy of text
classification models and sentiment analysis in the field of natural language processing (NLP). Several data
pretreatment procedures were tested to enhance the accuracy and consistency of sentiment analysis
prediction in binary text categorization.</p>
        <sec id="sec-3-1-1">
          <title>Lowercasing</title>
          <p>Converting all text to lowercase ofers several benefits, including text normalization, vocabulary
reduction, compatibility with word embedding models, and ensuring consistency in text processing
pipelines. This standardizes the text and ensures that the model does not treat words with diferent
cases as distinct entities. While this step improves consistency, it may also lead to a loss of information
from capitalized proper nouns or acronyms.</p>
        </sec>
        <sec id="sec-3-1-2">
          <title>Punctuation Removal Punctuation marks, though important for grammatical structure, generally do not add semantic value in this task. Removing punctuation reduces noise and simplifies tokenization, allowing the model to focus on word-level content without distraction from syntactic elements.</title>
        </sec>
        <sec id="sec-3-1-3">
          <title>Lemmatization Lemmatization reduces words to their dictionary base form, decreasing word variations to a single representative. This reduces feature sparsity and emphasizes core meanings. Unlike stemming, lemmatization considers part of speech and context, making it more linguistically accurate.</title>
        </sec>
        <sec id="sec-3-1-4">
          <title>Stemming</title>
          <p>Stemming is another normalization technique that removes sufixes to reduce words to their root form. It
is a faster, simpler alternative to lemmatization, but may produce non-standard root forms and overlook
contextual meaning. Despite its crudeness, stemming can be efective in reducing dimensionality when
linguistic precision is not critical.</p>
        </sec>
        <sec id="sec-3-1-5">
          <title>Stop Words Removal Stop words are common words that occur frequently in a language. These words do not carry significant sentiment value and can be safely removed from the text, reducing noise in the dataset and making it more focused and meaningful for sentiment analysis.</title>
        </sec>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Additional data used</title>
        <p>No external or augmented data was used. The model was trained strictly on the dataset provided in the
shared task. This was a conscious choice, as most available mental health datasets focus on issues like
depression or anxiety and don’t align well with the specific language and behaviors related to gambling
addiction. We prioritized maintaining domain relevance and consistency with the task’s objectives,
even if it limited the amount of training data available.</p>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Advanced Natural Language Processing Models</title>
        <p>
          To improve early detection of mental health conditions in Spanish social media comments, several
advanced NLP models were implemented. These models capture diferent linguistic patterns and
contextual nuances, each contributing to the classification process. The models used are:
• BiLSTM: Applied for capturing bidirectional context in text, which is valuable for detecting subtle
emotional expressions.
• CNN with FastText [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ]: Used to identify local linguistic patterns. This combination was efective
in processing short, informal messages.
• MobileNet [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ](Adapted for Text): A lightweight deep neural network adapted for text classification
using word/document embeddings. It achieved strong accuracy ( 0.73) while maintaining low
computational cost, making it suitable for real-time applications.
• Transformer-based Models (RoBERTa [9] and BETO [10]): These models were evaluated for their
deep contextual embeddings. Despite their potential, they achieved less than 60% accuracy in
this setting, likely due to domain mismatch and limited fine-tuning.
• LightGBM with FastText: This gradient boosting model efectively handled high-dimensional
text features and performed better than simpler models like logistic regression in several
configurations.
• Logistic Regression: Used as a baseline for comparison. When paired with quality embeddings
(e.g. FastText), it showed competitive results, validating its role as a strong, interpretable baseline.
        </p>
      </sec>
      <sec id="sec-3-4">
        <title>3.4. Classifiers and Ensemble Techniques</title>
        <p>To enhance performance, various classifiers and ensemble methods were applied. These ranged from
lightweight models to more complex deep learning architectures, ofering a broad perspective on their
trade-ofs:
• Logistic Regression with Embeddings: Served as a strong baseline when used with semantic-rich
embeddings like FastText and GloVe [11]. It was also useful in ensemble strategies due to its
probabilistic output.
• LightGBM: Demonstrated robustness and eficiency, especially in handling class imbalance. It
integrated well with FastText features and benefited from histogram-based optimization for faster
training.
• FastText + CNN: This configuration efectively captured local patterns indicative of mental distress
and provided interpretable, eficient classification.
• FastText + Logistic Regression: A computationally eficient strategy that maintained stable and
reliable performance, especially in resource-constrained settings.
• GloVe-Based Models: Both English and Spanish embeddings were tested, with English GloVe
surprisingly outperforming Spanish in certain cases, possibly due to language mixing in social
media text.
• MobileNet with Dimensionality Reduction: This setup delivered one of the best trade-ofs between
accuracy ( 0.73) and speed. It showed promise for applying transfer learning techniques from
vision models to NLP tasks.
• Transformer Models (RoBERTa and BETO): Although powerful, these models underperformed
(&lt;60% accuracy on validation data), likely due to insuficient fine-tuning and mismatch with the
target domain. Their computational cost also limited practical deployment in this project.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Results and Discussion</title>
      <p>Each model was evaluated using accuracy, F1 score and early detection performance, as summarized in
Table 1:
• BERT was initially expected to deliver the best results given its former state-of-the-art status
and good results in other studies [12][13], but in practice its performance sufered under tight
resource constraints. It required substantial training time and memory, leading us to remove it
from further consideration.
• GloVe embeddings were tested in both English and Spanish. The English models outperformed the</p>
      <p>Spanish ones, likely due to superior slang and informal-speech coverage in the English vectors.
• BiLSTM, as a powerful recurrent architecture, achieved solid efectiveness but at the cost of high
computational and memory demands when setting trainable=True.
• MobileNet proved to be fast, lightweight, and surprisingly accurate, making it well suited for
resource-limited environments, as illustrated by its confusion matrix in Figure 1.
• Inception achieved performance and inference speed on par with MobileNet, making it a reliable
choice among the tested models (Figure 2).
Applying dimensionality reduction dramatically reduced training time (especially for BERT) but did not
yield any appreciable gains in accuracy.</p>
      <sec id="sec-4-1">
        <title>4.1. Hyperparameters and Training Configuration</title>
        <p>We experimented with a set of standard hyperparameters. Minor tuning was performed using manual
adjustments and a grid search over key parameters, such as learning rate and batch size, to optimize
performance across diferent models. The goal was to balance training stability, speed, and generalization
while staying within computational constraints.</p>
        <p>The final configuration was:</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Final Model Selection</title>
        <p>Some preprocessing steps, like lemmatization, were tested early on but later excluded from the final
pipeline. They slowed down text processing and didn’t lead to any improvement in accuracy. Since
they added extra complexity without ofering real benefits, we chose to leave them out and stick to a
simpler setup that was faster and more eficient (lowercasing, punctuation and stopword removal).
Although multiple embeddings were tested, including FastText and GloVe, both English and
Spanishspecific versions, we ultimately kept only the English GloVe embeddings in the final models. Surprisingly,
they consistently outperformed the Spanish ones, likely due to better coverage of informal language
and cross-lingual slang often used in the dataset. This choice also helped simplify the pipeline and
reduced training time without hurting accuracy.</p>
        <p>The comparative performance of the final models on validation data is presented in Table 2.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Error Analysis</title>
      <p>Despite using various filters and preprocessing steps (stopword removal, punctuation stripping,
lowercasing, lemmatization), data issues persisted. The dataset was annotated at user level, which caused
mismatches, for example, meaningless messages like “Si” (“Yes”) were marked as risky. This limited
model learning and introduced noise.</p>
      <p>A detailed review of failed predictions revealed several causes:
• Noisy inputs: messages like "jajaja" (“hahaha”), emojis, or very short comments were incorrectly
labeled as risky, afecting training quality and creating a model prone to many false positive
labels. This is also reflected in the fact that several models have lower precision compared to
their F1 scores.
• Weak lexical distinction: no strong trends or word-level diferences between risk 0 and risk 1,
making separation dificult.
• Linguistic limitations: use of slang, regional Spanish, or code-switching (ES/EN) reduced
embedding efectiveness.
• Domain-specific overlap: both risk and non-risk users share the same community slang and
terminology (betting or crypto terms), reducing lexical contrast typically used for classification.
The comparative analysis of messages labeled risk 0 versus risk 1 revealed no meaningful diferences.
Although Figure 3 (hourly frequency plot) and Figure 4 (emotion-count chart) display certain trends,
such as peaks of activity at specific hours and a predominance of anticipation and negative emotions
(which is unsurprising in a gambling context); these patterns do not clearly separate the two risk groups.
Likewise, simple metrics like word count and other extracted features fail to distinguish risk 0 from risk
1 in any reliable way.</p>
    </sec>
    <sec id="sec-6">
      <title>6. Resource Impact and Emissions</title>
      <p>In addition to evaluating classification performance, it is increasingly important to consider the
computational and environmental cost of training and deploying NLP models. This is especially relevant
in shared tasks and evaluation campaigns where eficiency and sustainability are valued alongside
predictive accuracy.</p>
      <p>Our submissions included three distinct models, each trained and evaluated independently, yet all
sharing a common emphasis on energy eficiency and hardware-conscious design. The results for these
three runs and baseline models, including F1 score, emissions, and CPU usage, are summarized in Table
3.</p>
      <p>Precision (macro) F1 Score (macro)</p>
      <p>Duration Mean</p>
      <p>Emissions Mean</p>
      <p>CPU Ener</p>
      <p>These submissions reflect the team’s consistent focus on resource-conscious model design. Although
transformer-based architectures like RoBERTa and BETO were evaluated, they were ultimately excluded
due to modest performance and high resource demands.</p>
      <p>While the accuracy of these models ( 51-52%) did not reach the top rankings, the energy consumed was
of lower magnitude than that of GPU-based or transformer-based solutions. Duration and emissions
were similar between the runs. This highlights a crucial trade-of in applied machine learning: models
that are sustainable and eficient may be more viable for real-world deployment, especially in
resourceconstrained environments.</p>
    </sec>
    <sec id="sec-7">
      <title>7. Conclusions</title>
      <p>This study, carried out as part of the MentalRiskES 2025 shared task, underlined the dificulty of
detecting gambling addiction risk from short, informal social media messages in Spanish. Despite
testing a wide range of models, including traditional classifiers, CNNs, MobileNet, and transformers,
performance remained modest, with limited distinction between high-risk and low-risk users. MobileNet
and Inception achieved the best trade-of between speed and accuracy, while transformer models
underperformed in this domain. Much of the challenge stemmed from noisy data, inconsistent labeling,
and a lack of strong linguistic cues to separate classes.</p>
      <p>While the results were not especially strong, they reflect the complexity of the task and the need for
better data, clearer risk definitions, and more context-aware approaches. Future work should focus on
these aspects to improve real-world applicability.</p>
    </sec>
    <sec id="sec-8">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the author(s) used Grammarly and ChatGPT in order to: paraphrase,
identify and correct typos and grammatical mistakes. After using these tools, the author(s) reviewed
and edited the content as needed and take(s) full responsibility for the publication’s content.
[9] Liu, Y., Ott, M., Goyal, N., Du, J., Joshi, M., Chen, D., Levy, O., Lewis, M., Zettlemoyer, L., &amp; Stoyanov,
V. (2019). RoBERTa: A robustly optimized BERT pretraining approach [Technical report]. arXiv.
https://arxiv.org/abs/1907.11692
[10] Cañete, J., Chaperon, G., Fuentes, R., Ho, J.-H., Kang, H., &amp; Pérez, J. (2020). Spanish pre-trained
BERT model and evaluation data. In Proceedings of the Practical Machine Learning for Developing
Countries Workshop at ICLR 2020 (PML4DC).
[11] Pennington, J., Socher, R., &amp; Manning, C. D. (2014). GloVe: Global vectors for word representation.</p>
      <p>In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP)
(pp. 1532–1543). https://aclanthology.org/D14-1162
[12] Smith, E., Peters, J., &amp; Reiter, N. (2024). Automatic detection of problem-gambling signs from
online texts using large language models. PLOS Digital Health, 3(9), e0000605. https://doi.org/10.
1371/journal.pdig.0000605
[13] Bucur, A.-M., Cosma, A., &amp; Dinu, L. P. (2021). Early risk detection of pathological gambling,
self-harm and depression using BERT. arXiv. https://doi.org/10.48550/arXiv.2106.16175
[14] González-Barba, J. A., Chiruzzo, L., &amp; Jiménez-Zafra, S. M. (2025). Overview of IberLEF 2025:
Natural Language Processing Challenges for Spanish and other Iberian Languages. In Proceedings
of the Iberian Languages Evaluation Forum (IberLEF 2025), co-located with the 41st Conference of
the Spanish Society for Natural Language Processing (SEPLN 2025). CEUR-WS.org.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>Mármol-Romero</surname>
            ,
            <given-names>A. M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Álvarez-Ojeda</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Moreno-Muñoz</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <article-title>Plaza-del-</article-title>
          <string-name>
            <surname>Arco</surname>
            ,
            <given-names>F. M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>MolinaGonzález</surname>
          </string-name>
          , M. D.,
          <string-name>
            <surname>Martín-Valdivia</surname>
            ,
            <given-names>M. T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ureña-López</surname>
            ,
            <given-names>L. A.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Montejo-Ráez</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          (
          <year>2025</year>
          ). Overview of MentalRiskES at IberLEF 2025:
          <article-title>Early Detection of Mental Disorders Risk in Spanish</article-title>
          .
          <source>Procesamiento del Lenguaje Natural</source>
          ,
          <volume>75</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <surname>Álvarez-Ojeda</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cantero-Romero</surname>
            ,
            <given-names>M. V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Semikozova</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Montejo-Ráez</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          (
          <year>2025</year>
          ).
          <article-title>The PRECOM-SM corpus: Gambling in Spanish social media</article-title>
          .
          <source>In Proceedings of the 31st International Conference on Computational Linguistics</source>
          (pp.
          <fpage>17</fpage>
          -
          <lpage>28</lpage>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>González-Barba</surname>
            ,
            <given-names>J. A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chiruzzo</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Jiménez-Zafra</surname>
            ,
            <given-names>S. M.</given-names>
          </string-name>
          (
          <year>2025</year>
          ).
          <source>In Proceedings of the Iberian Languages Evaluation Forum (IberLEF</source>
          <year>2025</year>
          ),
          <article-title>co-located with the 41st Conference of the Spanish Society for Natural Language Processing (SEPLN</article-title>
          <year>2025</year>
          ).
          <article-title>CEUR-WS.org</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <surname>World Health Organization.</surname>
          </string-name>
          (
          <year>2022</year>
          , June 8).
          <article-title>Mental disorders</article-title>
          . Retrieved from https://www.who. int/news-room/fact-sheets/detail/mental-disorders
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5] World Health Organization.
          <article-title>(2020, October 5). COVID-19 disrupting mental health services in most countries, WHO survey</article-title>
          . Retrieved from https://www.who.int/news/item/ 05-10-2020-covid-19
          <string-name>
            <surname>-</surname>
          </string-name>
          disrupting
          <article-title>-mental-health-services-in-most-countries-who-survey</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>Montejo-Ráez</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mármol-Romero</surname>
            ,
            <given-names>A. M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Álvarez-Ojeda</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <article-title>Plaza-del-</article-title>
          <string-name>
            <surname>Arco</surname>
            ,
            <given-names>F. M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ureña-López</surname>
            ,
            <given-names>L. A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Martín-Valdivia</surname>
            ,
            <given-names>M. T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Molina-González</surname>
            ,
            <given-names>M. D.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Moreno-Muñoz</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          (
          <year>2025</year>
          ).
          <source>MentalRiskES</source>
          <year>2025</year>
          :
          <article-title>Early detection of mental disorders risk in Spanish</article-title>
          .
          <source>IberLEF</source>
          <year>2025</year>
          . Retrieved May 17,
          <year>2025</year>
          , from https://sites.google.com/view/mentalriskes2025/tasks
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>Bojanowski</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Grave</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Joulin</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Mikolov</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          (
          <year>2017</year>
          ).
          <article-title>Enriching word vectors with subword information</article-title>
          .
          <source>Transactions of the Association for Computational Linguistics</source>
          ,
          <volume>5</volume>
          ,
          <fpage>135</fpage>
          -
          <lpage>146</lpage>
          . https: //doi.org/10.1162/tacl_a_
          <fpage>00051</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <surname>Howard</surname>
            ,
            <given-names>A. G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhu</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kalenichenko</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Weyand</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Andreetto</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Adam</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          (
          <year>2017</year>
          ).
          <article-title>MobileNets: Eficient convolutional neural networks for mobile vision applications</article-title>
          [
          <source>Technical report]</source>
          . arXiv. https://arxiv.org/abs/1704.04861
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>