<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Balancing Translation Quality and Environmental Impact: Comparing Large and Small Language Models</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Antonio Castaldo</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Petra Giommarelli</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Johanna Monti</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>University of Naples L'Orientale</institution>
          ,
          <addr-line>Via Chiatamone, 61/62, 80121 Naples</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>University of Pisa, Largo Bruno Pontecorvo</institution>
          ,
          <addr-line>3, 56127 Pisa</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <abstract>
        <p>Large Language Models (LLMs) have demonstrated remarkable performance in machine translation (MT), specifically concerning high-resource European languages. However, their extensive computational requirements raise sustainability concerns. This paper investigates the potential of smaller, fine-tuned language models as a more sustainable alternative for MT tasks. We conduct a comparative analysis of model performance in terms of translation quality and CO2eq emissions, and examine the key errors associated with using smaller models. Furthermore, we propose a novel metric that balances translation quality against environmental impact, aiming to inform more sustainable model selection in MT research and practice.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;machine translation</kwd>
        <kwd>large language models</kwd>
        <kwd>sustainability</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        larger models. This setup allows us to assess the
realworld viability of small models for machine translation
MT has been a core topic in natural language processing when fine-tuned for specific language pairs and domains.
(NLP) for several decades, evolving from rule-based sys- We conduct a comprehensive analysis of model
perfortems to statistical methods, and more recently to neural mance, in terms of translation quality and CO2eq
emismachine translation (NMT) and transformer-based mod- sions, validating our results with a human evaluation
els. The emergence of LLMs has significantly advanced of the key errors associated with each model. Finally,
the state-of-the-art in MT, demonstrating remarkable we introduce a metric called Carbon-Adjusted Quality
performance on various NLP tasks [
        <xref ref-type="bibr" rid="ref8">1</xref>
        ]. Score (CAQS), designed to facilitate sustainable model
se
      </p>
      <p>
        Their ability to generate fluent, context-aware trans- lection, that quantifies the trade-of between translation
lations in diferent domains has positioned LLMs at the quality and sustainability.
forefront of MT research [
        <xref ref-type="bibr" rid="ref9">2</xref>
        ]. Their ability to model
context, semantics, and discourse phenomena makes them
highly attractive for both academic and industrial trans- 2. Background
lation applications.
      </p>
      <p>
        However, this performance comes at a significant envi- 2.1. LLMs and Translation
ronmental cost. Training and deploying LLMs consumes LLMs have achieved state-of-the-art results in MT, by
enormous computational resources, leading to consider- leveraging extensive pretraining on multilingual corpora,
able carbon emissions and infrastructure demands [
        <xref ref-type="bibr" rid="ref10 ref11">3, 4</xref>
        ]. enabling them to deliver remarkable performance across
These challenges have prompted the exploration of more a wide range of domains and language pairs [
        <xref ref-type="bibr" rid="ref2">6</xref>
        ]. In
consustainable alternatives. trast to NMT systems, which rely primarily on
paral
      </p>
      <p>
        This paper investigates whether smaller language mod- lel corpora, LLMs are pretrained on massive web-scale
els can serve as eficient and environmentally sustainable monolingual and multilingual datasets. This enables
valid alternatives to LLMs in MT. Specifically, we will them to generate high-quality translations even in
doifne-tune the Gemma-3-4B[
        <xref ref-type="bibr" rid="ref1">5</xref>
        ] model on a parallel English- mains where parallel data is limited [
        <xref ref-type="bibr" rid="ref3">7</xref>
        ].
Italian (EN-IT) parallel corpus, and evaluate its perfor- Notably, GPT-based models excel at producing
conmance, with human and automatic evaluation, against textually accurate translations, efectively capturing
disCLiC-it 2025: Eleventh Italian Conference on Computational Linguis- course relations and maintaining sentence-level
cohertics, September 24 — 26, 2025, Cagliari, Italy ence. They consistently outperform encoder-decoder
† The authors contributed equally. architectures such as Transformer-big and M2M100,
par$ antonio.castaldo@phd.unipi.it (A. Castaldo); ticularly in zero-shot and few-shot settings [
        <xref ref-type="bibr" rid="ref4">8</xref>
        ].
petra.giommarelli@phd.unipi.it (P. Giommarelli); jmonti@unior.it Moreover, LLMs support document-level translation
(J. M00o0n9t-i0)008-3325-787X (A. Castaldo) by leveraging discourse-aware context windows, which
© 2025 Copyright for this paper by its authors. Use permitted under Creative Commons License enable the maintenance of lexical cohesion and consistent
Attribution 4.0 International (CC BY 4.0).
resolution of anaphoric references across sentences [
        <xref ref-type="bibr" rid="ref5">9</xref>
        ]. due to optimizations in model architecture and training
This capability results in more fluent translations, making strategy [14].
      </p>
      <p>LLMs increasingly favored in professional translation Moreover, recent studies show that even highly
comsettings. plex capabilities like multi-step reasoning, previously</p>
      <p>
        The adoption of LLMs, however, requires substantial thought to emerge only in models over 100B parameters,
computational resources and infrastructure, which may can be acquired by SLMs through targeted fine-tuning
not be feasible for all organizations or languages. Beyond and distillation. Distilling chain-of-thought reasoning
these practical limitations, the widespread adoption of abilities, for instance, from GPT-3.5 into FlanT5 variants
LLMs also raises significant concerns about their envi- (250M to 3B) resulted in significant performance
improveronmental sustainability. ments on math reasoning tasks without the need for full
retraining of the model’s weights [
        <xref ref-type="bibr" rid="ref13">16</xref>
        ].
2.2. LLMs Sustainability A comprehensive survey of SLMs underscores the
value of model compression techniques such as pruning,
While Large Language Models (LLMs) have enabled re- quantization, and knowledge distillation. These enable
markable progress in NLP, their growing environmental the deployment of eficient models on mobile and edge
defootprint raises important sustainability concerns. Train- vices while maintaining competitive accuracy for many
ing large-scale models such as GPT-3, with hundreds tasks [
        <xref ref-type="bibr" rid="ref14">17</xref>
        ]. The adoption of SLMs is particularly
promisof billions of parameters, can consume up to 1.3 GWh ing for democratizing NLP, enabling smaller institutions
of electricity, comparable to the yearly energy usage of and low-resource languages to benefit from modern AI
more than 100 US homes [
        <xref ref-type="bibr" rid="ref6">10</xref>
        ]. This results in hundreds of without the environmental or infrastructural burden of
tons of CO2 emissions, depending on the carbon intensity LLMs.
of the power grid. In this study, we evaluate whether SLMs, when
com
      </p>
      <p>
        In addition to training, the inference phase of LLMs bined with modern fine-tuning strategies and lightweight
also significantly contributes to their overall carbon foot- architectures, could ofer a pragmatic and sustainable
print, particularly in large-scale deployments. While the path forward for machine translation and other NLP
apenergy cost of a single inference is lower than that of plications.
training, the cumulative emissions can become
substantial depending on usage patterns. For example, serving
a single ChatGPT prompt may emit over 4g of CO2eq, 3. Fine-tuning a SLM
more than 20 times the emissions of a typical web search
[
        <xref ref-type="bibr" rid="ref7">11</xref>
        ].
      </p>
      <p>The same study emphasizes that total environmental
impact depends on a combination of factors: model size,
batch size, and hardware type. The latter reflects the
impact of producing high-performance GPUs, which
involves substantial embodied carbon emissions. Although
these emissions occur at production time, they contribute
to the model’s overall environmental cost throughout its
operational lifetime.</p>
      <sec id="sec-1-1">
        <title>To demonstrate the efectiveness of using SLMs as sus</title>
        <p>
          tainable alternatives to larger, more resource-intensive
models in machine translation, we compare two
stateof-the-art models: GPT-4o-mini [
          <xref ref-type="bibr" rid="ref15">18</xref>
          ] and an open-source
model, Gemma-3-4B [
          <xref ref-type="bibr" rid="ref12">15</xref>
          ], which is significantly smaller
than its OpenAI counterpart.
        </p>
        <p>
          We fine-tune Gemma-3-4B on a carefully curated
subset of the OpenSubtitles corpus, obtained from the Opus
Corpus [
          <xref ref-type="bibr" rid="ref16">19</xref>
          ]. We evaluate both models on a held-out
test set of 400 segments for the English–Italian (EN-IT)
language pair and present our findings.
        </p>
        <sec id="sec-1-1-1">
          <title>2.3. Small Language Models</title>
        </sec>
      </sec>
      <sec id="sec-1-2">
        <title>Recent research has emphasized the growing feasibility</title>
        <p>and importance of SLMs as eficient alternatives to LLMs
in constrained environments [12, 13]. SLMs, typically
ranging from hundreds of millions to a few billion
parameters, are substantially more resource-eficient and
accessible, especially when tailored to specific tasks.</p>
        <p>
          SLMs benefit from architectural simplifications, such
as compact tokenizers and reduced model width and
depth, which are optimized to preserve key
capabilities while minimizing parameter overhead [14]. Small
models, like Gemma [
          <xref ref-type="bibr" rid="ref12">15</xref>
          ] and PanGu- -1.5B Pro model
with only a few billion parameters have recently
outperformed much larger models on several benchmarks
        </p>
        <sec id="sec-1-2-1">
          <title>3.1. Dataset Curation</title>
        </sec>
      </sec>
      <sec id="sec-1-3">
        <title>For our experiments, we focused on the EN–IT subset</title>
        <p>of the OpenSubtitles corpus, made available through the
Opus Corpus repository. While OpenSubtitles is a rich
resource for dialogue-based translation data, it also
contains a considerable amount of noise due to its automatic
extraction and alignment process. Therefore, careful
curation was necessary to ensure the quality and relevance
of the dataset.</p>
        <p>
          We began by removing duplicate entries and any
empty lines. Following this, we applied the langdetect
[
          <xref ref-type="bibr" rid="ref17">20</xref>
          ] tool to verify the language of each sentence. This
step was essential, as web-crawled corpora, although environmental impact of our training process.
intended to be language-specific, occasionally contain CodeCarbon is a Python library that estimates
carsegments in other languages. Sentences detected to be in bon emissions by tracking the energy consumption of
languages outside our target pair, and that could not be computing resources (CPU, GPU, RAM) during code
execlassified with a high confidence score, were filtered out. cution and combining this data with the carbon intensity
        </p>
        <p>
          Finally, we applied COMET-QE [21], a quality estima- of the electricity grid based on geographic location.
tion model, to score the remaining sentence pairs. Using The fine-tuning session consumed approximately 0.65
these scores, we selected the top 100,000 highest-quality kWh, resulting in an estimated 162 g CO2eq under an
translations for use in our fine-tuning experiments. The average EU grid intensity of 250 gCO2/kWh.
strategy of mining large datasets and selecting top-k
sentence pairs based on quality metrics for fine-tuning helps 3.3. Gemma-3 Evaluation
to further filter out noisy segments and ensures that the
limited available data contribute maximally to model
training [
          <xref ref-type="bibr" rid="ref18">22</xref>
          ]. This approach is consistent with our goal
of reducing computational costs. By carefully curating a
smaller but higher-quality dataset, we limit energy
consumption and the associated environmental costs, while
maximizing translation performance.
        </p>
      </sec>
      <sec id="sec-1-4">
        <title>We conduct our evaluation on a held-out test set of 400</title>
        <p>segments from the same corpus, ensuring no overlap
with the training data. Table 2 reports the evaluation
of EN–IT translation performance for Gemma-3-4B
before and after LoRA fine-tuning, using BLEU [ 27], chrF
[29], and COMET [30] as quality metrics. Our fine-tuned
Gemma-3-4B model, with only 0.42% of additional
trainable parameters, shows a notable improvement over the
base version, achieving a +4 point gain in BLEU, a modest
increase in chrF, and a +1 point gain in COMET. These
results place our model on par with GPT-4o in COMET
and above GPT-4o-mini in all three metrics.</p>
        <p>
          In addition to performance, we also measure the
environmental impact of inference using the CodeCarbon
library. The estimated carbon emissions per inference for
the fine-tuned model are approximately 0.028g CO 2eq,
twice that of the base model, but significantly lower than
GPT-4o models, each exceeding 0.42g per inference as
estimated in a relevant study [
          <xref ref-type="bibr" rid="ref20">31</xref>
          ].
        </p>
        <p>Our evaluation demonstrates that fine-tuning
Gemma3-4B with LoRA leads to competitive performance gains
with low additional environmental cost.</p>
        <sec id="sec-1-4-1">
          <title>3.2. Training</title>
          <p>
            The Gemma-3-4B model was fine-tuned for three epochs
using Low-Rank Adaptation (LoRA) [
            <xref ref-type="bibr" rid="ref19">23</xref>
            ], a fine-tuning
technique which injects small trainable matrices in the
model’s weights. The adoption of LoRA for fine-tuning
has shown strong empirical results in machine
translation [24, 25], enhancing eficiency, while reducing train- 4. Quality-Sustainability Trade-Of
ing time and computational costs. As demonstrated in
experiments conducted by [26], fine-tuning with LoRA In our second experiment, to further assess the viability
obtained the same improvements in terms of BLEU score of trading of quality for sustainability with the use of
[27], while drastically reducing training time and modify- SLMs, we extend our evaluation on a set of multilingual
ing only a small number of trainable parameters, with re- LMs, of diferent parameter sizes. We select the models
spect to supervised fine-tuning involving all parameters for our evaluation based on state-of-the-art performance
of the original network. In our case, we train efectively and usage in the research community. We benchmark
0.42% of the trainable parameters, corresponding to the each model on the same held-out EN–IT test set, using
LoRA adapter matrices injected in Gemma-3-4B. BLEU, chrF and COMET, and log the CO2eq emissions per
          </p>
          <p>
            Our fine-tuning pipeline was implemented using the inference using the CodeCarbon framework. Importantly,
Hugging Face Transformers library [28], leveraging its we emphasize in our approach that a sustainable model
integration with the PEFT library. For the LoRA config- choice should not be based on its parameter size alone,
uration, we set the rank (r) to 16 and the scaling factor but actual carbon emissions.
(alpha) to 16, with a dropout rate of 0.05 to improve As shown in Table 3, we highlight that the
relationgeneralization. The training was carried out on a single ship between model size and emissions is non-linear.
NVIDIA A100 GPU using mixed-precision (fp16) compu- For instance, Qwen-3B [
            <xref ref-type="bibr" rid="ref21">32</xref>
            ], despite its relatively small
tation. We used the CodeCarbon1 library to monitor the size, exhibits disproportionately high emissions. This can
be attributed to its reasoning behavior during inference,
which results in extended reasoning outputs before gen- calculating a carbon-adjusted score that considers both
erating a final answer. This behavior increases inference translation quality and sustainability.
latency and environmental cost.
          </p>
          <p>Similarly, the assumption that larger models
necessarily produces more carbon emissions does not always hold. 5. Error Analysis
This is the case for models developed with a
Mixture-ofExperts (MoE) architectures. In these models, only a sub- To complement the quantitative results and better
set of the total parameters is activated during inference. understand the practical implications of the
qualityAs a result, MoE models like Mixtral, although large in sustainability trade-of, we conduct a manual error
analaggregate size, can have lower or comparable emissions ysis on the translations generated by four representative
to smaller, densely activated models. This decoupling models: our fine-tuned version of Gemma-3-4B, and the
of parameter size and runtime eficiency highlights the baseline instruction-tuned Gemma-3-27B, Llama-3.2-3B
need for measuring more empirical results, such as CO2eq and Llama-3.3-70B.
emissions.</p>
          <p>Therefore, we introduce a Carbon-Adjusted
Quality Score (CAQS) metric as a measure of model
costefectiveness, and we calculate it on each corpus
translation generated by the models evaluated in our study. Our
CAQS score penalizes each gram of carbon emissions
exponentially, while ensuring that low-quality models
are not rewarded more than high-quality ones, regardless
of their eficiency. We define the CAQS metric as follows.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>6. Conclusions</title>
      <p>larger and environmentally demanding model, Llama-3.3- In this study, we investigated the potential of SLMs as
70B. In terms of weighted scores, both models show simi- sustainable alternatives to LLMs, for MT tasks focusing
lar results, with very few major errors and a comparable on the EN-IT language pair. Our results demonstrate
number of minor ones. The smallest Llama checkpoint that parameter-eficient fine-tuning of SLMs can achieve
presents a very high number of both major and minor competitive translation quality while dramatically
reducerrors, when compared to the Gemma-3-4B model. The ing environmental impact. The fine-tuned Gemma-3-4B
ifndings may suggest that Llama-3’s architecture is sub- model achieved performance comparable to GPT-4o and
optimal for translation tasks across model sizes, given outperformed GPT-4o-mini across all metrics, while
conthat Gemma-3-4B matches the performance of its largest suming approximately 15 times less energy per inference.
checkpoint. However, the results should be interpreted We complement these results with a MQM human
evalwith caution, as our evaluation was limited to a small uation across a set of representative models, confirming
test set and a single language pair. that Gemma-3-4B performed comparably to the much</p>
      <p>In terms of error category distribution increasing pa- larger Llama-3.3-70B, producing only minor fluency and
rameter size leads to an overall performance improve- spelling errors.
ment, as seen in Table 5. This trend is particularly evident We also highlighted that the relationship between
within the Gemma models, where the jump from 4B to model size and carbon emissions is non-linear and highly
27B parameters results in a significant drop in errors dependent on architectural choices, emphasizing the
across all categories. In contrast, Llama-3.2 models ex- need for accurate measurements of carbon emissions.
hibit a less linear improvement, suggesting diminishing Given the non-linear relation between model size
returns from scaling model size. This observation, how- and environmental impact, we introduced the CAQS, a
ever, is limited by the fact that only the smallest Gemma novel metric specifically designed to facilitate sustainable
model was LoRA-adapted, while the LLaMA models were model selection by integrating translation quality and
evaluated in their original form. A more rigorous com- carbon emissions. CAQS includes a sensitivity
paramparison, involving both original and adapted versions eter that allows users to adjust how strongly quality is
across model sizes, is left for future work. penalized by the model’s carbon footprint. According to</p>
      <p>When comparing Gemma-3-4B and Llama-3.3-70B, we this metric, Gemma-3-4B and Magistral-Small emerged
ifnd that most of the errors in the Gemma model are con- as the most eficient models in our study, ofering optimal
centrated in surface-level issues, especially in spelling trade-ofs between sustainability and translation quality.
diacritics. These errors, however, do not compromise</p>
    </sec>
    <sec id="sec-3">
      <title>7. Limitations</title>
    </sec>
    <sec id="sec-4">
      <title>Acknowledgments</title>
      <sec id="sec-4-1">
        <title>This work has been funded by the Italian National PhD</title>
        <p>programme in Artificial Intelligence, partnered by
University of Pisa and University of Naples “L’Orientale”,
through a doctoral grant (ID
39-411-24-DOT23A27WJ6603) established by Ex DM 318, of type 4.1, co-financed
by the National Recovery and Resilience Plan.
In light of practical constraints related to time and
resources, the main limitations of our study lie in the
relatively small sample of segments and the domain-specific
nature of the OpenSubtitles corpus, used for both
training and inference. For this reason, we highlight that
our evaluation results may not be reproducible in other
domains.</p>
        <p>As our evaluation focuses on a relatively high-resource
language pair (EN-IT), our findings may not be
applicable for distant or low-resource pairs. Finally, our carbon
emission measurements are specific to the computational
infrastructure used (NVIDIA A100 GPUs, EU
electricity grid). Results may difer when deploying models on
diferent hardware configurations, cloud providers, or
geographical regions.</p>
        <p>Declaration on Generative AI</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [5]
          <string-name>
            <surname>Google</surname>
            <given-names>DeepMind</given-names>
          </string-name>
          , Gemma:
          <article-title>Open models for responsible ai</article-title>
          , https://deepmind.google/models/ gemma/,
          <year>2024</year>
          . Accessed:
          <fpage>2025</fpage>
          -05-27.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>A.</given-names>
            <surname>Hendy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Abdelrehim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Sharaf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Raunak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Gabr</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Matsushita</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y. J.</given-names>
            <surname>Kim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Afify</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H. H.</given-names>
            <surname>Awadalla</surname>
          </string-name>
          ,
          <article-title>How good are gpt models at machine translation? a comprehensive evaluation, 2023</article-title>
          . URL: https://arxiv.org/abs/2302.09210. arXiv:
          <volume>2302</volume>
          .
          <fpage>09210</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>Z.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Liang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Jiao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Tu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Shi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <article-title>Exploring human-like translation strategy with large language models</article-title>
          ,
          <source>Transactions of the Association for Computational Linguistics</source>
          <volume>12</volume>
          (
          <year>2024</year>
          )
          <fpage>229</fpage>
          -
          <lpage>246</lpage>
          . doi:
          <volume>10</volume>
          .1162/tacl_a_
          <fpage>00642</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Moslem</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Haque</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. D.</given-names>
            <surname>Kelleher</surname>
          </string-name>
          ,
          <string-name>
            <surname>A. Way,</surname>
          </string-name>
          <article-title>Adaptive machine translation with large language models</article-title>
          ,
          <source>arXiv preprint arXiv:2301.13294</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>L.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Lyu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Ji</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Yu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Shi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Tu</surname>
          </string-name>
          ,
          <article-title>Document-level machine translation with large language models</article-title>
          ,
          <source>arXiv preprint arXiv:2304.02210</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>U.S.</given-names>
            <surname>Energy Information</surname>
          </string-name>
          <string-name>
            <surname>Administration</surname>
          </string-name>
          , Electricity use in homes,
          <year>2023</year>
          . URL: https://www.eia.gov/energyexplained/ use
          <article-title>-of-energy/electricity-use-in-homes</article-title>
          .php, accessed:
          <fpage>2025</fpage>
          -06-16.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>S.</given-names>
            <surname>Nguyen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Ding</surname>
          </string-name>
          , S. Liu,
          <article-title>Towards sustainable large language model serving</article-title>
          ,
          <year>2024</year>
          . URL: https: //arxiv.org/abs/2501.
          <year>01990</year>
          . arXiv:
          <fpage>2501</fpage>
          .
          <year>01990</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>C.</given-names>
            <surname>Lyu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Du</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Duan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Wang</surname>
          </string-name>
          , New trends [12]
          <string-name>
            <given-names>F.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Mo</surname>
          </string-name>
          ,
          <string-name>
            <surname>Q. Lu,</surname>
          </string-name>
          <article-title>in machine translation with large language models</article-title>
          , W. Wang,
          <string-name>
            <given-names>R.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Tang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Ma</surname>
          </string-name>
          ,
          <year>2023</year>
          . M.
          <string-name>
            <surname>Huang</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Wang</surname>
          </string-name>
          ,
          <article-title>A comprehensive survey of</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>W.</given-names>
            <surname>Jiao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Wang</surname>
          </string-name>
          , J.-t. Huang,
          <string-name>
            <given-names>X.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Shi</surname>
          </string-name>
          ,
          <string-name>
            <surname>Z.</surname>
          </string-name>
          <article-title>Tu, small language models in the era of large lanIs chatgpt a good translator? yes with gpt-4 as the guage models: Techniques, enhancements</article-title>
          , appliengine,
          <source>arXiv preprint arXiv:2301.08745</source>
          (
          <year>2023</year>
          ).
          <article-title>cations, collaboration with llms, and trustworthi-</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>A.</given-names>
            <surname>Singh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N. P.</given-names>
            <surname>Patel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ehtesham</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Kumar</surname>
          </string-name>
          , T. Ta- ness,
          <year>2024</year>
          . URL: https://arxiv.org/abs/2411.03350.
          <string-name>
            <surname>laei Khoei</surname>
          </string-name>
          ,
          <article-title>A survey of sustainability in large lan-</article-title>
          <source>arXiv:2411.03350. guage models: Applications</source>
          , economics, and chal- [13]
          <string-name>
            <surname>Y.-C. Lin</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Sharma</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Manikandan</surname>
          </string-name>
          , J. Kumar, lenges,
          <source>arXiv preprint arXiv:2412.04782</source>
          (
          <year>2025</year>
          ). T. H.
          <string-name>
            <surname>King</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Zheng</surname>
          </string-name>
          , Eficient multitask learning
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>M. C.</given-names>
            <surname>Rillig</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ågerstrand</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Bi</surname>
          </string-name>
          ,
          <string-name>
            <surname>K.</surname>
          </string-name>
          <article-title>A. Gould, in small language models through upside-down reU. Sauerland, Risks and benefits of large lan- inforcement learning</article-title>
          ,
          <year>2025</year>
          . URL: https://arxiv.org
          <article-title>/ guage models for the environment</article-title>
          ,
          <source>Environmen- abs/2502.09854. arXiv:2502.09854. tal Science &amp; Technology</source>
          <volume>57</volume>
          (
          <year>2023</year>
          )
          <fpage>3464</fpage>
          -
          <lpage>3466</lpage>
          . [14]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Tang</surname>
          </string-name>
          , K. Han,
          <string-name>
            <surname>F</surname>
          </string-name>
          . Liu,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Ni</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Tian</surname>
          </string-name>
          , et al.,
          <source>Rethinkdoi:10.1021/acs.est.3c01106. ing optimization and architecture for tiny language 2106</source>
          .09685, arXiv:
          <fpage>2106</fpage>
          .09685 [cs]. models,
          <source>in: Proceedings of the 41st International</source>
          [24]
          <string-name>
            <given-names>J.</given-names>
            <surname>Zheng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Hong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Su</surname>
          </string-name>
          ,
          <source>Conference on Machine Learning, PMLR</source>
          ,
          <year>2024</year>
          .
          <string-name>
            <given-names>Y.</given-names>
            <surname>Liang</surname>
          </string-name>
          , S. Wu,
          <article-title>Fine-tuning large language</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>G.</given-names>
            <surname>Team</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Kamath</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Ferret</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Pathak</surname>
          </string-name>
          ,
          <string-name>
            <surname>N.</surname>
          </string-name>
          <article-title>Vieil- models for domain-specific machine translalard</article-title>
          , R. Merhej, et al, Gemma 3 technical re- tion,
          <year>2024</year>
          . URL: https://arxiv.org/abs/2402.15061. port,
          <year>2025</year>
          . URL: https://arxiv.org/abs/2503.19786. arXiv:
          <volume>2402</volume>
          .15061. arXiv:
          <volume>2503</volume>
          .
          <fpage>19786</fpage>
          . [25]
          <string-name>
            <surname>D. M. Alves</surname>
            ,
            <given-names>N. M.</given-names>
          </string-name>
          <string-name>
            <surname>Guerreiro</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Alves</surname>
          </string-name>
          , J. Pombal,
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Fu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Peng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Ou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Sabharwal</surname>
          </string-name>
          , T. Khot,
          <string-name>
            <surname>Special- R. Rei</surname>
          </string-name>
          ,
          <string-name>
            <surname>J. G. C. de Souza</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Colombo</surname>
            ,
            <given-names>A. F. T.</given-names>
          </string-name>
          <article-title>Marizing smaller language models towards multi-step tins, Steering large language models for machine reasoning</article-title>
          ,
          <source>in: Proceedings of the 40th International translation with finetuning and in-context learnConference on Machine Learning, PMLR</source>
          ,
          <year>2023</year>
          . ing,
          <year>2023</year>
          . URL: https://arxiv.org/abs/2310.13448.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>C. V.</given-names>
            <surname>Nguyen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Shen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Aponte</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Xia</surname>
          </string-name>
          , arXiv:
          <fpage>2310</fpage>
          .
          <fpage>13448</fpage>
          . et al.,
          <source>A survey of small language models</source>
          ,
          <year>2024</year>
          . [26]
          <string-name>
            <given-names>X.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Rajabi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Duh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Koehn</surname>
          </string-name>
          , Machine arXiv:
          <volume>2410</volume>
          .20011.
          <article-title>translation with large language models: Prompting,</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [18]
          <string-name>
            <surname>OpenAI</surname>
          </string-name>
          , Gpt-4o, https://openai.com/gpt-4o,
          <year>2024</year>
          .
          <article-title>few-shot learning, and fine-tuning with QLoRA</article-title>
          , Accessed:
          <fpage>2025</fpage>
          -07-
          <lpage>23</lpage>
          . in: P.
          <string-name>
            <surname>Koehn</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Haddow</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Kocmi</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          Monz (Eds.),
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>J.</given-names>
            <surname>Tiedemann</surname>
          </string-name>
          , S. Thottingal,
          <article-title>OPUS-MT - Building Proceedings of the Eighth Conference on Machine open translation services for the World</article-title>
          , in: A. Mar- Translation, Association for Computational Lintins,
          <string-name>
            <given-names>H.</given-names>
            <surname>Moniz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Fumega</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Martins</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Batista</surname>
          </string-name>
          , guistics, Singapore,
          <year>2023</year>
          , pp.
          <fpage>468</fpage>
          -
          <lpage>481</lpage>
          . URL: https: L.
          <string-name>
            <surname>Coheur</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Parra</surname>
          </string-name>
          , I. Trancoso, M. Turchi, //aclanthology.org/
          <year>2023</year>
          .wmt-
          <volume>1</volume>
          .43/. doi:
          <volume>10</volume>
          .18653/ A. Bisazza,
          <string-name>
            <given-names>J.</given-names>
            <surname>Moorkens</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Guerberof</surname>
          </string-name>
          , M. Nurmi- v1/
          <year>2023</year>
          .wmt-
          <volume>1</volume>
          .43.
          <string-name>
            <surname>nen</surname>
            , L. Marg,
            <given-names>M. L.</given-names>
          </string-name>
          Forcada (Eds.), Proceedings of [27]
          <string-name>
            <given-names>K.</given-names>
            <surname>Papineni</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Roukos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Ward</surname>
          </string-name>
          , W.-J. Zhu,
          <article-title>Bleu: the 22nd Annual Conference of the European As- a method for automatic evaluation of machine sociation for Machine Translation</article-title>
          , European Asso- translation, in: P.
          <string-name>
            <surname>Isabelle</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          <string-name>
            <surname>Charniak</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <article-title>Lin ciation for Machine Translation</article-title>
          , Lisboa, Portugal, (Eds.),
          <source>Proceedings of the 40th Annual Meeting of 2020</source>
          , pp.
          <fpage>479</fpage>
          -
          <lpage>480</lpage>
          . URL: https://aclanthology.org/ the Association for Computational Linguistics,
          <source>As2020.eamt-1</source>
          .61.
          <article-title>sociation for Computational Linguistics</article-title>
          , Philadel-
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>S.</given-names>
            <surname>Nakatani</surname>
          </string-name>
          ,
          <article-title>Langdetect: Language detection library phia</article-title>
          , Pennsylvania, USA,
          <year>2002</year>
          , pp.
          <fpage>311</fpage>
          -
          <lpage>318</lpage>
          . URL: for python, https://pypi.org/project/langdetect/, https://aclanthology.org/P02-1040/. doi:
          <volume>10</volume>
          .3115/
          <year>2014</year>
          .
          <article-title>Port of Google's language-detection library</article-title>
          .
          <volume>1073083</volume>
          .1073135.
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>E. A.</given-names>
            <surname>Chimoto</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B. A.</given-names>
            <surname>Bassett</surname>
          </string-name>
          , Comet-qe and ac- //aclanthology.org/W15-3049/. doi:
          <volume>10</volume>
          .18653/v1/
          <article-title>tive learning for low-resource machine transla-</article-title>
          <source>W15-3049. tion</source>
          ,
          <year>2022</year>
          . URL: https://arxiv.org/abs/2210.15696. [30]
          <string-name>
            <given-names>R.</given-names>
            <surname>Rei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Stewart</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. C.</given-names>
            <surname>Farinha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Lavie</surname>
          </string-name>
          , COMET: arXiv:
          <fpage>2210</fpage>
          .15696.
          <article-title>A neural framework for MT evaluation</article-title>
          , in: B. Web-
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>E. J.</given-names>
            <surname>Hu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Shen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Wallis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Allen-Zhu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Li</surname>
          </string-name>
          , ber, T. Cohn,
          <string-name>
            <given-names>Y.</given-names>
            <surname>He</surname>
          </string-name>
          , Y. Liu (Eds.),
          <string-name>
            <surname>Proceedings</surname>
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Wang</surname>
          </string-name>
          , W. Chen, LoRA:
          <string-name>
            <surname>Low-Rank</surname>
          </string-name>
          Adap- of
          <source>the 2020 Conference on Empirical Methods tation of Large Language Models</source>
          ,
          <year>2021</year>
          . URL: http: in
          <source>Natural Language Processing (EMNLP)</source>
          , As//arxiv.org/abs/2106.09685. doi:
          <volume>10</volume>
          .48550/arXiv. sociation for Computational Linguistics, Online,
          <year>2020</year>
          , pp.
          <fpage>2685</fpage>
          -
          <lpage>2702</lpage>
          . URL: https://aclanthology.org/
          <year>2020</year>
          .emnlp-main.
          <volume>213</volume>
          /. doi:
          <volume>10</volume>
          .18653/v1/
          <year>2020</year>
          . emnlp-main.
          <volume>213</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [31]
          <string-name>
            <given-names>N.</given-names>
            <surname>Jegham</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Abdelatti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Elmoubarki</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Hendawi</surname>
          </string-name>
          ,
          <article-title>How hungry is ai? benchmarking energy, water, and carbon footprint of llm inference</article-title>
          ,
          <year>2025</year>
          . URL: https://arxiv.org/abs/2505.09598. arXiv:
          <volume>2505</volume>
          .
          <fpage>09598</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [32]
          <string-name>
            <given-names>A.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Hui</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Zheng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Yu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Gao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Lv</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Zheng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Hu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Ge</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Wei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Tang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Tu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Dang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Bao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Yu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Deng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Xue</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Zhu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Men</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Gao</surname>
          </string-name>
          , S. Liu,
          <string-name>
            <given-names>S.</given-names>
            <surname>Luo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Tang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Yin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Ren</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Ren</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Fan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Su</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Cui</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Qiu</surname>
          </string-name>
          ,
          <source>Qwen3 technical report</source>
          ,
          <year>2025</year>
          . URL: https://arxiv. org/abs/2505.09388. arXiv:
          <volume>2505</volume>
          .
          <fpage>09388</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [33]
          <string-name>
            <surname>Mistral-AI</surname>
            , :,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Rastogi</surname>
            ,
            <given-names>A. Q.</given-names>
          </string-name>
          <string-name>
            <surname>Jiang</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Lo</surname>
            , G. Berrada, G. Lample,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Rute</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Barmentlo</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Yadav</surname>
          </string-name>
          , e. a. Kartik Khandelwal, Magistral,
          <year>2025</year>
          . URL: https://arxiv.org/abs/2506.10910. arXiv:
          <volume>2506</volume>
          .
          <fpage>10910</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [34]
          <string-name>
            <given-names>A.</given-names>
            <surname>Grattafiori</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Dubey</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Jauhri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Pandey</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Kadian</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Al-Dahle</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Letman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mathur</surname>
          </string-name>
          ,
          <string-name>
            <surname>A</surname>
          </string-name>
          . Schelten, e. a. Alex Vaughan,
          <source>The llama 3 herd of models</source>
          ,
          <year>2024</year>
          . URL: https://arxiv.org/abs/2407.21783. arXiv:
          <volume>2407</volume>
          .
          <fpage>21783</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [35]
          <string-name>
            <given-names>M.</given-names>
            <surname>Abdin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Aneja</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Awadalla</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Awadallah</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. A.</given-names>
            <surname>Awan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Bach</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Bahree</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Bakhtiari</surname>
          </string-name>
          , J. Bao, e. a. Harkirat Behl, Phi-3
          <source>technical report: A highly capable language model locally on your phone</source>
          ,
          <year>2024</year>
          . URL: https://arxiv.org/abs/2404.14219. arXiv:
          <volume>2404</volume>
          .
          <fpage>14219</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [36]
          <string-name>
            <given-names>A.</given-names>
            <surname>Lommel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Gladkof</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Melby</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. E.</given-names>
            <surname>Wright</surname>
          </string-name>
          , I. Strandvik,
          <string-name>
            <given-names>K.</given-names>
            <surname>Gasova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Vaasa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Benzo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. M.</given-names>
            <surname>Sparano</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Foresi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Innis</surname>
          </string-name>
          , L. Han,
          <string-name>
            <surname>G</surname>
          </string-name>
          . Nenadic,
          <article-title>The multi-range theory of translation quality measurement: Mqm scoring models and statistical quality control</article-title>
          ,
          <year>2024</year>
          . URL: https://arxiv.org/abs/2405.16969. arXiv:
          <volume>2405</volume>
          .
          <fpage>16969</fpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>