<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>mdok of KInIT: Robustly Fine-tuned LLM for Binary and Multiclass AI-Generated Text Detection</article-title>
      </title-group>
      <contrib-group>
        <aff id="aff0">
          <label>0</label>
          <institution>Kempelen Institute of Intelligent Technologies</institution>
          ,
          <addr-line>Bratislava</addr-line>
          ,
          <country country="SK">Slovakia</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <abstract>
        <p>The large language models (LLMs) are able to generate high-quality texts in multiple languages. Such texts are often not recognizable by humans as generated, and therefore present a potential of LLMs for misuse (e.g., plagiarism, spams, disinformation spreading). An automated detection is able to assist humans to indicate the machine-generated texts; however, its robustness to out-of-distribution data is still challenging. This notebook describes our mdok approach in robust detection, based on fine-tuning smaller LLMs for text classification. It is applied to both subtasks of Voight-Kampf Generative AI Detection 2025, providing remarkable performance (1st rank) in both, the binary detection as well as the multiclass classification of various cases of human-AI collaboration.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;PAN 2025</kwd>
        <kwd>Voight-Kampf Generative AI Detection 2025</kwd>
        <kwd>Large language models</kwd>
        <kwd>Machine-generated text detection</kwd>
        <kwd>AI-content detection</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Continuously increasing quality of texts generated by artificial intelligence (AI) technology, such as large
language models (LLMs), causes that humans are no longer able to diferentiate between human-written
and high-quality machine-generated texts. Naturally, this arises concerns about potential LLM misuse,
e.g., for accelerated generation of disinformation [
        <xref ref-type="bibr" rid="ref1 ref2">1, 2</xref>
        ], plagiarism [3], or frauds for academic exams [4].
The automated means, also utilizing AI technology, are able to help humans to diferentiate such texts;
however, the automated detection performance is also not perfect (errors occur, such as false positives
or false negatives). Further challenges are in application of the trained detectors to out-of-distribution
data, i.e. data that are significantly diferent to data used for training (surprise data).
      </p>
      <p>The Voight-Kampf Generative AI Detection 2025 shared task [5], as a part of PAN lab [6] at the CLEF
2025 conference, addresses two challenges in the machine-generated text detection problem area in
two subtasks. Subtask 1 is focused on classical binary detection task distinguishing between
machineauthored (class 1) and human-authored (class 0) texts, which addresses also the robustness of detectors
to obfuscation and other surprise data. Subtask 2 is focused on human-AI collaborative text classification,
covering six classes of collaboration, namely class 0 representing fully human-written texts; class 1
representing human-written, then machine-polished texts, class 2 representing machine-written, then
machine-humanized texts, class 3 representing human-initiated, then machine-continued texts, class 4
representing deeply-mixed text (where some parts are written by a human and some are generated by a
machine), and class 5 representing machine-written, then human-edited texts.</p>
      <p>In this notebook, we describe our mdok approach addressing both subtasks (two separate systems).
We build on our previously developed robust LLM fine-tuning for sequence classification [ 7], but adjust
the training process to the shared task training data and binary/multiclass problem formulation. We
further explore the usage of the most recent LLMs, size of which ranging from 1B to 14B parameters.
The submitted systems are based on Qwen3-4B and Qwen3-14B models [8]. For the replication purpose,
we publish the source code of the mdok approach1.</p>
      <p>The contributions of the proposed approach are as follows:
• We have proposed the usage of most modern Qwen3 LLM in robust mdok fine-tuning
pipeline and evaluated its robustness in sensitivity to obfuscation.
• We have proposed a modification of robust fine-tuning mdok approach for multiclass
detection of AI-human collaboration.
• We have benchmarked multiple most modern LLMs on the provided evaluation datasets,
making comparison of other models and approaches straightforward and fair.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Background</title>
      <p>A binary machine-generated text detection is a well researched task, typically addressed by stylometric
methods (e.g., a machine learning classifier trained on TF-IDF features), statistical methods (e.g., utilizing
perplexity, entropy, or likelihood) [9, 10], or fine-tuned language models for classification task (e.g., by
supervised or contrastive learning) [11, 12]. Most of the detection methods can be directly applied by
existing frameworks, such as MGTBench [13] or IMGTB [14].</p>
      <p>A multiclass machine-generated text classification is mostly researched in related authorship
attribution task, identifying the author (generator) of the text. The problem of diferent classes of AI-human
collaboration can be approached by existing authorship attribution methods [15, 16], only slightly
diferent from binary (two-class) methods.</p>
      <p>The robustness of machine-generated text detection methods against authorship obfuscation methods
has been explored in [17]. Although it has been focused on multilingual settings and the results
difer among languages, it has shown that fine-tuned detection methods are more robust against
obfuscation than the statistical methods, while ofering significantly higher detection performance.
Further it has shown that including obfuscated data into fine-tuning process increases the detector’s
robustness against obfuscation. Such approach has been proposed in [7] and shown to increase even
the generalization to out-of-distribution data. The training data mixture consists of social-media texts
from the MultiSocial [18] dataset, news articles from the MULTITuDE [19] dataset, and obfuscated
texts from [17]. For validation (i.e., model-checkpoint selection), the approach uses a unique massively
multi-generator (75 generators) multilingual (7 languages) data of MIX2k composition of 18 existing
labeled datasets to represent out-of-distribution data.</p>
    </sec>
    <sec id="sec-3">
      <title>3. System Overview</title>
      <p>For this shared task, we aimed to keep the system as simple as possible, ideally resulting in a
singlemodel detection system for each task (avoiding ensembles), just by tweaking the data and the training
process.</p>
      <p>Although Subtask 1 does not explicitly mention any limitation in usage of additional training data,
the Subtask 2 conditions are clear in this regard, not allowing additional training data. Therefore, for
adoption of robust fine-tuning approach of [ 7], we do not include additional training data. Instead, for
Subtask 1, we combine train and validation sets into training data and modify a small portion of them by
using a generic homoglyph attack of [17]. For validation (model selection), we use MIX2k dataset only
to ensure the best generalization to out-of-distribution data. For Subtask 2, we also combine train and
validation sets for training, but we leave a pseudo-randomly balanced (500 samples per class) holdout
portion for validation (in order to minimize bias due to imbalanced evaluation). An overview of this
approach is illustrated in Figure 1.</p>
      <p>For the fine-tuning process, we have used the adjusted published script for QLoRA based robust
ifne-tuning of [ 7]. Besides of modification of training data mixture, we have used learning rate of 2 − 4,
avoided gradient accumulation, and limited the training time to 3 epochs. Due to already incorporated
weighted cross entropy for loss calculations, it was straightforward to extend it to multiclass classification
by calculating weights based on the training data class distribution.</p>
      <p>Train + validation splits
for training</p>
      <p>MIX2k
for validation
Train + validation splits</p>
      <p>for training
Balanced hold-out</p>
      <p>Qwen3-14B</p>
      <p>Base
mdok binary</p>
      <p>detector
Qwen3-4B-Base
mdok multiclass
detector</p>
      <p>Since the evaluation on the original validation sets is not usable (due to data leakage) when combined
in the training, we have used the original training sets for training to select suitable LLMs to robustly
ifne-tune for the two subtasks and compared them on validation sets (as well as on MIX2k data). Based
on such comparison, we have decided to use the Qwen3-14B model for binary detection in Subtask 1
and the Qwen3-4B for multiclass classification in Subtask 2.</p>
      <sec id="sec-3-1">
        <title>3.1. Homoglyph Attack</title>
        <p>As mentioned, we have used a generic homoglyph attack of [17] to modify (in the pre-processing step) a
small portion of training data (pseudo-randomly selected 10% of machine-generated texts) for Subtask 1
to increase its robustness to obfuscation. It uses the whole confusables table2 to psudorandomly replace
letters for their homoglyphs. The probability of a character to be replaced was set to 0.05 (i.e., about 5%
of characters). It further integrates a pseudorandom insertion of visually unseenable zero-width-joiner
character in the text (also with a probability of 0.05). We have used the random seed of 42. It has been
previously shown that seeing such obfuscated data in the training (fine-tuning) efectively eliminates
their highly negative efect on the detection.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. MIX2k Dataset</title>
        <p>For validation in out-of-distribution settings in Subtask 1, we have used MIX2k dataset, introduced
by [7]. It contains 1,000 samples for each class (human and machine), pseudo-randomly sampled from
18 existing labeled datasets. Such data composition consists of data from 75 generators in 7 languages,
thus being mostly out-of-distribution (in comparison to training data). Validation on such a dataset
during fine-tuning procedure helps to select the model checkpoint, which has the best generalization to
out-of-distribution data, efectively avoiding overfit to training data.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Results</title>
      <p>The results are provided by using standard/oficial evaluation metrics specified for the shared task.
Namely, in Subtask 1, the following metrics are used:
• ROC-AUC: The area under the ROC (Receiver Operating Characteristic) curve.
• Brier: The complement of the Brier score (mean squared loss).
• C@1: A modified accuracy score that assigns non-answers (score = 0.5) the average accuracy of
the remaining cases.</p>
      <sec id="sec-4-1">
        <title>2https://www.unicode.org/Public/security/8.0.0/confusables.txt</title>
        <p>0.984
0.757
0.844
0.6230
0.5920
0.5710
0.5720
0.6480
0.6995</p>
        <p>F1
• F0.5u: A modified F0.5 measure (precision-weighted F measure) that treats non-answers (score =
0.5) as false negatives.</p>
        <p>• The arithmetic mean of all the metrics above.</p>
        <p>For Subtask 2, standard metrics utilize macro averages of Recall, Precision, and F1 scores, as well as
overall Accuracy.</p>
        <p>In Table 1, the detection performance for oficial validation set of Subtask 1 for each tested detector
is provided. The results indicate that all the fine-tuned models perform almost ideally (very close to 1
values, outperforming all the baselines), which also indicates a potential overfit to in-distribution data.</p>
        <p>Therefore, we have also tested the detectors on the MIX2k out-of-distribution dataset (see Table 2).
As the results show, the Qwen3-14B-Base fine-tuned detector generalizes the best to out-of-distribution
data. Therefore, it was our natural selection for robust fine-tuning using the mdok approach
(described in Section 3), resulting in the mdok (binary) detector, which further boosted the performance,
outperforming the baseline detectors.</p>
        <p>For Subtask 2, the results for the oficial validation set are provided in Table 3. Although the results
indicate that the Qwen2.5-1.5B models performs the best in oficial metric of Macro Recall, we selected
newer and bigger Qwen3-4B-Base (with similar performance) for robust fine-tuning. Using the mdok
approach for the fine-tuning resulted in mdok (multiclass) detector, hopefully outperforming the other
variants. However, since it uses combined training and validation sets for training, the results indicated
in Table 3 for this detector are not representative.</p>
        <p>F1
0.898
0.865
0.860
0.856
0.854
0.845
0.855
0.851
0.847
0.854
0.827</p>
        <p>F0.5u</p>
        <sec id="sec-4-1-1">
          <title>4.1. Subtask 1 Oficial Results</title>
          <p>The oficial results are provided in Table 4, where Mean of the oficial metrics has been used for ranking.
The results show that our approach outperforms all the others in all metrics except of ROC-AUC, where
the system of one team outperformed our detector. The performances of the systems are close enough
to yield a more eficient detection system, since we have relied on a rather large (14B of parameters)
model, consuming more resources for inference. However, the mdok single-model detection system
can still be more eficient than some more complex ensemble systems, relying on multiple inference
rounds of multiple (although smaller) models. If the labeled test set will be released, an ablation study
could pinpoint the strong parts of our system that make it the best performing and/or identify a smaller
model to be used instead to provide a more eficient solution.</p>
        </sec>
        <sec id="sec-4-1-2">
          <title>4.2. Subtask 2 Oficial Results</title>
          <p>The oficial results are provided in Table 5, where Macro Recall is the oficial ranking metric. The results
show that our approach outperforms all the others by a margin of 3%. Our decision to keep the system
simple and focus on the training data and fine-tuning process paid of. If the labeled test set will be
released, the further analysis and ablation study reveals whether we have made the right selection of
the system to submit. However, even the top performance of our system (Macro Recall of 64.46%) is far
from perfect (Macro Recall of 100%). This provides further space for improvements.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion</title>
      <p>The robustness of machine-generated text detection methods to out-of-distribution data is still
challenging. However, we can cope with this problem by better mixture of training data, as we have show
in the proposed mdok approach for both binary and multiclass detection. The resulting detectors are
very competitive (ranking 1st in both subtasks), outperforming all the baselines, and promising better
generalization to out-of-distribution data. Due to gold labels of the test data are not available yet,
any ablation study pinpointing crucial parts of the systems and their specific efects was not possible.
Further work is still needed to refine the process to achieve ideal performance.</p>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgments</title>
      <p>This work was partially supported by the European Union NextGenerationEU through the Recovery
and Resilience Plan for Slovakia under the project No. 09I01-03-V04-00059 and partially by LorAI –
Low Resource Artificial Intelligence , a project funded by Horizon Europe under GA No.101136646. We
acknowledge EuroHPC Joint Undertaking for awarding us access to Leonardo at CINECA, Italy.</p>
    </sec>
    <sec id="sec-7">
      <title>Declaration on Generative AI</title>
      <sec id="sec-7-1">
        <title>The author(s) have not employed any Generative AI tools.</title>
        <p>[3] J. P. Wahle, T. Ruas, F. Kirstein, B. Gipp, How large language models are transforming
machineparaphrase plagiarism, in: Y. Goldberg, Z. Kozareva, Y. Zhang (Eds.), Proceedings of the 2022
Conference on Empirical Methods in Natural Language Processing, Association for Computational
Linguistics, Abu Dhabi, United Arab Emirates, 2022, pp. 952–963. URL: https://aclanthology.org/
2022.emnlp-main.62/. doi:10.18653/v1/2022.emnlp-main.62.
[4] J. Achiam, S. Adler, S. Agarwal, L. Ahmad, I. Akkaya, F. L. Aleman, D. Almeida, J. Altenschmidt,</p>
        <p>S. Altman, S. Anadkat, et al., GPT-4 technical report, arXiv preprint arXiv:2303.08774 (2023).
[5] J. Bevendorf, Y. Wang, J. Karlgren, M. Wiegmann, M. Fröbe, A. Tsivgun, J. Su, Z. Xie, M. Abassy,
J. Mansurov, R. Xing, M. N. Ta, K. A. Elozeiri, T. Gu, R. V. Tomar, J. Geng, E. Artemova, A. Shelmanov,
N. Habash, E. Stamatatos, I. Gurevych, P. Nakov, M. Potthast, B. Stein, Overview of the
“VoightKampf” Generative AI Authorship Verification Task at PAN and ELOQUENT 2025, in: G. Faggioli,
N. Ferro, P. Rosso, D. Spina (Eds.), Working Notes of CLEF 2025 – Conference and Labs of the
Evaluation Forum, CEUR Workshop Proceedings, CEUR-WS.org, 2025.
[6] J. Bevendorf, D. Dementieva, M. Fröbe, B. Gipp, A. Greiner-Petter, J. Karlgren, M. Mayerl, P. Nakov,
A. Panchenko, M. Potthast, A. Shelmanov, E. Stamatatos, B. Stein, Y. Wang, M. Wiegmann,
E. Zangerle, Overview of PAN 2025: Voight-Kampf Generative AI Detection, Multilingual Text
Detoxification, Multi-Author Writing Style Analysis, and Generative Plagiarism Detection, in:
J. C. de Albornoz, J. Gonzalo, L. Plaza, A. G. S. de Herrera, J. Mothe, F. Piroi, P. Rosso, D. Spina,
G. Faggioli, N. Ferro (Eds.), Experimental IR Meets Multilinguality, Multimodality, and Interaction.
Proceedings of the Sixteenth International Conference of the CLEF Association (CLEF 2025),
Lecture Notes in Computer Science, Springer, Berlin Heidelberg New York, 2025.
[7] D. Macko, R. Moro, I. Srba, Increasing the robustness of the fine-tuned multilingual
machinegenerated text detectors, 2025. URL: https://arxiv.org/abs/2503.15128. arXiv:2503.15128.
[8] Q. Team, Qwen3 technical report, 2025. URL: https://arxiv.org/abs/2505.09388.</p>
        <p>arXiv:2505.09388.
[9] A. Hans, A. Schwarzschild, V. Cherepanova, H. Kazemi, A. Saha, M. Goldblum, J. Geiping, T.
Goldstein, Spotting llms with binoculars: Zero-shot detection of machine-generated text, 2024. URL:
https://arxiv.org/abs/2401.12070. arXiv:2401.12070.
[10] G. Bao, Y. Zhao, Z. Teng, L. Yang, Y. Zhang, Fast-DetectGPT: Eficient zero-shot detection of
machine-generated text via conditional probability curvature, in: The Twelfth International
Conference on Learning Representations, 2023.
[11] M. Spiegel, D. Macko, KInIT at SemEval-2024 task 8: Fine-tuned LLMs for multilingual
machinegenerated text detection, in: A. K. Ojha, A. S. Doğruöz, H. Tayyar Madabushi, G. Da San
Martino, S. Rosenthal, A. Rosá (Eds.), Proceedings of the 18th International Workshop on Semantic
Evaluation (SemEval-2024), Association for Computational Linguistics, Mexico City, Mexico,
2024, pp. 558–564. URL: https://aclanthology.org/2024.semeval-1.84/. doi:10.18653/v1/2024.
semeval-1.84.
[12] S. R. Dipta, S. Shahriar, Hu at semeval-2024 task 8a: Can contrastive learning learn embeddings to
detect machine-generated text?, 2024. URL: https://arxiv.org/abs/2402.11815. arXiv:2402.11815.
[13] X. He, X. Shen, Z. Chen, M. Backes, Y. Zhang, Mgtbench: Benchmarking machine-generated text
detection, in: Proceedings of the 2024 on ACM SIGSAC Conference on Computer and
Communications Security, CCS ’24, Association for Computing Machinery, New York, NY, USA, 2024, p.
2251–2265. URL: https://doi.org/10.1145/3658644.3670344. doi:10.1145/3658644.3670344.
[14] M. Spiegel, D. Macko, IMGTB: A framework for machine-generated text detection benchmarking,
in: Y. Cao, Y. Feng, D. Xiong (Eds.), Proceedings of the 62nd Annual Meeting of the Association for
Computational Linguistics (Volume 3: System Demonstrations), Association for Computational
Linguistics, Bangkok, Thailand, 2024, pp. 172–179. URL: https://aclanthology.org/2024.acl-demos.
17/. doi:10.18653/v1/2024.acl-demos.17.
[15] A. Uchendu, Z. Ma, T. Le, R. Zhang, D. Lee, TURINGBENCH: A benchmark environment for
Turing test in the age of neural text generation, in: M.-F. Moens, X. Huang, L. Specia, S. W.-t. Yih
(Eds.), Findings of the Association for Computational Linguistics: EMNLP 2021, Association for
Computational Linguistics, Punta Cana, Dominican Republic, 2021, pp. 2001–2016. URL: https:
//aclanthology.org/2021.findings-emnlp.172/. doi: 10.18653/v1/2021.findings-emnlp.172.
[16] L. La Cava, D. Costa, A. Tagarelli, Is contrasting all you need? contrastive learning for the detection
and attribution of ai-generated text, in: ECAI 2024, IOS Press, 2024, pp. 3179–3186.
[17] D. Macko, R. Moro, A. Uchendu, I. Srba, J. S. Lucas, M. Yamashita, N. I. Tripto, D. Lee, J. Simko,
M. Bielikova, Authorship obfuscation in multilingual machine-generated text detection, in:
Y. Al-Onaizan, M. Bansal, Y.-N. Chen (Eds.), Findings of the Association for Computational
Linguistics: EMNLP 2024, Association for Computational Linguistics, Miami, Florida, USA, 2024,
pp. 6348–6368. URL: https://aclanthology.org/2024.findings-emnlp.369/. doi: 10.18653/v1/2024.
findings-emnlp.369.
[18] D. Macko, J. Kopal, R. Moro, I. Srba, MultiSocial: Multilingual benchmark of machine-generated text
detection of social-media texts, 2024. URL: https://arxiv.org/abs/2406.12549. arXiv:2406.12549.
[19] D. Macko, R. Moro, A. Uchendu, J. Lucas, M. Yamashita, M. Pikuliak, I. Srba, T. Le, D. Lee, J. Simko,
M. Bielikova, MULTITuDE: Large-scale multilingual machine-generated text detection benchmark,
in: H. Bouamor, J. Pino, K. Bali (Eds.), Proceedings of the 2023 Conference on Empirical Methods
in Natural Language Processing, Association for Computational Linguistics, Singapore, 2023,
pp. 9960–9987. URL: https://aclanthology.org/2023.emnlp-main.616/. doi:10.18653/v1/2023.
emnlp-main.616.</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>I.</given-names>
            <surname>Vykopal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Pikuliak</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Srba</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Moro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Macko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Bielikova</surname>
          </string-name>
          ,
          <article-title>Disinformation capabilities of large language models</article-title>
          , in: L.
          <string-name>
            <surname>-W. Ku</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Martins</surname>
          </string-name>
          , V. Srikumar (Eds.),
          <source>Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume</source>
          <volume>1</volume>
          :
          <string-name>
            <surname>Long</surname>
            <given-names>Papers)</given-names>
          </string-name>
          ,
          <source>Association for Computational Linguistics</source>
          , Bangkok, Thailand,
          <year>2024</year>
          , pp.
          <fpage>14830</fpage>
          -
          <lpage>14847</lpage>
          . URL: https: //aclanthology.org/
          <year>2024</year>
          .
          <article-title>acl-long</article-title>
          .
          <volume>793</volume>
          /. doi:
          <volume>10</volume>
          .18653/v1/
          <year>2024</year>
          .
          <article-title>acl-long</article-title>
          .
          <volume>793</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>A.</given-names>
            <surname>Zugecova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Macko</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Srba</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Moro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Kopal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Marcincinova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Mesarcik</surname>
          </string-name>
          ,
          <article-title>Evaluation of llm vulnerabilities to being misused for personalized disinformation generation</article-title>
          ,
          <year>2024</year>
          . URL: https://arxiv.org/abs/2412.13666. arXiv:
          <volume>2412</volume>
          .
          <fpage>13666</fpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>