<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>ORCID:</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Writing from Computer Generated Writing</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Luis Enrique Morales-Márquez</string-name>
          <email>luise.morales@viep.com.mx</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Erick Barrios-González</string-name>
          <email>erick.barrios@viep.com.mx</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>David Eduardo Pinto-Avendaño</string-name>
          <email>david.pinto@correo.buap.mx</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Autonomous University of Puebla</institution>
          ,
          <addr-line>San Claudio Av., 14 Sur Blvd., Puebla, ZIP Code: 72592</addr-line>
          ,
          <country country="MX">México</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Computer Generated Texts</institution>
          ,
          <addr-line>Texts Classification, Machine Learning</addr-line>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Proceedings IberLEF</institution>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2023</year>
      </pub-date>
      <volume>000</volume>
      <fpage>0</fpage>
      <lpage>0001</lpage>
      <abstract>
        <p>In recent years, linguistic computational models have advanced to the point where they can generate stories and hold conversations. Despite being a useful tool, computer generated texts can be used for unethical purposes, which underscores the need to identify these texts. This paper presents a model that classifies human- or computer-generated texts, using vocabulary richness metrics and POS label ratios to train a simple artificial neural network for spanish classification, and some others features to build a Naïve Bayes Model for english classification. The objective is to classify texts in both English and Spanish. The results show a Macro F1 of 0.67 for the texts in English and 0.6441 for the texts in Spanish. Therefore, the performance of this model is consistent with that achieved in previous research.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>2023 Copyright for this paper by its authors.
the proposed method and the associated theoretical concepts; Section 4 presents the results obtained
and their analysis; and finally, Section 5 contains the conclusions.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <sec id="sec-2-1">
        <title>Below are some recent proposals for the classification of texts generated by humans or by some</title>
        <p>TGM.</p>
        <p>Gehrmann, Strobelt &amp; Rush (2019) developed Giant language Model Test Room (GLTR), a
computer-generated text detector based on text statistics. The central idea is that the generated texts are
written based on a very limited set of language distributions. The tests use the probability of occurrence
of a word given that other words are already written, the rank of a word, and the entropy of the
distribution model predicted by the text generator. By calculating the rank and entropy of each text,
they showed that these metrics are higher in human texts, thus obtaining a 72% hit rate in text detection
[3].</p>
      </sec>
      <sec id="sec-2-2">
        <title>Ippolito et al. (2020) collected 250,000 texts generated by the GPT-2 model and an additional 5,000</title>
        <p>for testing. They proposed a refined variant of the BERT linguistic model and obtained approximately
70% accuracy with prior knowledge of how GPT-2 generates texts. By using decoding strategies, which
consist of selecting the next word to write based on probability distributions, they determined that
knowing the generation process has a large impact on the performance of the classifier [4].</p>
      </sec>
      <sec id="sec-2-3">
        <title>Kirchenbauer et al. (2023), instead of proposing an AI-generated text detection model, suggest the</title>
        <p>idea of watermarking TGMs without retraining them. The water-mark consists of the use of certain
words that must appear when generating a text, appearance determined by existing probability models.</p>
      </sec>
      <sec id="sec-2-4">
        <title>The selected words are those that present a greater entropy in the texts generated by the model, and</title>
        <p>must have a much higher frequency of use to allow the detection of the watermark. Although this idea
is subject to the weaknesses of the watermarking mechanisms of any other multimedia element, it is a
potential idea that, applied correctly, may allow detection in the future [5].</p>
      </sec>
      <sec id="sec-2-5">
        <title>Some researchers find the task so difficult that they prefer to concentrate on detecting texts written</title>
        <p>by the same model. This is the case of Gritsay, Grabovoy &amp; Chekhovic (2022), who developed a model
based on ROBERTA for the similarity detection of texts generated by the same TGM GPT-2. The
results show that a large number of tokens and very wide windows are required to improve the results,
although this may result in overfitting if it is decided to analyze very wide neighborhoods for the text
tokens. They report accuracy of up to 97% in detection [6].</p>
      </sec>
      <sec id="sec-2-6">
        <title>No related works have been found that have the objective of detecting computer-generated texts in</title>
      </sec>
      <sec id="sec-2-7">
        <title>Spanish, so it will be one of the tasks addressed in this work.</title>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Proposed Method</title>
      <sec id="sec-3-1">
        <title>The used corpus is the corpus provided in AuTexTification subtask 1 from IberLEF 2023 [9], this dataset contains 33845 texts in english for training and 21832 for tests and contains 32062 texts in spanish for training with 20129 for test. Two specialized models were implemented for english and spanish, the methods will be described below.</title>
        <p>3.1.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Subtask 1 - English</title>
      <sec id="sec-4-1">
        <title>The implementation oriented to the English language, is based on a naive bayes model. The Naive</title>
      </sec>
      <sec id="sec-4-2">
        <title>Bayes algorithm is frequently used in binary classification, it is used as a base model to compare the results of new models [7, 8].</title>
      </sec>
      <sec id="sec-4-3">
        <title>To implement this model, a pre-processing is necessary. This process consists of using the library</title>
        <p>provided by Spacy to tag the texts in English. The tags that interest us are those of part of speech and
the dependency. Once the tags have been created, we create n-grams of the tokenized words (not of the
tags), in the end both the words, as well as the n-grams, and the tags are counted to see their occurrences
by class and measure their frequency of occurrence. This information will be used to later calculate the
probabilities of occurrence. In Fig. 1, this process is observed.</p>
      </sec>
      <sec id="sec-4-4">
        <title>In the pre-processing process we obtain information related to the frequency of appearance of words, n-grams (N-gram to 2-grams, 3-grams and 4-grams), POS labels and dependency labels, this same information was evaluated implementing a naive bayes model, to choose the best characteristics a search of grid was done to explore the best combination of data.</title>
        <p>In Figure 2 there is an example of how the input to the naive bayes model would be, the final features
selected were the frequencies of appearance of the following elements:
• N-gram data: 3-gram words (combinations of 3 words in the text).
• POS data: Individual tags for verbs (VB, VBN, VBD), adverbs (RB, RBR, RBS, WBR, RP),
nouns (NNP), punctuation (LS) and spaces (_SP).
• Dependency data: Index of token head dependency (the position with respect to a sentence of
the token heads of the dependency relations).</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Subtask 1 - Spanish</title>
      <sec id="sec-5-1">
        <title>It is decided to use statistical characteristics for the analysis of the texts. As demonstrated in [3],</title>
        <p>identification based on these metrics is possible. The method proposed here is based on the idea that a</p>
      </sec>
      <sec id="sec-5-2">
        <title>TGM will tend to use certain types of words more than others. This same rule applies to the use of</title>
        <p>bigrams, it is expected that computational models will tend to use some type of bigram mostly. In the
same case with the richness of the vocabulary, naturally the generated texts could have a lower than
human richness, depending on the texts with which the model that writes them has been trained.</p>
      </sec>
      <sec id="sec-5-3">
        <title>AuTexTification task [9], belonging to the 2023 edition of IberLEF has a useful dataset, it consists</title>
        <p>of a corpus of texts in English and one of texts in Spanish, both corpus with the instances labeled as
human and generated by computer. The method is applied to the texts of both languages. In total, 20,129
texts in Spanish were used for tests and 21,832 in English, for training 33,845 texts in English and
32,062 in Spanish were used.</p>
      </sec>
      <sec id="sec-5-4">
        <title>First, punctuation marks are removed to preserve only the words or tokens that compose the text. At this point, stop words are eliminated , that is, those high-frequency words that do not add any value to the text such as articles, prepositions, pronouns, etc. [10].</title>
      </sec>
      <sec id="sec-5-5">
        <title>Subsequently, the prevailing text is tagged using Part -of- Speech tagging (POS tagging), in this</title>
        <p>process each token is tagged within a morphosyntactic category such as noun, adjective, personal
pronoun, etc. [11]. The purpose is to obtain information on the structure of the sentences that are present
in the text. The SpaCy library for Python performs this procedure, recognizing 20 different tags. An
example of POS tagging made in the parts-of-speech.info site can be seen in Figure 3</p>
      </sec>
      <sec id="sec-5-6">
        <title>I am a text being tagged with POS tagging</title>
        <sec id="sec-5-6-1">
          <title>Pronoun Verb Determiner Noun Verb Verb Preposition Noun Noun</title>
          <p>,</p>
          <p>Where   is the proportion of the i-th POS tag, the numerator is the number of occurrences of the
ith POS tag, and the denominator is the total number of words in the text. For each of the i tags considered
by the Spacy library, in this case 20, we get the first 20 features.</p>
          <p>
            POS bigram ratio is also examined. Since we have 20 possible tags, when generating bigrams up to
400 different pairs can be established. The proportion of bigrams is obtained by the expression:
(
            <xref ref-type="bibr" rid="ref1">1</xref>
            )
(
            <xref ref-type="bibr" rid="ref2">2</xref>
            )
 
( ,  )
,
          </p>
        </sec>
      </sec>
      <sec id="sec-5-7">
        <title>Where pbij is the proportion of tagged i,j bigrams for each pair of consecutive words. The numerator</title>
        <p>indicates the number of labeled bigrams i,j for each pair of consecutive words, and the denominator is
the total bigrams of consecutive words. For a text with n words, there are n-1 bigrams, this procedure
yields 400 additional features.</p>
      </sec>
      <sec id="sec-5-8">
        <title>In addition, two vocabulary richness metrics are defined. The first of these is the Standardized</title>
      </sec>
      <sec id="sec-5-9">
        <title>Token-Type Ratio (STTR). This calculation determines the proportion of different tokens, unique words or types used by a given number of words [12]. The number of words may vary depending on the purpose; for this job the number of tokens from the text is used, see Equation 3.</title>
        <p>=
Another useful measure is the indicator λ defined by [13]:</p>
        <p>,
,
 =</p>
        <p>∑ =1 √(  −   +1)2 + 1
 
,</p>
      </sec>
      <sec id="sec-5-10">
        <title>Where N is the size of the text, fi are the absolute frequencies in ascending order of each of the types,</title>
        <p>and V is the number of types. With these metrics, two more features are added.</p>
      </sec>
      <sec id="sec-5-11">
        <title>Finally, the proportion of collocations will be added. It is expected that a TGM will have less use of</title>
        <p>this type of bigrams. Collocations are multiple word expressions, usually bigrams, that often go
together, but none of them can be changed to a synonym without losing their meaning. For example,
"red wine" cannot be substituted for "reddish wine" [14]. The NLTK library for Python is capable of
placing collocations in a text. Therefore, the ratio of collocations to the number of bigrams in the text
is calculated:</p>
        <p>
          In total, there are 423 features for each text, which are passed to a neural network for training as a
classifier. The Neural Network was an architecture Feed Forward with 2 hidden layers with 423 neurons
per layer, batch size of 200 elements, adaptive learning rate with value of 0.001 and momentum of 0.9,
and 500 epochs, the activation function was ReLu. The diagram of the method can be seen in Figure 4.
(
          <xref ref-type="bibr" rid="ref3">3</xref>
          )
(
          <xref ref-type="bibr" rid="ref4">4</xref>
          )
(
          <xref ref-type="bibr" rid="ref5">5</xref>
          )
        </p>
        <p>Remove
punctuation</p>
        <p>POS
tagging</p>
        <p>Remove
stop words</p>
        <p>POS
proportion
POS bigram
proportion
Calculate</p>
        <p>STTR
Calculate</p>
        <p>Collocations
proportion</p>
        <p>Artificial
Neural
Network</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>4. Results</title>
      <sec id="sec-6-1">
        <title>Two neural networks were trained, one for English texts and one for Spanish texts. The data was</title>
        <p>split into training set with 80% of the data and test set with the remaining 20%, this was applied to both
languajes. The final performance metric of the classifier is the Macro F1 Score, although precision and
recall are also shown. To understand how it works, it is necessary to refer to the confusion matrix.</p>
      </sec>
      <sec id="sec-6-2">
        <title>In a diagnostic test or binary classifier, whose results can be positive or negative, we have the class labels for each instance and it is assumed that they are correct</title>
      </sec>
      <sec id="sec-6-3">
        <title>From which the following metrics [8] are derived: Precision: Measures how many instances predicted as positive are actually positive, it is often used when seeking to reduce the number of false positives:</title>
      </sec>
      <sec id="sec-6-4">
        <title>Recall: Measures how many positive instances are captured by positive predictions, it is used when trying to avoid false negatives:</title>
        <p>
          From where the F1 metric is calculated as a way of combining both results:
 =
 =
,
,
,
(
          <xref ref-type="bibr" rid="ref6">6</xref>
          )
(
          <xref ref-type="bibr" rid="ref7">7</xref>
          )
(
          <xref ref-type="bibr" rid="ref8">8</xref>
          )
(9)
        </p>
        <p>These metrics assume that we are forcing a class to be positive, however it is worth checking the F1
when the positive prediction is the human class and when it is the generated class, so the F1 Macro is
used, it’s applied for multi class classification:
,</p>
        <p>Now, the confusion matrices for the English texts are presented in Fig 6. It can be seen that there are
more positive than negative instances, and the system has better identified negative instances than
positive ones
ty 9665
i
l
a
e
R 5385
(a)</p>
        <sec id="sec-6-4-1">
          <title>Diagnoses 1525 5257</title>
        </sec>
        <sec id="sec-6-4-2">
          <title>Diagnoses</title>
          <p>ty 5257
i
l
a
e
R 1525
5385
9665
(b)</p>
        </sec>
      </sec>
      <sec id="sec-6-5">
        <title>Confusion matrix for the texts in Spanish is presented in Fig 7:</title>
        <sec id="sec-6-5-1">
          <title>Diagnoses</title>
        </sec>
      </sec>
      <sec id="sec-6-6">
        <title>It can be seen that the classifiers have a higher precision for the human class. The recall suggests that real-world cases of computer-generated text are better covered.</title>
      </sec>
      <sec id="sec-6-7">
        <title>Also, the F1 overall is superior when evaluating the computer generated class. The results are not</title>
        <p>given in terms of accuracy, so a direct comparison with related works cannot be made. However, the
overall results of Macro F1 indicate a modest performance that is close to the work done so far and
without the need to modify deep networks. The classifier in English could result in a higher score
because the SpaCy library has a more extensive corpus in that language, which could make POS tagging
more appropriate than that done in Spanish.</p>
      </sec>
      <sec id="sec-6-8">
        <title>Also, we can compare our results with the published results for the test set in AuTexTification subtask 1 [9], see Table 2.</title>
        <p>Proposed Classifiers
Languaje
English
Spanish
English
Spanish
English
Spanish
English
English
Spanish
English
Spanish
English</p>
      </sec>
      <sec id="sec-6-9">
        <title>It can be seen that the classifiers have a higher precision for the human class. The recall suggests that real-world cases of computer-generated text are better covered.</title>
      </sec>
      <sec id="sec-6-10">
        <title>Also, for texts in English, it can be observed that a Macro F1 higher by 1.22 units is obtained with</title>
        <p>respect to Logistic Regression, which is the best baseline result, in addition to exceeding Symanto</p>
      </sec>
      <sec id="sec-6-11">
        <title>Brain's 59.44 (Few-shot) and above RoBERTa's and its 57.1. On the other hand, the method for texts is</title>
      </sec>
      <sec id="sec-6-12">
        <title>Spanish obtained a Macro F1 of 64.41, which places it below the 68.52 of RoBERTa but above the 62.4</title>
        <p>of Logistic Regression and the rest of the baseline methods. In English, a notable improvement has been
obtained with respect to the baseline classifiers, while in Spanish the result of the majority has been
exceeded, being only surpassed by RoBERTa.</p>
      </sec>
    </sec>
    <sec id="sec-7">
      <title>5. Conclusions</title>
      <sec id="sec-7-1">
        <title>In this work, a text classification task was carried out seeking to differentiate between those</title>
        <p>generated by TGM and those generated by humans. Preprocessing was carried out on the texts in</p>
      </sec>
      <sec id="sec-7-2">
        <title>English and Spanish, which included removal of punctuation, stop words, and POS tagging. In addition,</title>
        <p>a series of statistical descriptors, including bigram usage and vocabulary richness, were determined and
used to train a neural network. The results indicate that the model is moderately successful, reaching</p>
      </sec>
      <sec id="sec-7-3">
        <title>Macro F1 scores of 0.67 for English texts and 0.64 for Spanish texts. A higher precision was obtained</title>
        <p>with the human texts, but a better recall with the generated ones. Also, the F1 suggests better detection
of generated texts.</p>
      </sec>
      <sec id="sec-7-4">
        <title>These numbers show that the classifier can distinguish computer-generated texts from human texts</title>
        <p>with some reliability, although it is clear that there is a lot of room for improvement. Future works may
include the use of additional features or the implementation of different artificial intelligence models.</p>
      </sec>
      <sec id="sec-7-5">
        <title>Obtaining better results in English suggests that there may be language factors that could be explored</title>
        <p>further. In the same way, the use of other ways of text processing can be reviewed.</p>
        <p>In general, it was shown that the use of statistical features instead of the use of pre-trained deep
networks is feasible and can present better results with proper direction and methodology. The
importance of artificial text detection will increase in the near future due to the rapid growth of generator
models and the risks associated with their unethical use. This work represents an initial step in that
direction.</p>
      </sec>
    </sec>
    <sec id="sec-8">
      <title>6. References</title>
      <p>[9] Sarvazyan, A. M., Gonzalez, J., Franco Salvador, M., Rangel, F., Chulvi, B., &amp; Rosso, P. (2023).</p>
      <sec id="sec-8-1">
        <title>Overview of AuTexTification at IberLEF 2023: Detection and Attribution of Machine-Generated</title>
      </sec>
      <sec id="sec-8-2">
        <title>Text in Multiple Domains. In Procesamiento del Lenguaje Natural..</title>
        <p>[10] Müller, A. C., &amp; Guido, S. (2016b). Introduction to Machine Learning with Python: A Guide for</p>
      </sec>
      <sec id="sec-8-3">
        <title>Data Scientists. O’Reilly Media.</title>
        <p>[11] Sammut, C., &amp; Webb, G. I. (2017). Encyclopedia of Machine Learning and Data Mining. Springer.
[12] WordSmith Tools. (s. f.).</p>
        <p>https://lexically.net/downloads/version5/HTML/index.html?type_to%20ken_ratio_pr%20oc.htm.
[13] Fang, Y., &amp; Liu, H. (2015). Comparison of vocabulary richness in two translated Honglou-meng.</p>
      </sec>
      <sec id="sec-8-4">
        <title>Glottometrics, 31, 54-75.</title>
        <p>[14] Bird, S., Klein, E., &amp; Loper, E. (2009). Natural Language Processing with Python. O’Reilly Media.</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>Jawahar</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Abdul-Mageed</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Lakshmanan</surname>
            ,
            <given-names>L. V. S.</given-names>
          </string-name>
          (
          <year>2020</year>
          ).
          <article-title>Automatic Detection of Machine Generated Text: A Critical Survey</article-title>
          . https://doi.org/10.18653/v1/
          <year>2020</year>
          .coling-main.
          <volume>208</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <surname>Massarelli</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Petroni</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Piktus</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ott</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rocktäschel</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Plachouras</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Silvestri</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Riedel</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          (
          <year>2019</year>
          ).
          <article-title>How Decoding Strategies Affect the Verifiability of Generated Text</article-title>
          . arXiv (Cornell University). https://doi.org/10.48550/arxiv.
          <year>1911</year>
          .
          <volume>03587</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>Gehrmann</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Strobelt</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Rush</surname>
            ,
            <given-names>A. M.</given-names>
          </string-name>
          (
          <year>2019</year>
          ).
          <article-title>GLTR: Statistical Detection and Visualization of Generated Text</article-title>
          . https://doi.org/10.18653/v1/p19-
          <fpage>3019</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <surname>Ippolito</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Duckworth</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Callison-Burch</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Eck</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          (
          <year>2020b</year>
          ).
          <article-title>Automatic Detection of Generated Text is Easiest when Humans are Fooled</article-title>
          . https://doi.org/10.18653/v1/
          <year>2020</year>
          .aclmain.164
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>Kirchenbauer</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Geiping</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wen</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Katz</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Miers</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Goldstein</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          (
          <year>2023</year>
          ).
          <article-title>A Wa-termark for Large Language Models</article-title>
          . arXiv (Cornell University). https://doi.org/10.48550/arxiv.2301.10226.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>Gritsay</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Grabovoy</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Chekhovich</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          (
          <year>2022</year>
          ).
          <source>Automatic Detection of Machine Generated Texts: Need More Tokens</source>
          . https://doi.org/10.1109/ivmem57067.
          <year>2022</year>
          .
          <volume>9983964</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>AminiMotlagh</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shahhoseini</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Fatehi</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          (
          <year>2023</year>
          ).
          <article-title>A reliable sentiment analysis for classification of tweets in social networks</article-title>
          .
          <source>Soc. Netw. Anal. Min</source>
          .
          <volume>13</volume>
          , 7 https://doi.org/10.1007/s13278-022-00998-2.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <surname>Dimlioglu</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bisla</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          et al. (
          <year>2023</year>
          ).
          <article-title>Automatic document classification via transformers for regulations compliance management in large utility companies</article-title>
          .
          <source>Neural Comput &amp; Applic</source>
          https://doi.org/10.1007/s00521-023-08555-4.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>