<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <issn pub-type="ppub">1613-0073</issn>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Models⋆</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Renan Lirio de Souza</string-name>
          <email>rliriodesouza@fbk.eu</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Mauro Dragoni</string-name>
          <email>dragoni@fbk.eu</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Workshop</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Data Augmentation</institution>
          ,
          <addr-line>Natural Language Processing, Large Language Models, Persuasive Dialog Systems, Argu-</addr-line>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Fondazione Bruno Kessler (FBK)</institution>
          ,
          <addr-line>Via Sommarive 18, 38123 Trento</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Free University of Bolzano (UNIBZ)</institution>
          ,
          <addr-line>piazza Università 1, 39100 Bozen-Bolzano</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2024</year>
      </pub-date>
      <fpage>25</fpage>
      <lpage>28</lpage>
      <abstract>
        <p>Data augmentation comprises a set of techniques that are used to improve the performance of machine learning models. In the Natural Language Processing (NLP) field, the discrete nature of language represents a challenge to data augmentation, as text can lose coherence or syntactic accuracy during the augmentation process. This challenge is particularly pronounced in persuasive dialog systems, where high-quality data is crucial and scarce due to privacy regulations. In this study, we investigate the application of Large Language Models (LLMs) to enhance persuasive arguments from an Automatic Persuasive System (APS). Using a limited COVID-19 dialogue dataset of user-machine persuasive interactions, our goal is to evaluate the eficacy of diferent augmentation techniques, combined with LLMs, in generating syntactically coherent and accurate dialogues, with a specific emphasis on dialogue quality and persuasiveness. The results show that integrating LLMs with augmentation generates realistic and diverse examples, aiding the overall quality and efectiveness of the persuasive dialogues produced by the system.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ].
      </p>
      <p>
        One common approach to address data scarcity issues is applying text-data augmentation (DA)
techniques. DA involves generating new text data from existing samples, without the need for additional
real-world data collection [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. This process can enhance the volume and diversity of training datasets
and improve the performance of the model in NLP tasks such as text classification, machine translation,
and dialogue generation by generating additional samples with label-preserving transformations [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ].
However, traditional DA strategies often rely on simple textual transformations, such as random word
replacement, addition, removal, or swapping [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. Although these methods can enhance model robustness
against minor variations, they may not be suficient to generate good persuasive arguments.
      </p>
      <p>
        The emergence of Large Language Models (LLMs) such as GPT [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] has introduced a new era of
possibilities for textual DA. Trained on vast and diverse datasets, these models excel at generating
coherent and contextually appropriate text across various domains and can be tailored to act as possible
data augmentators. The main goal of my PhD is to develop and evaluate a method that implements
LLM’s capabilities to generate novel and contextually accurate persuasive dialogues and arguments for
an APS system. We applied this method as a case study using a COVID-19 dataset of user-machine
persuasive interactions, composed of user arguments and concerns. In this paper, we used a small
      </p>
      <p>CEUR</p>
      <p>ceur-ws.org
sample size of arguments from the COVID-19 dataset to test and evaluate the method by applying three
diferent augmentation approaches: paraphrasing, backtranslation, and masking.</p>
    </sec>
    <sec id="sec-2">
      <title>1. Dataset</title>
      <p>
        The dataset used to evaluate our approach is based on the work of [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. They developed and
provided a knowledge base consisting of a set of aggregated arguments and counterarguments related
to common concerns about COVID-19 vaccination. From the available datasets, we choose the
  _   _ _ dataset, which is composed of tuples of arguments ( ) and concerns ( ).
We defined it as   , as presented in Table 1, where:
 
  = {(  ,   , C )}=1
with each tuple (  ) consisting of a unique identifier (  ), an argument ( ) and a concern ( ), where:
•   represents the unique identifier for the  ℎ argument.
•   contains a user interaction of the  ℎ argument, which includes a user’s dialogue against getting
a COVID-19 vaccine.
•   is the category of concern that the  ℎ argument addresses, such as side efects, vaccine
development speed, eficacy, safety, etc.
•   is the total number of arguments in the base dataset.
      </p>
      <p>id
1
2
3
4
5
6
7</p>
      <p>Concern
healthy
healthy
healthy
healthy
healthy
healthy
healthy</p>
      <p>Argument</p>
      <p>I have strong immunity</p>
      <p>I am not worried about my health anyway
I am healthy and not worried to catch COVID</p>
      <p>I know that, but I’m in good health for my age</p>
      <p>I’d rather let my immune system develop naturally</p>
      <p>I prefer my chances against Covid 19, I am young and healthy
True but I am young and healthy and I don’t think I will need the vaccine</p>
      <p>We randomly selected these 7 arguments of the ℎℎ concern to be used as original data for
augmentation purposes and evaluated our proposed augmentation method. We also selected only 7
arguments because of the limited number of pages allowed for the paper, as more selected arguments
could lead to more augmented output examples. As the main idea is to increase the number of existing
examples in   with synthetic arguments, we choose the ℎℎ concern as it is a class with a total
amount of values in the middle compared with other concerns, as presented in 2.</p>
      <p>The complete dataset comprises, in total, 820 user arguments ( ), divided into 15 distinct concerns
( ). Table 2 also provides a breakdown of the frequency of each concern, measured in terms of the
number of arguments associated with it. The most common concerns are ”long-term efects” (19%),
”safety” (17%), and ”side efects” (15%). The dataset presents a clear imbalance in the distribution of
arguments, a common and well-documented issue in the literature. Data augmentation (DA) techniques
ofer an efective solution to address this imbalance, especially in improving representation among the
underrepresented classes.</p>
    </sec>
    <sec id="sec-3">
      <title>2. Method</title>
      <p>
        Traditional DA techniques typically use a simplistic strategy that modifies a given sentence to improve
classification algorithms’ generalization capabilities [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. While this approach is efective to a certain
      </p>
      <p>Concerns
long_term_efects</p>
      <p>⋮
healthy</p>
      <p>
        ⋮
already_had
15
extent, it may lack the ability to generate novel coherent data for humans [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. However, this paradigm
is changing with the advent of Large Language Models (LLMs). Researchers are currently using LLMs
to develop novel augmentation strategies tailored to the challenges of the NLP domain [
        <xref ref-type="bibr" rid="ref10 ref7">7, 10</xref>
        ]. The
study in [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] uses LLMs like ChatGPT for text augmentation by rephrasing sentences into semantically
distinct variants, yielding improvements even in few-shot learning scenarios. In [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], GPT-generated
samples enhanced the classification of underrepresented vaccine hesitancy in Dutch social media, and
with back-translation, improved model accuracy and F1 scores. These findings show that GPT models
can create realistic, diverse examples, enhancing training in imbalanced datasets.
      </p>
      <p>
        Although LLM models have shown good performance in data augmentation, their performance
varies with diferent datasets and tasks [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]. These advancements have expanded the possibilities for
data augmentation applications, enabling more advanced and eficient approaches. However, more
research is needed to establish standardized practices, especially for complex tasks such as argument
augmentation for persuasive dialog systems.
      </p>
      <p>In response to the lack of DA methods for persuasive argumentation, we investigated the use of
LLMs to generate synthetic arguments. By exploiting the advanced capabilities of these models, we aim
to develop diverse and contextually relevant arguments to enhance   . Several models were evaluated,
including GPT, Gemini, BERT, and DistilBERT, but we are going to use only GPT. As a result, we
formulated two specific prompts to interact with these models, as shown below.
2.1. Prompts Layout
We designed three diferent prompts to interact with the LLM, by organizing them following a structure
(Figures 1a and 1b) with three main sections: Instruction, Examples, and Output.</p>
      <p>###Instruction:
Create 2 sentences for each example
using paraphrasing augmentation.
###Examples (Healthy concern):
1. I have strong immunity
2. Not worried about my health anyway
  .   (Table 1)
###Instruction:
Create 5 new sentences diferent from the examples,
following the ###Output logic.
###Examples (Healthy concern):
1. I have strong immunity
2. Not worried about my health anyway
  .   (Table 1)
###Output:
&lt;mask&gt; &lt;mask&gt; healthy &lt;mask&gt; &lt;mask&gt; fate &lt;mask&gt;
(a)  and  augmentation prompt.</p>
      <p>(b) Masking augmentation prompt.</p>
      <p>• Instruction. Main description or information for the LLM executing a task. The prompt 1 (Figure
1a) has a unique parameter that is highlighted as underline, which was used withparaphrasing
( ) and back-translation ( ).
• Examples. Examples of arguments from   to provide the model with a comprehensive data
context. During our tests, we used seven arguments from healthy concerns (Table 1).
• Output. This section provides the template or structure (in the form of a sentence with blanks) that
the model should follow when creating the augmented sentences. The placeholders, represented
by ”&lt;mask&gt;”, indicate where the model needs to fill in words that will create a coherent sentence.
In the example provided in Figure 1b, the model needs to fill the &lt;mask&gt; values around ”healthy”
and ”fate” with words that match the tone and meaning of the examples. We used these two
words with the &lt;mask&gt; values to have more control over sentence length, generation process, and
contextual integrity. The word ”healthy” was chosen because it addresses the concern  itself.
Conversely, the word ”fate” was randomly selected from a bag of words with partial relevance to
the topic, while also introducing subtle variation into the final sentence and forcing the model to
follow the implied concept.</p>
    </sec>
    <sec id="sec-4">
      <title>3. Experiments</title>
      <p>We execute all prompts using Python code using the API access for GPT model. However, all prompts
can be easily executed via the standard web browsers for GPT or Gemini. In this section, we will present
the results obtained using the GPT model. The sentences resulting from the first two prompts (Figure
1a, with paraphrasing and backtranslation), are shown in Tables 3 and 4.</p>
      <p># Paraphrasing ( )
1 My immune system is quite robust
2 I don’t have any concerns about my health
3 I’m in good health and not concerned about getting COVID
4 I understand, but I’m in good shape for my age
5 I’d rather let my body build its immunity naturally
6 I feel confident about facing COVID-19 since I’m young and fit
7 That’s true, but I’m young and healthy, so I don’t think I’ll need the vaccine</p>
      <p># Backtranslation ( )
1 I possess strong immunity
2 I’m not concerned about my health anyway
3 I’m healthy and not worried about getting COVID
4 I know that, but I am in good health for my age
5 I prefer to let my immune system develop naturally
6 I prefer my chances against COVID-19, I’m young and healthy
7 It’s true, but I’m young and healthy, and I don’t think I’ll need the vaccine</p>
      <p>A qualitative comparison of the tables with the original sentences reveals distinct variations for
each technique. The paraphrased sentences show minor changes, adopting a conversational tone,
while retaining the original meaning. In contrast, backtranslation results in sentences with negligible
changes. This can be attributed to the choice of language, suggesting that a combination of diferent
languages may yield better results. The results of the masking augmentation, as presented in table 5,
show significant variation in both structure and meaning, incorporating an element of randomness
while maintaining the integrity of individual words. The generated sentences are also similar in length
and include both predefined words. The terms ”healthy” and ”fate” have been added to the table (last
two inputs) as ”fake” sentences to evaluate the impact that each word has when compared with the
original sentences.</p>
      <p>#
3
5
6</p>
      <p>healthy</p>
      <sec id="sec-4-1">
        <title>7 fate</title>
      </sec>
      <sec id="sec-4-2">
        <title>Masking ( )</title>
        <p>1 I trust being healthy and believe my fate will protect me
2 I am staying healthy and leaving my fate to nature’s course</p>
        <p>Remaining healthy is important, but I accept my fate either way
4 I focus on staying healthy and let fate decide the rest</p>
        <p>Being healthy gives me confidence to leave fate as it is
is suitable for clustering, semantic search, and other tasks.</p>
        <p>Initially, we obtained the embeddings of the original arguments (Table 2) as well as the embeddings
of the masked results (Table 5), and performed a comprehensive comparison of all sentence embeddings.
The resulting similarity matrix, shown in Table 6a, provides a complete comparison between the original
arguments and the corresponding masked versions. Since the masked results are not generated on a
one-to-one basis (i.e., original sentence 1 to masked sentence 1), the comprehensive matrix provides
a more detailed understanding of the similarity results. Paraphrasing and back-translation (Table 6b)
only need diagonal entries of the matrix; as they achieve modifications at the sentence level, hence, it is
not necessary to perform evaluations with the others.</p>
        <p>1
2
3
4
5
6
7</p>
        <p>1
0.34
0.41
0.39
0.42
0.37
0.32
0.40</p>
        <p>2
0.25
0.44
0.46
0.46
0.34
0.36
0.29</p>
        <p>3
0.30
0.45
0.45
0.43
0.38
0.37
0.34</p>
        <p>4
0.13
0.43
0.47
0.40
0.34
0.39
0.27</p>
        <p>5
0.22
0.40
0.30
0.41
0.35
0.25
0.31</p>
        <p>6
0.36
0.50
0.50
0.56
0.30
0.31
0.37</p>
        <p>7
0.08
-0.02
0.07
0.04
0.20
0.11
0.10
(a) Full matrix of   and Mask 
  #
1
2
3
4
5
6
7

0.34
0.44
0.45
0.40
0.35
0.31
0.10

0.78
0.79
0.84
0.77
0.86
0.83
0.94</p>
        <p>(b) Main diagonal
0.98
0.92
0.92
0.98
0.91
0.98
0.93
original   . Both paraphrasing and backtranslation scores show much higher similarity than masking  .
The  method yields results ranging from 0.77 to 0.94, making it ideal for varied, yet faithful sentences.
The  method produces an almost identical structure with the highest similarity scores, providing
a safer technique for preserving sentences within the same class. Finally, masking introduces more
randomness and diversity, while providing control by allowing users to define specific keywords or
masks to generate more varied arguments.</p>
        <p>In this study, we explored the integration of Large Language Models (LLMs) with data augmentation
techniques to improve the quality and efectiveness of persuasive dialogues generated by an Automated
Persuasion System (APS). By applying methods such as paraphrasing, backtranslation, and masking to
a limited COVID-19 dialogue dataset, we were able to produce syntactically coherent and contextually
accurate examples. These techniques not only could address the data scarcity issue but also improve
the overall efectiveness of the APS. Future work will focus on refining these approaches and exploring
additional augmentation strategies like user behavior, profile and clustering by types of users.</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>A.</given-names>
            <surname>Hunter</surname>
          </string-name>
          ,
          <article-title>Computational persuasion with applications in behaviour change</article-title>
          .,
          <source>in: COMMA</source>
          ,
          <year>2016</year>
          , pp.
          <fpage>5</fpage>
          -
          <lpage>18</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>M.</given-names>
            <surname>Bayer</surname>
          </string-name>
          , M.
          <article-title>-</article-title>
          <string-name>
            <surname>A. Kaufhold</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Reuter</surname>
          </string-name>
          ,
          <article-title>A survey on data augmentation for text classification</article-title>
          ,
          <source>ACM Computing Surveys</source>
          <volume>55</volume>
          (
          <year>2022</year>
          )
          <fpage>1</fpage>
          -
          <lpage>39</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>L.</given-names>
            <surname>Chalaguine</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Hunter</surname>
          </string-name>
          ,
          <article-title>Addressing popular concerns regarding covid-19 vaccination with natural language argumentation dialogues, in: Symbolic and Quantitative Approaches to Reasoning with Uncertainty: 16th European Conference</article-title>
          ,
          <string-name>
            <surname>ECSQARU</surname>
          </string-name>
          <year>2021</year>
          , Prague, Czech Republic,
          <source>September 21-24</source>
          ,
          <year>2021</year>
          , Proceedings 16, Springer,
          <year>2021</year>
          , pp.
          <fpage>59</fpage>
          -
          <lpage>73</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>S.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Ao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Pan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <article-title>Learning policy scheduling for text augmentation</article-title>
          ,
          <source>Neural Networks</source>
          <volume>145</volume>
          (
          <year>2022</year>
          )
          <fpage>121</fpage>
          -
          <lpage>127</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>J.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Tam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Rafel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Bansal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <article-title>An Empirical Survey of Data Augmentation for Limited Data Learning in NLP, Transactions of the Association for Computational Linguistics 11 (</article-title>
          <year>2023</year>
          )
          <fpage>191</fpage>
          -
          <lpage>211</lpage>
          . URL: https://doi.org/10.1162/tacl_a_
          <fpage>00542</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>A.</given-names>
            <surname>Anaby-Tavor</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Carmeli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Goldbraich</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Kantor</surname>
          </string-name>
          , G. Kour,
          <string-name>
            <given-names>S.</given-names>
            <surname>Shlomov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Tepper</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Zwerdling</surname>
          </string-name>
          ,
          <article-title>Do not have enough data? deep learning to the rescue!</article-title>
          ,
          <source>in: Proceedings of the AAAI Conference on Artificial Intelligence</source>
          , volume
          <volume>34</volume>
          ,
          <year>2020</year>
          , pp.
          <fpage>7383</fpage>
          -
          <lpage>7390</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>H.</given-names>
            <surname>Dai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Liao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Cao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Xu</surname>
          </string-name>
          , W. Liu,
          <string-name>
            <given-names>N.</given-names>
            <surname>Liu</surname>
          </string-name>
          , et al.,
          <article-title>Auggpt: Leveraging chatgpt for text data augmentation</article-title>
          ,
          <source>arXiv preprint arXiv:2302.13007</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Hu</surname>
          </string-name>
          ,
          <string-name>
            <surname>R. K.-W. Lee</surname>
            ,
            <given-names>N. F.</given-names>
          </string-name>
          <string-name>
            <surname>Chen</surname>
          </string-name>
          ,
          <article-title>Are current task-oriented dialogue systems able to satisfy impolite users?</article-title>
          ,
          <source>arXiv preprint arXiv:2210.12942</source>
          (
          <year>2022</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>H.</given-names>
            <surname>Queiroz Abonizio</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. Barbon</given-names>
            <surname>Junior</surname>
          </string-name>
          ,
          <article-title>Pre-trained data augmentation for text classification</article-title>
          ,
          <source>in: Intelligent Systems: 9th Brazilian Conference, BRACIS</source>
          <year>2020</year>
          ,
          <string-name>
            <given-names>Rio</given-names>
            <surname>Grande</surname>
          </string-name>
          , Brazil,
          <source>October 20-23</source>
          ,
          <year>2020</year>
          , Proceedings,
          <string-name>
            <surname>Part</surname>
            <given-names>I</given-names>
          </string-name>
          , Springer,
          <year>2020</year>
          , pp.
          <fpage>551</fpage>
          -
          <lpage>565</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>J. Van Nooten</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          <string-name>
            <surname>Daelemans</surname>
          </string-name>
          ,
          <article-title>Improving dutch vaccine hesitancy monitoring via multi-label data augmentation with gpt-3.5</article-title>
          , in:
          <source>Proceedings of the 13th Workshop on Computational Approaches</source>
          to Subjectivity, Sentiment, &amp;
          <article-title>Social Media Analysis</article-title>
          ,
          <year>July 2023</year>
          ; Toronto, Canada, volume
          <volume>1</volume>
          ,
          <year>2023</year>
          , pp.
          <fpage>251</fpage>
          -
          <lpage>270</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>F.</given-names>
            <surname>Piedboeuf</surname>
          </string-name>
          , P. Langlais,
          <article-title>Is chatgpt the ultimate data augmentation algorithm?</article-title>
          ,
          <source>in: Findings of the Association for Computational Linguistics: EMNLP</source>
          <year>2023</year>
          ,
          <year>2023</year>
          , pp.
          <fpage>15606</fpage>
          -
          <lpage>15615</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>N.</given-names>
            <surname>Reimers</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Gurevych</surname>
          </string-name>
          ,
          <article-title>Sentence-bert: Sentence embeddings using siamese bert-networks</article-title>
          ,
          <source>in: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing, Association for Computational Linguistics</source>
          ,
          <year>2019</year>
          . URL: http://arxiv.org/abs/
          <year>1908</year>
          .10084.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>