<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Automatic detection of Russia-Ukraine war euphemisms ⋆</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Iryna Dilai</string-name>
          <email>iryna.dilay@lnu.edu.ua</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Maksym Davydov</string-name>
          <email>maks.davydov@ucu.edu.ua</email>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Anna Feldman</string-name>
          <email>feldmana@mail.montclair.edu</email>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Olha Oleksyn</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Svitlana Kohut</string-name>
          <email>svitlana.kohut@lnu.edu.ua</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Olha Baranovska</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Ivan Franko National University of Lviv</institution>
          ,
          <addr-line>Universytetska Street 1, 79000, Lviv</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>MoMLeT-2025: 7th International Workshop on Modern Machine Learning Technologies</institution>
          ,
          <addr-line>June, 14, 2025, Lviv-Shatsk</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Montclair State University</institution>
          ,
          <addr-line>Normal Avenue 1, 07043-1624, Montclair, NJ</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>Ukrainian Catholic University</institution>
          ,
          <addr-line>Kozelnytska Street 2, 79026, Lviv</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Automatic detection of figurative language is one of the major directions in modern NLP. Euphemisms are words or phrases used to mitigate the expression. By and large, they are socially and culturally determined, naming the sensitive entities in an indirect, softened way. The problems of the automatic detection of euphemisms arise when words can be used both literally (non-euphemistically) and euphemistically. We refer to such usages as PETs (potentially euphemistic terms). The attempts to detect/disambiguate euphemisms cross-linguistically have reported a high performance of transformerbased neural models. Nonetheless, such models have not been tested on Ukrainian datasets. The purpose of this endeavor is to test LLMs on the collected, annotated, and processed Ukrainian dataset, exemplified in this paper by the newly coined during the Russia-Ukraine war PETs. Employing prompt engineering, the study has revealed a high performance of GPT-4o and GPT-4o-mini on the Ukrainian PET dataset.</p>
      </abstract>
      <kwd-group>
        <kwd>Euphemism</kwd>
        <kwd>automatic detection</kwd>
        <kwd>NLP</kwd>
        <kwd>FLP</kwd>
        <kwd>LLM</kwd>
        <kwd>prompt engineering</kwd>
        <kwd>Russia-Ukraine war</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Despite being an important element of language use, the figurative nature of euphemisms poses
a challenge for natural language processing (NLP). Due to the polysemous nature of the potentially
euphemistic terms (PETs), the detection and recognition of their euphemistic usages requires the
elaboration of viable mechanisms of word sense disambiguation. The semantic annotation scheme
applied to PETs poses difficulties as it needs to consider subtle context-sensitive instances with
various shades of meaning. Thus, we hypothesize that drawing on manual annotation of multiple
instances of PETs allows for approaching the full specification principle (Lakoff) [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] in the
description of the word meaning (conceptual category), which can further feed large language
models (LLMs) and train them to detect and recognize euphemistic expressions of the Ukrainian
language.
      </p>
      <p>Thus, the aim of this paper is to discover efficient techniques for the automatic detection of
Ukrainian euphemisms related to the topic of the Russia-Ukraine war, which presupposes the
completion of the following tasks:




</p>
      <p>To collect a dataset of Ukrainian war-related euphemisms based on the corpus of modern
web communication.</p>
      <p>To elaborate, standardize, and apply the annotation scheme for PETs.</p>
      <p>To elicit key difficulties in the recognition and annotation of the PETs.</p>
      <p>To leverage machine learning techniques for the automatic detection of euphemisms.
To assess the performance of the models.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Works</title>
      <p>
        In recent years, there has been a surge of interest in computational approaches to euphemism
detection in the NLP community. [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] introduce the recognition of euphemisms and dysphemisms
using NLP, generating near-synonym phrases for sensitive topics. [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] propose euphemism
detection and identification tasks using masked language modeling with BERT. [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] create an
extensive corpus of potentially euphemistic terms (PETs). In [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], they develop a linguistically
driven approach for identifying PETs using distributional similarities. BERT-based systems that
participated in a shared task on euphemism disambiguation they organized showed promise [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ].
[
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] experiment with classifying PETs unseen during training. In [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ], they perform
transformerbased euphemism disambiguation experiments, exploring vagueness as one of the properties of
euphemisms.
      </p>
      <p>
        The work of R. Choenni et al. [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] explores the multilingual and cross-lingual transfer
capabilities of LLMs. They find that multilingual LLMs rely on data from multiple languages to a
large extent, learning both complementary and reinforcing information. The authors of [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] find
cases where transfer learning from out-of-language data in a particular domain performed better
than the same-language data in a different domain.
      </p>
      <p>While euphemisms are culturally dependent, the need to discuss sensitive topics in a
nonoffensive way is universal, suggesting similarities in the way euphemisms are used across
languages and cultures. Euphemisms are found across the world’s languages, making them a
universal linguistic phenomenon. As such, euphemistic data may have useful properties for
computational tasks across languages. A. Feldman and her team have explored this premise by
training a multilingual transformer model (XLM-RoBERTa) to disambiguate potentially
euphemistic terms (PETs) in multilingual and cross-lingual settings. They have conducted
experiments on English, Spanish, Chinese, Turkish, and some other low-resource languages. In line
with current trends, they demonstrate that zero-shot learning across languages takes place. They
also showcase where multilingual models perform better on the task compared to monolingual
models by a statistically significant margin, indicating that multilingual data presents additional
opportunities for models to learn about cross-lingual, computational properties of euphemisms. In
a follow-up analysis, they focus on universal euphemistic "categories" such as death and bodily
functions, among others. They test to see whether cross-lingual data of the same domain is more
important than within-language data of other domains to further understand the nature of the
cross-lingual transfer.</p>
      <p>
        In June 2024, a special FigLang (Figurative Language Processing) workshop was held in Mexico,
where the findings of the shared tasks on multilingual euphemism detection were presented.
Among others, [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] tried to test whether Chat GPT can detect euphemisms across multiple
languages.
      </p>
      <p>
        To the best of our knowledge, no similar study on the automatic detection of euphemisms has
been conducted in Ukraine or on Ukrainian language material. Nonetheless, there are works with
valuable observations related to the linguistic aspect of the newly coined Ukrainian military
vocabulary and euphemisms in particular [
        <xref ref-type="bibr" rid="ref13 ref14 ref15 ref16">13–16</xref>
        ]. A revised approach to labeling sensitive
language related to the ongoing war was proposed in [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ].
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Methods and Materials</title>
      <p>The technical solution to the problem of automated Ukrainian euphemisms detection involves
prompt engineering for LLMs. We test both zero-shot prompting, which does not contain any
examples or demonstrations while interacting with the model, and few-shot prompting,
accompanied by illustrations and enabling in-context learning of the model.</p>
      <p>The experimental research is based on GPT-4o, the flagship LLM by OpenAI, GPT-4o-mini, and
DeepSeek, though other models have also been tested but showed worse performance.
GPT-4.5preview was rejected due to its high pricing at the time of testing. Older models (o1, o3-mini,
o1mini) performed notably worse on smaller datasets and were also rejected. DeepSeek-Chat was
chosen as a widely advertised, cheaper alternative to OpenAI models.</p>
      <p>We draw on the F1 score to elaborate on class-wise performance of the LLMs. The overall
workflow consists of the following stages:




</p>
      <p>PET dataset collection.</p>
      <p>PET dataset annotation.</p>
      <p>PET dataset processing.</p>
      <p>Testing zero-shot and few-shot prompting performance of the LLMs.</p>
      <p>Prompt engineering to enhance the performance of the models.</p>
      <p>
        PET samples have been collected from the Polish Automatic Web corpus of the Ukrainian
language (PAWUK) [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ]. It was built and is being maintained by our partner institution the
Linguistic Engineering Group of the Institute of Computer Science of the Polish Academy of
Sciences. The corpus contains Ukrainian texts selected from the web pages and social network
posts starting from February 24, 2022, and is daily updated. As of March 2025, it consists of over
700 million tokens. The PAWUK is lemmatized and accompanied by automatic POS tagging.
      </p>
      <p>The seed list for the dataset encompasses 21 PETs together with their derivatives featuring the
war-related vocabulary. Since most of the identified PETs are polysemous items (e.g., пташка,
бавовна, ціль, мінусувати, etc.), not always used euphemistically, the key problem both for
human annotators and for the AI lies in disambiguating their senses. It can be tackled by
accomplishing their fine-grained annotation and elaborating a machine learning model that would
achieve high performance in the recognition of euphemistic usages.</p>
      <p>The initial stage of testing a euphemism detection model is the collection of a PET dataset. Our
dataset consists of 4,258 instances of Ukrainian PETs referring to the ongoing Russia-Ukraine war,
encompassing both euphemistic and non-euphemistic usages. Table 1 features the resultant
dataset.</p>
      <p>When collecting the dataset based on the PAWUK, we were guided by the following principles:
(1) tried to get a balanced representation of the PET across the three-year period (2022–2025), (2)
tried to represent all wordforms for both euphemistic and non-euphemistic usages, (3) tried to use
a corpus-driven approach, proportionately representing euphemistic and non-euphemistic usages,
their wordforms, etc., (4) tried to identify and include cases hard to classify: instances of pun,
symbolism, intentional vagueness, etc.</p>
      <p>The dataset collection consists of the sentences with the PET, ideally, with the preceding and/or
following sentences to introduce broader context. The node, e.g., &lt;бавовна&gt;, was enclosed in angle
brackets in each sample to facilitate further processing. The annotation stage lies in labeling PETs
as euphemistic with a label (0) and non-euphemistic with a label (1).</p>
      <sec id="sec-3-1">
        <title>Euphemistic</title>
        <p>instances</p>
      </sec>
      <sec id="sec-3-2">
        <title>Non-euphemistic instances</title>
        <p>бавовна
(за)бавовнитися
пташка
дискотека
двохсотий
((за)двохсотити)
трьохсотий
(за)трьохсотити
на щиті
мінусувати
відпрацювати
мопед
приліт
(прилетіти)
втомитися
ціль
м'ясо
спеціальна воєнна
операція
зоряні війни
приземлити
на концерт
Кобзона
закобзонити
батальйон
Монако
за руски/ім
кораблем
дружній вогонь
нуль</p>
        <p>The annotation was done manually by four annotators who are expert linguists. The
interannotator agreement was measured using Cohen’s kappa (к). For the resultant dataset, к = 0.89. The
annotators were asked to mark the cases of uncertainty, attaching the most likely label to the
respective sentences.</p>
        <p>The cases of uncertainty encompassed the samples of distinct play on words (pun), in which the
euphemistic usage keeps traces of the literal one and cannot be discerned without it, e.g., зацвіла
бавовна. Also, we identified the cases of the literal use of this PET with a noticeable shade of the</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Experiment</title>
      <p>new euphemistic sense. In these cases, the PET бавовна is used in the sense of “a flower/plant”
referred to as a symbol of “(victorious) explosion”. Such nuances contribute to the complexity of
PET annotation.</p>
      <p>The dataset processing pinpoints its major quantitative characteristics. Though we tried to
obtain a balanced and representative dataset of the PET, it is limited to (1) the time span (2022–
2025) and (2) the web communication register. Thus, the obvious bias will be towards euphemistic
usage, often accompanied by metaphor, irony, and sarcasm.</p>
      <p>The performance of LLMs is highly affected by a prompt that is passed to interact with the model
and perform the detection of euphemisms. To reduce the mutual impact of data samples on each
other, we rejected batching several data samples into one prompt, although batching reduces the
overall pricing of data processing without a high impact on the performance.</p>
      <p>The experiments were planned to understand the impact of the prompt on LLM
performance. Four types of prompts were chosen (Table 2) for experiments with different
scopes of additional information provided. The language of the prompts (Ukrainian/ English)
did not substantially affect the performance.
[Prompt 1] For each sentence in the set, determine whether the term enclosed in
angle brackets is used as a euphemism (1) or not (0).</p>
      <p>[Prompt 2] For each sentence in the set, determine whether the term enclosed
in angle brackets is used as a euphemism (1) or not (0). Consider the example of
labeling.</p>
      <p>[Prompt 3] You are a linguist. For each sentence in the set, determine whether
the term enclosed in angle brackets is used as a euphemism (1) or not (0).</p>
      <sec id="sec-4-1">
        <title>Consider the terms to be euphemistic in the context of war, the dictionary</title>
        <p>definitions are attached. (The dictionary definitions generated by the GPT-4o
model are provided in Appendix A).</p>
        <p>[Prompt 4] You are a linguist. For each sentence in the set, determine whether
the term enclosed in angle brackets is used as a euphemism (1) or not (0).</p>
      </sec>
      <sec id="sec-4-2">
        <title>Consider the terms to be euphemistic in the context of war; the list of euphemisms is attached. (The attached list of euphemisms without definitions is provided in Appendix B).</title>
        <p>The initial (context-free/zero-shot) prompt was: “For each sentence in the set, determine
whether the term enclosed in angle brackets is used as a euphemism (1) or not (0)”. Table 3 shows
the sample of PET labeling in comparison with the annotators’ labeling.</p>
        <p>The agreement between the annotators and GPT-4o was estimated. For the PET бавовна, the F1
score is equal to 0.77 (Precision = 0.82, Recall = 0.72). For the whole dataset, the F1 score amounts
to 0.81, which is rather high, though individual PETs show different performances (from 0.5 to 0.9).</p>
        <p>The next step was to check if the performance could be augmented after refining the prompt,
providing the AI with a few-shot prompting.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Results</title>
      <p>The performance of LLMs on the PETs dataset largely depends on the type of the model and the
prompt. The DeepSeek-chat model was not significantly affected by the prompt type and its
performance was considerably worse than GPT-4o-mini and GPT-4o models. GPT-4o-mini
performed unexpectedly better than GPT-4o on context-free prompts, regardless of its smaller size.</p>
      <p>Providing definitions of the war-related euphemisms was beneficial for the GPT-4o-mini and
GPT-4o models, but the performance boost was considerably higher for the GPT-4o model (+11%).</p>
      <p>One of the hypotheses was that the LLM improves performance by utilizing a list of
euphemisms in the context of the war without focusing on the meaning of the euphemisms
themselves. To prove or disprove this, we provided a list of euphemisms without an explanation of
their meaning (Prompt 4). The result turned out to be worse than when providing no word samples
at all, as in Prompt 1. The inclusion of 10 random labeled examples of euphemistic and
noneuphemistic usage of words in the prompt (Prompt 2) had no significant impact on the
performance of the models. Moreover, it significantly increases the prompt size, thereby raising
inference costs.</p>
      <p>Table 4 shows the performance of all the models tested on Prompts 1, 3, and 4. The results of
Prompt 2 are omitted because they do not differ significantly from the results of the context-free
prompt.
deepSeek-chat</p>
      <p>Context-free prompt
deepSeek-chat
deepSeek-chat
gpt-4o-mini
gpt-4o-mini
gpt-4o-mini
gpt-4o
gpt-4o
gpt-4o</p>
      <sec id="sec-5-1">
        <title>Prompt with a dictionary definition</title>
      </sec>
      <sec id="sec-5-2">
        <title>Prompt with a word list</title>
      </sec>
      <sec id="sec-5-3">
        <title>Context-free prompt</title>
      </sec>
      <sec id="sec-5-4">
        <title>Prompt with a dictionary definition</title>
      </sec>
      <sec id="sec-5-5">
        <title>Prompt with a word list</title>
      </sec>
      <sec id="sec-5-6">
        <title>Context-free prompt</title>
      </sec>
      <sec id="sec-5-7">
        <title>Prompt with a dictionary definition</title>
      </sec>
      <sec id="sec-5-8">
        <title>Prompt with a word list</title>
        <p>Precision
0.762
0.754</p>
        <p>The detection rate is unevenly distributed among the euphemisms (Table 5). The LLMs’
performance on the PETs відпрацювати, зоряні війни, дискотека, ціль, втомитися was much
worse than the performance on other terms.</p>
        <p>PET
бавовна
двохсотий
приліт
трьохсотий
прилетіти
втомитися
пташка
ціль
спеціальна воєнна операція
на щиті
приземлити
мопед
Батальон Монако
дискотека
зоряні війни
за рускім кораблем
дружній вогонь
на концерт до Кобзона
мінусувати
м'ясо
відпрацювати
нуль</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>6. Discussion</title>
      <p>The quality of annotation has largely predetermined the performance of the model. Among the
major challenges for the annotators were:
1. Labeling instances with play on words (pun).
2. Handling the symbolic usage and metaphoric extensions of different types.
3. Adopting a vantage point, as the same PETs appeared to be euphemistic and have more
positive sentiment when referred to enemy losses but looked dysphemistic and acquired
negative sentiment when referred to one’s own losses (сomp., бавовна в Тернополі).
4. Annotators’ inner bias.
5. Already noticed euphemism treadmill resulting in the tendency to gradually treat them as
rather dysphemistic within broader contexts.</p>
      <p>Engineering LLMs’ prompts that can best detect euphemistic usages in context involved
experimenting with zero-shot and few-shot modes. The highest F1 scores have been achieved by
GPT-4o and GPT-4o-mini for the whole dataset while interacting with a prompt accompanied by
dictionary definitions of the euphemisms under scrutiny.</p>
      <p>It is worth mentioning that the PET dataset is not homogeneous; it comprises clear-cut
instances of euphemisms always labeled with (1), which are generally easier to detect, and
ambiguous instances of polysemous PETs where either euphemistic or non-euphemistic usages
prevail based on the corpus data. The task was also complicated by insufficiency of context in some
cases. Besides, some PETs refer to more than one euphemistic category and, as a result, were
ignored due to the focus of the prompts on the war-related vocabulary.</p>
      <p>Another observation is that though the overall performance of GPT-4o and GPT-4o-mini
achieved on the Ukrainian PET dataset is strikingly high, the models often fail to explain why a
certain word or phrase is euphemistic (they provide wrong synonyms, hypernyms, or definitions).
It demonstrates that even though the correct label has been attached, the models’ understanding of
the sense/usage is incorrect.</p>
    </sec>
    <sec id="sec-7">
      <title>7. Conclusions</title>
      <p>The euphemism treadmill illustrates how language evolves in response to societal attitudes and
how efforts to soften language often fall short of removing the negative associations that these
terms might evoke. It highlights a tension between the desire to use language to be more sensitive
and inclusive and the reality that such efforts can sometimes inadvertently create new stigmas.</p>
      <p>The study has proven that the war-related euphemisms manifest the vast creative potential of
the users and are particularly context-sensitive. As mostly newly coined, euphemisms are a
challenging problem for detection and proper understanding by humans, let alone AI. Nonetheless,
neural network models relying on efficient techniques can easily recognize them and use them in
other applications, including generative AI.</p>
      <p>The implications of this research go beyond computational linguistics and NLP. Ukrainian
warrelated euphemisms designating sensitive topics are a rapidly developing category in the Ukrainian
language, reflecting the new reality and its perception. Thus, the results can be of interest to the
social sciences.</p>
      <p>The prospects of further study lie in testing the models on a larger dataset of Ukrainian PETs
belonging to other categories and employing other, more advanced LLMs.</p>
    </sec>
    <sec id="sec-8">
      <title>Acknowledgements</title>
      <p>This study is supported by the STCU and is part of the broader international collaboration within
the IMPRESS-U project #7132 “DARE: Detecting and Recognizing Euphemisms”, which is a
supplement to the NSF grant #2226006.</p>
    </sec>
    <sec id="sec-9">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the authors used GPT-4o in order to: Paraphrase and reword
in Prompt 3 (Appendix A) and integrate the generated definitions in the experiment.</p>
    </sec>
    <sec id="sec-10">
      <title>Appendices</title>
      <p>Appendix A. Definitions of euphemisms for Prompt 3</p>
      <p>Бавовна – іронічне позначення вибухів, яке виникло через цензуру в російських медіа.
Замінює слово «вибух» у контексті ударів по ворожих об’єктах.</p>
      <p>Двохсотий – військовий термін, що означає загиблого солдата (походить від кодової
назви «вантаж 200» для транспортування тіл загиблих).</p>
      <p>Приліт – потрапляння ракети, снаряду або дрону в ціль, зазвичай супроводжується
вибухом.</p>
      <p>Трьохсотий – військовий термін, що означає пораненого солдата (походить від кодової
назви «вантаж 300» для евакуації поранених).</p>
      <p>Прилетіти – отримати влучання ракетою чи снарядом, зазвичай використовується щодо
обстрілів міст, військових об’єктів або техніки.</p>
      <p>Втомитися – евфемізм, яким часто описують стан російських систем ППО або техніки
після удару ЗСУ.</p>
      <p>Пташка – безпілотник або літальний апарат, який виконує розвідувальні чи ударні
завдання.</p>
      <p>Ціль – об’єкт, по якому планується завдати удару (наприклад, військова техніка,
командний пункт, склад боєприпасів).</p>
      <p>Спеціальна воєнна операція – евфемістичний термін, який росія використовує для
позначення свого повномасштабного вторгнення в Україну з метою уникнення слова
«війна».</p>
      <p>Приземлити – збити ворожий літак, безпілотник чи ракету.</p>
      <p>Мопед – іронічна назва іранського дрона-камікадзе «Shahed», який використовується
для ударів по українській інфраструктурі (через характерний звук двигуна, схожий на
мопед).</p>
      <p>Батальйон Монако – саркастичний термін для українських багатіїв та політиків, які
втекли за кордон під час війни, особливо в дорогі курортні місця на кшталт Монако.</p>
      <p>Дискотека – масований обстріл або бомбардування, часто супроводжується вибухами та
загравою.</p>
      <p>Зоряні війни – протиповітряний бій із застосуванням ППО, коли в небі видно сліди від
збитих ракет або дронів.</p>
      <p>За рускім кораблем – скорочена форма українського військового мему «Русскій корабль,
іді *!», що став символом спротиву російській агресії.</p>
      <p>Дружній вогонь – випадковий обстріл своїх військ або техніки, часто через погану
координацію або паніку.</p>
      <p>На концерт до Кобзона – евфемізм, який означає загибель російських військових чи
командирів (Йосип Кобзон – радянський співак, що підтримував російську агресію, помер у
2018 році).</p>
      <p>Мінусувати – знищувати ворожу техніку або живу силу (наприклад, «мінуснули танк» –
знищили танк).</p>
      <p>М’ясо – мобілізовані солдати, яких російське командування кидає в бій без належної
підготовки та забезпечення (також відоме як «м’ясні штурми»).</p>
      <p>Відпрацювати – завдати удару по ворожій позиції або техніці (наприклад, «артилерія
відпрацювала по складу БК»).</p>
      <p>Нуль – передова лінія фронту, найнебезпечніше місце, де тривають активні бойові дії.
На щиті – вираз, що означає загибель військового у бою. Походить із давньої традиції,
коли загиблих воїнів приносили з поля бою на щитах. У сучасному контексті
використовується як синонім терміна «двохсотий».</p>
      <sec id="sec-10-1">
        <title>Appendix B. The list of euphemisms for Prompt 4</title>
        <p>Бавовна, двохсотий, приліт, трьохсотий, прилетіти, втомитися, пташка, ціль, спеціальна
воєнна операція, на щиті, приземлити, мопед, Батальон Монако, дискотека, зоряні війни, за
рускім кораблем, дружній вогонь, на концерт до Кобзона, мінусувати, м’ясо, відпрацювати,
нуль.</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>S.</given-names>
            <surname>Pinker</surname>
          </string-name>
          ,
          <article-title>The Blank Slate: The Modern Denial of Human Nature</article-title>
          . Viking,
          <year>2003</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>G.</given-names>
            <surname>Lakoff</surname>
          </string-name>
          , Women, Fire, and
          <string-name>
            <given-names>Dangerous</given-names>
            <surname>Things</surname>
          </string-name>
          .
          <source>What Categories Reveal about the Mind</source>
          , University of Chicago Press, Chicago,
          <year>1987</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>C.</given-names>
            <surname>Felt</surname>
          </string-name>
          , E. Riloff,
          <article-title>Recognizing euphemisms and dysphemisms using sentiment analysis</article-title>
          ,
          <source>in: Proceedings of the Second Workshop on Figurative Language Processing</source>
          , pp.
          <fpage>136</fpage>
          -
          <lpage>145</lpage>
          , Online,
          <year>July 2020</year>
          ,
          <article-title>Association for Computational Linguistics</article-title>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2020</year>
          .figlang-
          <volume>1</volume>
          .
          <fpage>20</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>W.</given-names>
            <surname>Zhu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Gong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Bansal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Weinberg</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Christin</surname>
          </string-name>
          , G. Fanti,
          <string-name>
            <given-names>S.</given-names>
            <surname>Bhat</surname>
          </string-name>
          ,
          <article-title>Self-supervised euphemism detection and identification for content moderation</article-title>
          ,
          <source>in: 42nd IEEE Symposium on Security &amp; Privacy</source>
          ,
          <year>2021</year>
          , arXiv preprint arXiv:
          <volume>2103</volume>
          .16808. doi:
          <volume>10</volume>
          .48550/arXiv.2103.16808
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>M.</given-names>
            <surname>Gavidia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Feldman</surname>
          </string-name>
          ,
          <string-name>
            <surname>J. Peng.</surname>
          </string-name>
          <article-title>CATs are fuzzy PETs: A corpus and analysis of potentially euphemistic terms</article-title>
          ,
          <source>in: Proceedings of the Thirteenth Language Resources and Evaluation Conference</source>
          , pp.
          <fpage>2658</fpage>
          -
          <lpage>2671</lpage>
          , Marseille, France,
          <year>June 2022</year>
          . European Language Resources Association.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>P.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Gavidia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Feldman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Peng</surname>
          </string-name>
          ,
          <article-title>Searching for PETs: Using distributional and sentiment-based methods to find potentially euphemistic terms</article-title>
          ,
          <source>in: Proceedings of the Second Workshop on Understanding Implicit and Underspecified Language</source>
          , pp.
          <fpage>22</fpage>
          -
          <lpage>32</lpage>
          , Seattle, USA,
          <year>July 2022</year>
          .
          <article-title>Association for Computational Linguistics</article-title>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2022</year>
          .unimplicit-
          <volume>1</volume>
          .
          <fpage>4</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>P.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Feldman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Peng</surname>
          </string-name>
          ,
          <article-title>A report on the euphemisms detection shared task</article-title>
          ,
          <source>in: Proceedings of the 3rd Workshop on Figurative Language Processing (FLP)</source>
          , pp.
          <fpage>184</fpage>
          -
          <lpage>190</lpage>
          ,
          <string-name>
            <surname>Abu</surname>
            <given-names>Dhabi</given-names>
          </string-name>
          , United Arab Emirates (Hybrid),
          <year>December 2022</year>
          .
          <article-title>Association for Computational Linguistics</article-title>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2022</year>
          .flp-
          <volume>1</volume>
          .
          <fpage>27</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>S. S.</given-names>
            <surname>Keh</surname>
          </string-name>
          ,
          <article-title>Exploring euphemism detection in few-shot and zero-shot settings</article-title>
          ,
          <source>in: Proceedings of the 3rd Workshop on Figurative Language Processing (FLP)</source>
          , pp.
          <fpage>167</fpage>
          -
          <lpage>172</lpage>
          ,
          <string-name>
            <surname>Abu</surname>
            <given-names>Dhabi</given-names>
          </string-name>
          , United Arab Emirates (Hybrid),
          <year>December 2022</year>
          .
          <article-title>Association for Computational Linguistics</article-title>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2022</year>
          .flp-
          <volume>1</volume>
          .
          <fpage>24</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>P.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Shode</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. Chirino</given-names>
            <surname>Trujillo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O. E.</given-names>
            <surname>Ojo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. Cuevas</given-names>
            <surname>Plancarte</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Feldman</surname>
          </string-name>
          , J. Peng, FEED PETs:
          <article-title>Further Experimentation and Expansion on the Disambiguation of Potentially Euphemistic Terms</article-title>
          ,
          <source>in: Proceedings of the 12th Joint Conference on Lexical and Computational Semantics (*SEM</source>
          <year>2023</year>
          ), pp.
          <fpage>437</fpage>
          -
          <lpage>448</lpage>
          , Toronto, Canada,
          <year>July 2023</year>
          .
          <article-title>Association for Computational Linguistics</article-title>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2023</year>
          .starsem1.
          <fpage>38</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>R.</given-names>
            <surname>Choenni</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Garrette</surname>
          </string-name>
          ,
          <string-name>
            <surname>E. Shutova.</surname>
          </string-name>
          <article-title>How do languages influence each other? Studying crosslingual data sharing during LLM fine-tuning</article-title>
          ,
          <source>in: Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing</source>
          , pp.
          <fpage>13244</fpage>
          -
          <lpage>13257</lpage>
          , Singapore,
          <year>December 2023</year>
          .
          <article-title>Association for Computational Linguistics</article-title>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2023</year>
          .emnlp-main.
          <fpage>818</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>I.</given-names>
            <surname>Shode</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. Ifeoluwa</given-names>
            <surname>Adelani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Peng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Feldman</surname>
          </string-name>
          , Nollysenti:
          <article-title>Leveraging transfer learning and machine translation for Nigerian movie sentiment classification, in: Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics</article-title>
          (Volume
          <volume>2</volume>
          :
          <string-name>
            <surname>Short</surname>
            <given-names>Papers)</given-names>
          </string-name>
          , May
          <year>2023</year>
          .
          <article-title>Association for Computational Linguistics</article-title>
          .
          <source>doi: 10.48550/arXiv.2305.10971</source>
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>T.</given-names>
            <surname>Firsich</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Rios</surname>
          </string-name>
          ,
          <source>Can GPT-4 Detect Euphemisms across Multiple Languages? in: Proceedings of the 4th Workshop on Figurative Language Processing (FigLang)</source>
          , June 21,
          <year>2024</year>
          , pp.
          <fpage>65</fpage>
          -
          <lpage>72</lpage>
          .
          <article-title>Association for Computational Linguistics</article-title>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2024</year>
          .figlang-
          <volume>1</volume>
          .
          <fpage>9</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>O.</given-names>
            <surname>Levchenko</surname>
          </string-name>
          ,
          <article-title>Specifics of Ukrainian military discourse</article-title>
          , in: Ucrainica X, Vydala Univerzita Palackého v Olomouci, Olomouc, Czechia,
          <year>2023</year>
          , pp.
          <fpage>199</fpage>
          -
          <lpage>207</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>V.</given-names>
            <surname>Balazh</surname>
          </string-name>
          ,
          <article-title>Pragmalinguistic aspects of the research into euphemisms and dysphemisms (based on Ukrainian news Telegram channels</article-title>
          ),
          <source>New Philology</source>
          ,
          <volume>89</volume>
          (
          <year>2023</year>
          )
          <fpage>35</fpage>
          -
          <lpage>42</lpage>
          . doi:
          <volume>10</volume>
          .26661/
          <fpage>2414</fpage>
          - 1135-2023-89-5
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>O.</given-names>
            <surname>Kharchenko</surname>
          </string-name>
          ,
          <article-title>Ukrainian metaphorical euphemisms during the Russian-Ukrainian war, Printing Horizon, National Technical University of Ukraine "Kyiv Polytechnic Institute named after Igor Sikorsky"</article-title>
          , Кyiv, Ukraine,
          <volume>2</volume>
          /14 (
          <year>2023</year>
          )
          <fpage>90</fpage>
          -
          <lpage>101</lpage>
          . doi:
          <volume>10</volume>
          .20535/
          <fpage>2522</fpage>
          -
          <lpage>1078</lpage>
          .
          <year>2023</year>
          .
          <volume>2</volume>
          (
          <issue>14</issue>
          ).
          <fpage>295247</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>O.</given-names>
            <surname>Taranenko</surname>
          </string-name>
          ,
          <article-title>Euphemization in the Ukrainian media discourse of the hybrid war period</article-title>
          ,
          <source>Social communication: theory and practice</source>
          ,
          <volume>4</volume>
          (
          <year>2017</year>
          )
          <fpage>19</fpage>
          -
          <lpage>27</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>D.</given-names>
            <surname>Stetsenko</surname>
          </string-name>
          ,
          <article-title>When a Language Question Is at Stake. A Revisited Approach to Label Sensitive Content</article-title>
          ,
          <source>arXiv:2311.10514v1 [cs.CL], November</source>
          <volume>17</volume>
          ,
          <year>2023</year>
          . URL: https://arxiv.org/abs/2311.10514
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>W.</given-names>
            <surname>Kieraś</surname>
          </string-name>
          , Ł. Kobyliński,
          <string-name>
            <given-names>D.</given-names>
            <surname>Komosińska</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Nitoń</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Rudolf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Shvedova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Zwierzchowska</surname>
          </string-name>
          , PAWUK:
          <article-title>Polish Automatic Web corpus of UKrainian language</article-title>
          ,
          <source>Instytut Podstaw Informatyki PAN</source>
          ,
          <year>Warszawa 2023</year>
          . URL: https://pawuk.ipipan.waw.pl.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>