<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>A Mandarin-Cantonese Parallel Corpus with Formality Ranking</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>John S. Y. Lee</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Qiong Wang</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Linguistics and Translation, City University of Hong Kong</institution>
        </aff>
      </contrib-group>
      <abstract>
        <p>Formality-controlled machine translation allows users to specify the formality level of the target sentence, so that it would be suitable for the intended audience. While formality-annotated datasets have been constructed for some major languages, no such resource is currently available for Cantonese. This paper presents a Mandarin-Cantonese parallel corpus with 300 Mandarin sentences, each of which is aligned to a list of five or more Cantonese sentences ranked according to their level of formality. To our knowledge, this is the first parallel translation corpus with manual formality ranking, which provides more nuanced judgment than the formal/informal dichotomy in most current formality-annotated datasets. This corpus can support future research towards more fine-grained notions of formality in terminology, translation and text style transfer.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;formality ranking</kwd>
        <kwd>formality-controlled machine translation</kwd>
        <kwd>parallel corpus</kwd>
        <kwd>Large Language Models</kwd>
        <kwd>Cantonese</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>It is important for a translated text to have a level of formality that is appropriate for the target audience.
Consider the following sentences, ordered from the most formal to the least:</p>
      <sec id="sec-1-1">
        <title>1. According to the staf, the lavatory was defective.</title>
        <p>2. We were informed that the lavatory was not functioning.
3. They told us that the toilet wasn’t working.</p>
        <p>4. The loo’s broken, that’s what those guys said.</p>
        <p>
          Sentence (1) would be appropriate for formal communication, for example reports and public speeches,
while sentence (4) may be suitable for casual communication such as everyday conversations. Between
these two extremes of the continuum of formality [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ], sentence (2) or (3) could be the preferred choice
in other contexts, such as polite conversations or social media posts.
        </p>
        <p>
          Formality-controlled machine translation (FCMT) allows users to specify the formality level of
the translation output [
          <xref ref-type="bibr" rid="ref2 ref3">2, 3</xref>
          ]. Since it is expensive to collect parallel bilingual data that include both
formal and informal target sentences, FCMT can be challenging for low-resource languages. Recent
initiatives, such as the Special Task on Formality Control for Spoken Language Translation, have
provided formality-annotated datasets for some major languages [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ]. No such resource has yet been
developed for Cantonese, a variety of Chinese that has 85 million speakers worldwide [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ]. With its lack
of standard written forms, Cantonese expresses nuanced diferences in formality using a considerable
number of linguistic devices, including newly-coined Chinese characters, code-switching with English,
and an elaborate system of sentence-final particles (SFPs). Table 1 shows paraphrases of the same
sentence across the spectrum of formality through vocabulary choices and SFP usage.
        </p>
        <p>
          This paper presents a Mandarin-Cantonese parallel corpus in which each Mandarin sentence is
aligned to multiple formal and informal Cantonese sentences. Notably, these Cantonese sentences are
manually ranked according to their level of formality. Most current language resources for FCMT and
Formality Style Transfer (FST) [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] adopt the formal/informal dichotomy, which can hardly reflect the
diversity of linguistic contexts in the continuum of formality [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ]. To our knowledge, this is the first
        </p>
        <sec id="sec-1-1-1">
          <title>Language Example sentence</title>
          <p>Mandarin 不需要害怕癌症
‘(You) do not need to be
afraid of cancer’
↑ More 唔需要害怕癌症
↑ formal ‘(You) don’t need to be
afraid of cancer’
唔洗怕癌症
‘Don’t worry about
cancer’
Cantonese 唔洗驚癌症
‘Don’t get scared by
cancer’
癌症使乜驚
‘Cancer? Why get
scared?’
↓ More Cancer 咋嘛使乜驚呀
↓ informal ‘Cancer my foot! Why
get scared?’
(1)
(2)
(3)
(4)
(5)
不需要 害怕
buxuyao haipa
‘do not need’ ‘afraid’
唔需要 害怕
m4 seoi1 jiu3 hoi6 paa3</p>
          <p>Lexical diferences
癌症
aizheng
‘cancer’
癌症
ngaam4 zing3
唔洗
m4 sai2
唔洗
m4 sai2
使乜
sai2 mat1
使乜
sai2 mat1
怕
paa3
驚
geng1
驚
geng1
驚
geng1
(No SFP)
(No SFP)
(No SFP)
(No SFP)
(No SFP)
癌症
ngaam4 zing3
癌症
ngaam4 zing3
癌症
ngaam4 zing3
Cancer 咋嘛; 呀
ngaam4 zing3 zaa3 maa3; aa3
(SFP)
attempt to annotate a parallel corpus with formality ranking, which can support research towards more
nuanced gradations of formality.</p>
          <p>The contribution of this paper is two-fold. First, we have constructed a parallel corpus with 300
Mandarin sentences and over 2000 Cantonese paraphrases with formality ranking. This corpus can
facilitate more fine-grained FCMT and FST, as well as the development of formality-annotated lexica.
Second, we describe and evaluate the use of Large Language Models (LLMs) for semi-automatic
corpus construction, which can inform future eforts in creating similar corpora for other low-resource
languages.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>2. Research Background</title>
      <p>
        2.1. Linguistic resources
Parallel corpora with formality-annotated target sentences have been constructed for a number of major
languages. For example, the CoCoA-MT dataset [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], adopted by the Special Task on Formality Control
for Spoken Language Translation, provides formal and informal translations from English into German,
Spanish, Hindi, and Japanese. With the exception of the Japanese subset, which makes a three-way
distinction (“formal”, “polite”, “informal”), all other languages have a binary annotation of formality
only. Since the formal/informal dichotomy can hardly reflect the continuum of formality [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], there may
be many linguistic contexts that call for translations between these two options.
      </p>
      <p>
        Monolingual corpora of formal-informal sentence pairs are available for English, Brazilian Portuguese,
French, and Italian [
        <xref ref-type="bibr" rid="ref6 ref7">6, 7</xref>
        ]. A more fine-grained, 7-point Likert scale on formality has been applied in
annotating English [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] and German sentences of a variety of genres [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. Since these datasets do not
contain paraphrases of the same sentence at diferent formality levels, they could not be used directly
in evaluating formality ranking of candidate translations.
      </p>
      <p>
        As the target language in this paper, Cantonese is the “most widely known and influential variety
of Chinese other than Mandarin” [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], which is generally considered to be standard Chinese. Though
Cantonese and Mandarin share similar writing systems and many cognates, they are mutually
unintelligible in their spoken form. Despite its 85 million speakers worldwide, Cantonese is a low-resource
language [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. With its lack of standard written forms, Cantonese expresses nuanced diferences in
formality using a considerable number of linguistic devices, including newly-coined Chinese characters,
code-switching with English, and an elaborate system of sentence-final particles (SFPs). Current
Cantonese resources include monolingual corpora (e.g., [11]), English-Cantonese parallel corpora (e.g., [12])
and Mandarin-Cantonese parallel corpora (e.g., [13, 14, 15]). Since none has been annotated according
to formality, they cannot support development of formality-controlled machine translation or formality
style transfer. Table 1 shows paraphrases of the same sentence across the spectrum of formality through
lexical diferences (e.g., from formal to informal, 唔需要 m4 seoi1 jiu3, 唔洗 m4 sai2, 使乜 sai2 mat1)
and SFP usage (e.g., 咋嘛 zaa3 maa3).
2.2. Formality-controlled machine translation (FCMT)
In a study on Mandarin-to-Cantonese FCMT, Wong and Lee [14] proposed a rule-based system to
generate low-register and high-register output based on dictionary look-up and syntactic transformation.
In a human evaluation, 62% of the output sentences were judged as satisfactory. Other FCMT approaches
include the use of a “side constraint” to facilitate generation at diferent levels of formality [ 16, 17],
for example by placing tags in front of the input sentence [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. Pre-trained language models, such as
mT5-large and mBART-large, have been fine-tuned to translate English into six languages, in both
informal style and formal style [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. Recent advances in AI have led to Large Language Models (LLMs)
that ofer superior performance in many NLP tasks. Zero-shot and one-shot prompting of LLMs have
been demonstrated to produce high-quality translations with appropriate levels of formality [18]. Due
to the lack of formality-annotated corpus, however, FCMT studies typically rely on manual evaluation
or predictions of formality classifiers or regression models [19].
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Dataset</title>
      <p>Our Mandarin-Cantonese parallel corpus was constructed out of 300 Mandarin-Cantonese sentence
pairs. These sentence pairs were drawn from a parallel corpus of Mandarin and Cantonese [13], which
contains the Mandarin subtitles and Cantonese speech transcriptions of television programs broadcast
in Hong Kong; and from a parallel treebank of Mandarin1 [20] and Cantonese2 [21], annotated with
Universal Dependencies, based on Mandarin subtitles and Cantonese speech transcriptions of short
iflms produced by undergraduate students in Hong Kong. The average sentence length is 13.7 characters
in Mandarin sentences and 14.5 in Cantonese sentences.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Corpus Construction</title>
      <p>
        We adopted a two-stage process to expedite corpus construction. In the first stage (Section 4.1),
translation drafts were automatically generated with Large Language Models (LLMs). Subsequently,
these drafts were annotated and edited by human judges (Section 4.2).
4.1. Generation of translation drafts
In order to generate translations with more diverse levels of formality, we implemented two FCMT
approaches. The “Direct” method prompts an LLM to directly generate formal and informal translations.
In contrast, the “Pipeline” method first performs formality-agnostic MT, and then applies Formality
Style Transfer (FST) [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] to modify the level of formality of the MT output.
      </p>
      <p>Direct method This method uses few-shot prompting to directly generate formal and informal
target sentences with an LLM. The FCMT prompt (Table 2) incorporates ten example
MandarinCantonese sentence pairs, consisting of five formal and five informal samples. To generate formal</p>
      <sec id="sec-4-1">
        <title>1https://universaldependencies.org/treebanks/zh_hk/index.html 2https://universaldependencies.org/treebanks/yue_hk/index.html</title>
        <p>Pipeline method This method performs formality-agnostic MT, and then revises the MT output to
the desired formality level via FST. For formality-agnostic MT, the &lt;formality&gt; in the FCMT
prompt (Table 2) is substituted with “Cantonese” (粵語), i.e., without any formality specification.
For FST, the &lt;formality&gt; in the prompt (Table 2) is substituted with “formal Cantonese” to
generate formal output, and “informal Cantonese” to generate informal output. The FST prompt
uses the chain-of-thought strategy, which has been shown to produce higher-quality output than
a number of competitive baselines [22].</p>
        <p>We used GPT-4o, accessed through the Azure OpenAI Library, to implement the two methods
above. For each of the 300 Mandarin sentences in our dataset (Section 3), the Direct method
generated a formal output (henceforth, Direct-formal) and an informal (Direct-informal) output;
the Pipeline method also generated a formal output (Pipeline-formal) and an informal output
(Pipeline-informal). The sentence lengths were similar across these four output categories, with
average length ranging from 14.1 to 15.0 characters. Through this procedure, each Mandarin sentence
was aligned with five Cantonese paraphrases: four produced by GPT-4o, along with the Cantonese
speech transcript (henceforth, Transcript) harvested from the original dataset (Section 3).</p>
        <sec id="sec-4-1-1">
          <title>Method</title>
          <p>Transcript
Direct-informal
Pipeline-informal
Direct-formal
Pipeline-formal
4.2. Annotation
We recruited 9 undergraduate students, all native speakers of Cantonese, to judge the degree of formality
and quality of these Cantonese sentences. They were shown the Mandarin sentence, followed by the
ifve Cantonese sentences in random order. Each output was independently annotated by two of the
judges. They were instructed to perform the following two tasks:
Formality ranking Rank the five candidate translations from 1 (most formal) to 5 (most informal),
without ties. An example is provided in Table 1.</p>
          <p>Revision Determine if the candidate translation is acceptable. If not, edit the translation while
preserving its level of formality.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Results</title>
      <p>
        5.1. Inter-annotator agreement
In terms of formality ranking, the annotators attained a Kendall’s  of 0.6579. This level of correlation in
ranking compares favorably with those in previous studies [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. In terms of revision, we measured how
often two annotators agreed on the need (or lack thereof) to edit the translation. Depending on one’s
strictness, it could be subjective to draw the line between acceptable and unacceptable translations.
The annotators achieved a Kappa of 0.4116, which corresponds to ‘moderate’ level of agreement [23].
5.2. Analysis
Table 3 reports the average formality ranking of sentences obtained with diferent methods, and the
proportion of Cantonese sentences that were deemed incorrect and therefore required revision. The
larger the value of the average rank, the more informal the sentence.
      </p>
      <p>Informal output. The Cantonese speech transcripts were considered to be most informal (average
rank 4.44 out of 5). A substantial proportion (24.8%) of these transcripts were revised, mostly due to
discrepancies in content with the Mandarin subtitles. The subtitles did not always include all details in
the speech, perhaps due to constraints over screen size.</p>
      <p>The Direct and Pipeline methods were both successful in producing informal Cantonese. The output
of the Direct method was judged to be slightly more informal (rank 3.63) than that of the Pipeline
(rank 3.30). The fact that only 23.7% to 26.8% of the output was edited suggests that these automatic
methods do not necessarily require more manual revision efort than the transcripts. However, we
found that LLM-generated content was less likely than the transcripts to include English code-switching
and appropriate sentence-final particles. Their presence, which leads to more colloquial and engaging
text, was often favored by judges for the informal style.</p>
      <p>Formal output. The Pipeline method produced more formal output (rank 1.20) than the Direct
method (rank 2.35). However, it required significantly more intervention, with 79.7% of the output
edited by the judges. Our post-hoc analysis showed that this was mainly due to an overuse of Mandarin
terms, making the sentence too formal to sound natural in Cantonese. Overall, these results suggest
that LLMs could be efective in producing high-quality drafts for semi-automatic corpus construction.</p>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusions</title>
      <p>We have presented the first Mandarin-Cantonese parallel corpus with formality ranking. For each
Mandarin sentence, at least five Cantonese paraphrases are manually ranked according to their degree of
formality. Most current approaches in formality-controlled machine translation (FCMT) and formality
style transfer (FST) adopt the formal/informal dichotomy. This corpus can support future FCMT and FST
research in exploring more fine-grained notions of formality in terminology and translation. Further,
we have reported our experience with an LLM-based, semi-automatic approach for corpus construction.
Evaluation results suggest that this approach is feasible, and may be considered for future development
of formality-ranked corpora and lexica for other low-resource languages.</p>
    </sec>
    <sec id="sec-7">
      <title>Acknowledgments</title>
      <p>This work is partially supported by a Strategic Research Grant (project number 70006037) from City
University of Hong Kong.</p>
    </sec>
    <sec id="sec-8">
      <title>Declaration on Generative AI</title>
      <sec id="sec-8-1">
        <title>The author(s) have not employed any Generative AI tools.</title>
        <p>[11] O. Y. Kwong, Probing a Two-Way Parallel T&amp;I Corpus for the Lexical Choices of Translators
and Interpreters, New Perspectives on Corpus Translation Studies, New Frontiers in Translation
Studies (2021) 101–132.
[12] V. Yip, S. Matthews, The bilingual child: Early development and language contact, Cambridge</p>
        <p>University Press, 2007.
[13] J. S. Y. Lee, Toward a Parallel Corpus of Spoken Cantonese and Written Chinese, in: Proc. 5th</p>
        <p>International Joint Conference on Natural Language Processing (IJCNLP), 2011, p. 1462–1466.
[14] T.-S. Wong, J. S. Y. Lee, Register-sensitive Translation: A Case Study of Mandarin and Cantonese,
in: Proc. Association for Machine Translation in the Americas (AMTA), 2018.
[15] H. Y. Mak, T. Lee, Low-Resource NMT: A Case Study on the Written and Spoken Languages
in Hong Kong, in: Proc. 5th International Conference on Natural Language Processing and
Information Retrieval (NLPIR), 2021.
[16] Y. Wang, J. Zhang, F. Zhai, J. Xu, C. Zong, Three Strategies to Improve One-to-Many Multilingual</p>
        <p>Translation, in: Proc. EMNLP, 2018.
[17] G. Lample, S. Subramanian, E. Smith, L. Denoyer, M. Ranzato, Y. L. Boureau, Multiple-attribute</p>
        <p>Text Rewriting, in: Proc. ICLR, 2019.
[18] E. Marrese-Taylor, P. C. Wang, Y. Matsuo, Towards Better Evaluation for Formality-Controlled</p>
        <p>English-Japanese Machine Translation, in: Proc. 8th Conference on Machine Translation, 2023.
[19] E. Briakou, S. Agrawal, J. Tetreault, M. Carpuat, Evaluating the Evaluation Metrics for Style
Transfer: A Case Study in Multilingual Formality Transfer, in: Proc. 2021 Conference on Empirical
Methods in Natural Language Processing, 2021.
[20] H. Leung, R. Poiret, T. sum Wong, X. Chen, K. Gerdes, J. Lee, Developing Universal Dependencies
for Mandarin Chinese, in: Proc. Workshop on Asian Language Resources, 2016.
[21] T.-S. Wong, K. Gerdes, H. Leung, J. S. Y. Lee, Quantitative Comparative Syntax on the
CantoneseMandarin Parallel Dependency Treebank, in: Proc. 4th International Conference on Dependency
Linguistics (Depling), 2017, pp. 266–275.
[22] C. Zhang, H. Cai, Y. Li, Y. Wu, L. Hou, M. Abdul-Mageed, Distilling Text Style Transfer With
Self-Explanation From LLMs, in: Proc. 2024 Conference of the North American Chapter of the
Association for Computational Linguistics: Human Language Technologies (Volume 4: Student
Research Workshop), 2024.
[23] J. R. Landis, G. G. Koch, The Measurement of Observer Agreement for Categorical Data, Biometrics
33 (1977) 159–174.</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>F.</given-names>
            <surname>Heylighen</surname>
          </string-name>
          ,
          <string-name>
            <surname>J.-M. Dewaele</surname>
          </string-name>
          ,
          <article-title>Formality of language: Definition, measurement and behavioral determinants</article-title>
          ,
          <source>in: Technical report</source>
          , Center “Leo Apostel”, Free University of Brussels, Brussels, Belgium,
          <year>1999</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>X.</given-names>
            <surname>Niu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Rao</surname>
          </string-name>
          ,
          <string-name>
            <surname>M.</surname>
          </string-name>
          <article-title>Carpuat, Multi-Task Neural Models for Translating Between Styles Within and Across Languages</article-title>
          ,
          <source>in: Proc. 27th International Conference on Computational Linguistics (COLING)</source>
          ,
          <year>2018</year>
          , pp.
          <fpage>1008</fpage>
          -
          <lpage>1021</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>E.</given-names>
            <surname>Rippeth</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Agrawal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Carpuat</surname>
          </string-name>
          ,
          <article-title>Controlling Translation Formality Using Pre-trained Multilingual Language Models</article-title>
          ,
          <source>in: 19th International Conference on Spoken Language Translation (IWSLT</source>
          <year>2022</year>
          ),
          <year>2022</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>M.</given-names>
            <surname>Naˇdejde</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Currey</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Hsu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Niu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Federico</surname>
          </string-name>
          , G. Dinu,
          <string-name>
            <surname>CoCoA-MT</surname>
          </string-name>
          :
          <article-title>A Dataset and Benchmark for Contrastive Controlled MT with Application to Formality, in: Findings of the Association for Computational Linguistics: NAACL</article-title>
          <year>2022</year>
          ,
          <year>2022</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>R.</given-names>
            <surname>Xiang</surname>
          </string-name>
          , E. Chersoni,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Pan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <article-title>Cantonese natural language processing in the transformers era: a survey and current challenges, Language Resources and Evaluation (</article-title>
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>S.</given-names>
            <surname>Rao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Tetreault</surname>
          </string-name>
          ,
          <article-title>Dear Sir or Madam, May I Introduce the GYAFC Dataset: Corpus, Benchmarks and Metrics for Formality Style Transfer</article-title>
          , in: Proc.
          <article-title>Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL-HLT</article-title>
          ),
          <year>2018</year>
          , p.
          <fpage>129</fpage>
          -
          <lpage>140</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>E.</given-names>
            <surname>Briakou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Lu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , J. Tetreault, Olá, Bonjour, Salve! XFORMAL:
          <article-title>A Benchmark for Multilingual Formality Style Transfer</article-title>
          , in: Proc.
          <article-title>Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL-HLT</article-title>
          ),
          <year>2021</year>
          , p.
          <fpage>3199</fpage>
          -
          <lpage>3216</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>E.</given-names>
            <surname>Pavlick</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Tetreault</surname>
          </string-name>
          ,
          <article-title>An Empirical Analysis of Formality in Online Communication, Transactions of the Association for Computational Linguistics 4 (</article-title>
          <year>2016</year>
          )
          <fpage>61</fpage>
          -
          <lpage>74</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>E.</given-names>
            <surname>Eder</surname>
          </string-name>
          ,
          <string-name>
            <given-names>U.</given-names>
            <surname>Krieg-Holz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Wiegand</surname>
          </string-name>
          ,
          <article-title>A Question of Style: A Dataset for Analyzing Formality on Diferent Levels, in: Findings of the Association for Computational Linguistics: EACL</article-title>
          <year>2023</year>
          ,
          <year>2023</year>
          , p.
          <fpage>580</fpage>
          -
          <lpage>593</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>S.</given-names>
            <surname>Matthews</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Yip</surname>
          </string-name>
          , Cantonese:
          <string-name>
            <given-names>A Comprehensive</given-names>
            <surname>Grammar</surname>
          </string-name>
          , Routledge, New York,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>