<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">&quot;An bhfuil Gaeilge agat?&quot;: Differences in User Interaction and Assistant Responses Across Languages of European Origin in Large-scale Conversational Datasets</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author role="corresp">
							<persName><forename type="first">Aldan</forename><surname>Creo</surname></persName>
							<email>research@acmc.fyi</email>
							<affiliation key="aff0">
								<orgName type="institution">Independent author</orgName>
								<address>
									<settlement>Dublin</settlement>
									<region>IE</region>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">&quot;An bhfuil Gaeilge agat?&quot;: Differences in User Interaction and Assistant Responses Across Languages of European Origin in Large-scale Conversational Datasets</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">F01B370CD9C6ED62A7EFCE4322EEE296</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T18:14+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Artificial Intelligence</term>
					<term>Natural Language Processing</term>
					<term>Conversational AI</term>
					<term>Multilingualism</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>This study presents a comprehensive analysis of user interactions and assistant responses across 28 European languages using the WildChat and LMSYS datasets, addressing an existing gap in the understanding of multilingual conversational AI. We examine five specific dimensions: the topics discussed, the length of the conversations, the sentiment expressed, the toxicity of the interactions, and the quality of the responses. Our findings indicate significant cross-linguistic variations that have significant implications for the development and deployment of language models. Topic analysis shows a high degree of overlap across languages, indicating that users engage with similar subjects regardless of their linguistic background. We observe a positive correlation between the frequency of language use and conversation length, which suggests that different engagement patterns may be at play across language communities. Sentiment analysis indicates a high degree of consistency in neutral tones across languages, whereas toxicity levels vary considerably, with some languages exhibiting notably elevated scores. To assess response quality, we introduce a custom neural architecture based on the classification of user-assistant interaction triples. Our model achieved an accuracy of 0.82 and served to uncover variations in user satisfaction across language groups. Speakers of Romance languages exhibited higher levels of satisfaction, whereas those of Eastern European languages tended to show lower satisfaction with their interactions with the assistant. Our findings underscore the need for language-specific strategies in conversational AI development, particularly in content moderation, conversation design, and quality assessment. By highlighting the differences and commonalities in conversational interactions across languages, our work provides insights for researchers and developers seeking to better understand and address the needs of users across a diverse linguistic landscape.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>The development of conversational assistants has seen significant advancements in recent years, following the advent of Transformer-based architectures <ref type="bibr" target="#b0">[1]</ref>, as exemplified by the GPT family of models. These assistants are capable of engaging in conversations with users on a wide range of topics, providing information, answering questions, and even engaging in small talk <ref type="bibr" target="#b1">[2]</ref>.</p><p>The popularity of conversational assistants has therefore surged across various applications, including customer service and language learning <ref type="bibr" target="#b2">[3,</ref><ref type="bibr" target="#b3">4]</ref>. Notably, ChatGPT exceeded 100 million monthly active users within its first two months, a testament to the rising interest in conversational AI <ref type="bibr" target="#b4">[5]</ref>. However, despite this booming interest, research and development in the field remain predominantly Englishcentric, with limited focus on other languages <ref type="bibr" target="#b5">[6,</ref><ref type="bibr" target="#b6">7]</ref>.</p><p>A branch of research that is key to the development of conversational assistants is the analysis of user interactions and assistant responses in a large scale. Two main works in this area due to their size and diversity are the WildChat and LMSYS datasets <ref type="bibr" target="#b7">[8,</ref><ref type="bibr" target="#b8">9]</ref>, which contain a large number of user interactions with conversational assistants and are publicly available. However, there is still work to be done in analyzing these datasets, due to their size and complexity.</p><p>This study aims to explore user interactions and assistant responses, focusing on linguistic differences  across European languages. Previous research has only examined datasets as a whole <ref type="bibr" target="#b9">[10]</ref>, without delving into cross-language discrepancies-a gap our study seeks to fill. In contrast, we seek to provide insights into cross-language differences within large-scale conversational datasets, in order to inform the development of assistants that are better tailored to the needs of users of a wider range of languages. Our analysis focuses on language groups rather than the countries in which users reside. We acknowledge that languages like Spanish or Portuguese may have significant representation from Latin American users, contributing to the diversity of our findings. The rest of this paper is structured as follows. The subsequent sections examine whether notable differences exist across languages in terms of topics discussed (RQ1), length of interactions (RQ2), sentiment expressed (RQ3), toxicity (RQ4), and quality (user satisfaction) of responses (RQ5). We then proceed to an integrative discussion and conclusion.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Datasets</head><p>In this section, we describe the datasets used for our analysis.</p><p>We combined the WildChat and LMSYS datasets <ref type="bibr" target="#b7">[8,</ref><ref type="bibr" target="#b8">9]</ref>, which contain 990,372 and 1,000,000 examples, respectively. Nevertheless, the count of examples we analyze is lower, as we applied several preprocessing steps to clean the data. First, we discarded examples where the user's initial message contained fewer than five words, as short inputs hinder the accuracy of language detection. For instance, messages like "Test" were often misclassified, such as being incorrectly identified as Estonian. Furthermore, we excluded interactions where the toxicity scores were not defined, as these annotations are essential for our toxicity analysis.</p><p>To ensure a comprehensive representation of European languages, we included 28 languages, shown in Figure <ref type="figure" target="#fig_0">1</ref>, together with the number of examples in each. However, we excluded languages such as Basque and Breton because of the lack of a sufficient number of examples in the datasets, as well as the lack of language models trained specifically for these languages, which could introduce bias when using them in our analysis. After applying these filtering criteria, we obtained a total of 781,376 examples.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">RQ1: Topics</head><p>In this section, we analyze the distribution of topics across languages. We utilize language tags as a clustering feature for semantically-informed embeddings. We hypothesize that if speakers of different languages tend to discuss distinct topics, language tags should form clusters with clear boundaries. However, if topics are more uniformly distributed across languages, the boundaries will be blurrier.</p><p>We commence our analysis by selecting all initial user messages in the dataset, as these are the most representative, establishing the context for subsequent discourse. We then use two multilingual models from the Sentence Transformers library <ref type="bibr" target="#b10">[11]</ref>-paraphrase-multilingual-mpnet-base-v2 and paraphrase-multilingual-MiniLM-L12-v2-to generate embeddings of these messages.</p><p>To evaluate how well messages cluster based on language, we conduct a silhouette score analysis (ranging from -1 to 1), which measures cluster cohesion and separation. A score of 1 indicates welldefined, separated clusters, while a score of 0 suggests overlap. Negative scores indicate poorly defined or incorrect clusters <ref type="bibr" target="#b11">[12]</ref>.</p><p>Due to the high computational cost of this analysis, which scales quadratically with the number of examples, we performed the silhouette score calculation on 20 randomly-selected subsets, each representing 5% of the dataset, and then averaged the results. The mean silhouette scores for the two models were −0.121(24) -mean(s.d.) -and −0.115(30). These negative scores indicate significant semantic overlap between clusters, confirming that languages are not effective clustering tags.</p><p>Per-language silhouette scores, which align with the overall results, are provided in the supplementary material. Based on this analysis, we conclude that users across linguistic groups engage with similar themes, rather than showing strong language-specific patterns. This finding highlights the universality of topics and suggests that conversational assistants must prioritize general topic coverage and flexibility. Moreover, the lack of clustering implies that cultural or regional nuances may play a smaller role in topic differentiation than previously expected. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">RQ2: Length</head><p>In this section, we explore the correlation between the mean number of words written by users in a conversation and the number of examples per language. As both variables are not normally distributed, we study their interaction by calculating the Spearman Rank correlation coefficient, which ranges from -1 to 1. A value of -1 indicates a perfect negative correlation; 0, no correlation; and 1 signifies a perfect positive correlation. The coefficient evaluates to 0.51, indicating a moderate positive correlation between the number of user words and the number of examples in a language. With a p-value of 0.0055 &lt; 𝛼 = 0.05, the correlation is statistically significant.</p><p>While not a perfect correlation, the strength of the relationship suggests that the number of user words can be considered a reasonably reliable predictor of the number of examples in a language. Figure <ref type="figure" target="#fig_1">2</ref> presents a scatter plot illustrating this relationship. Additionally, Table <ref type="table" target="#tab_1">1</ref> provides the per-conversation mean number of messages, user words, and assistant words for each language.</p><p>In essence, we find that speakers of languages with more interactions in conversational datasets tend to engage in longer conversations, as measured by the number of user words. This trend provides moderate support for the notion that there exist behavioral differences across groups of language users, which may be indicative of cultural or linguistic factors influencing conversation length. However, further research is warranted to ascertain the underlying reasons for these differences.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">RQ3: Sentiment</head><p>In this section, we explore sentiment differences across languages, as it allows to understand the emotional tone of conversations and provides insights into both user experience and interaction quality.</p><p>We use the pre-trained multilingual classification model twitter-XLM-roBERTa-base <ref type="bibr" target="#b12">[13]</ref> for our analysis. This model is particularly suited to our multilingual setting, as it has been trained on a large, diverse corpus of Twitter messages in various languages.</p><p>Our analysis focuses on both the user and the assistant first messages in each conversation, as these set the tone and are likely the most representative of the overall sentiment. We calculate sentiment scores for both messages, then aggregate these scores across languages. We show the results in Table <ref type="table">2</ref>.</p><p>One clear observation is that the assistant's messages tend to be more positive than the users', with an overall mean of 0.2205(15) compared to 0.16 <ref type="bibr" target="#b10">(11)</ref>. This difference likely reflects the Reinforcement</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 2</head><p>Mean and standard deviation for the sentiment scores of the first user (U) and assistant (A) messages. Learning from Human Feedback (RLHF) paradigm used for training <ref type="bibr" target="#b13">[14]</ref>, which encourages the assistant to maintain a more positive and helpful tone. Interestingly, the assistant also shows a slightly higher negative sentiment score (0.28 <ref type="bibr" target="#b19">(20)</ref>) than the users (0.26 <ref type="bibr" target="#b19">(20)</ref>). This could also be attributed to RLHF, which prompts it to refrain from engaging in potentially toxic conversations, thus increasing the frequency of negative sentiment classifications in those contexts <ref type="bibr" target="#b14">[15]</ref>. Overall, however, both user and assistant messages tend to be neutral, with a mean neutral score of 0.50 <ref type="bibr" target="#b17">(18)</ref> and 0.58 <ref type="bibr" target="#b18">(19)</ref>, respectively. These observations show that while there are differences in the sentiment expressed by users and assistants, the sentiment across languages tends to remain fairly neutral.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Language Negative (U) Neutral (U) Positive (U) Negative (A) Neutral (A) Positive</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">RQ4: Toxicity</head><p>We perform a toxicity analysis to identify potential differences, for every message in the dataset (a total of 2,951,678). There exist 11 types of toxicity annotations generated by the dataset creators with the OpenAI Moderations API. To simplify our analysis, we aggregate these categories into two general toxicity scores using the mean and maximum values across the categories. We do this for each example. Then, we calculate the mean of the two across the examples, which we present in Table <ref type="table" target="#tab_3">3</ref>.</p><p>In the rest of this section, we focus on the perspective of the user messages, as they are generally causative of the toxicity in the assistant's side of the conversation. To identify significant differences in toxicity across languages, we employ the Kruskal-Wallis test. This is a non-parametric method well-suited to compare medians of toxicity scores across multiple independent samples, especially given the non-normal distribution of toxicity values, which cluster near the extremes. We test a null hypothesis 𝐻 0 = "The medians of the toxicity scores are equal across languages" with 𝛼 = 0.05. Since toxicity scores are continuous and generally skewed towards values close to 0, directly applying the Kruskal-Wallis test might exaggerate differences between languages due to minor deviations in low-toxicity messages. To mitigate this, we round the scores to two decimal places, reducing the number of unique values and treating very close values as ties. The outcomes of the Kruskal-Wallis test, as well as independent analyses for each toxicity category and aggregated scores, are presented in Table <ref type="table" target="#tab_4">4</ref>.</p><p>With the exception of the "self-harm" category, the p-values for all other categories are less than 𝛼, indicating significant differences in toxicity across languages. This necessitates pairwise comparisons to determine which languages exhibit meaningful disparities. We use Dunn's test for this purpose, with the null hypothesis 𝐻 0 = "The probability that a randomly selected message from one language has a higher toxicity score than one from another language is 0.5. " To control for the increased risk of Type I errors due to multiple comparisons, we apply the Bonferroni correction to the p-values. The results of these pairwise comparisons for the averaged toxicity scores are shown in Figure <ref type="figure" target="#fig_2">3</ref>, while results for other configurations are included in the supplementary materials.</p><p>Key findings from the pairwise comparisons include:</p><p>1. No Statistically Significant Differences: For most languages-such as Bulgarian, Croatian, Czech, Danish, Estonian, Finnish, Galician, Greek, Hungarian, Irish, Italian, Latvian, Lithuanian, Maltese, Norwegian, Polish, Romanian, Slovak, and Slovenian-the comparisons are not statistically significant. In these cases, we cannot reject 𝐻 0 , indicating it is equally likely that a randomly selected message from one of these languages has a higher or lower toxicity score than another. 2. Significant Differences in Specific Languages: Pairwise comparisons involving Dutch, English,  French, German, Spanish, Swedish, and Ukrainian reveal significant differences compared to several languages from the first group (Greek, Italian, Polish, Portuguese, Romanian, Spanish, Swedish and Ukrainian). These languages show distinct toxicity profiles, although notable exceptions exist within this subset. 3. Group of Similar Toxicity: While Dutch, English, French, and German exhibit significant differences compared to other languages, they are not significantly different from one another, suggesting a group of languages with similar, higher-than-average toxicity scores. 4. Highest Average Toxicity: Languages such as Catalan, Dutch, English, German, and Swedish display the highest average toxicity scores, with Dutch, English, German, and Swedish identified  as significantly different in the majority of pairwise comparisons.</p><p>Overall, the results highlight considerable variability in the toxicity of conversations across languages. The distribution of toxicity scores varies significantly between some languages, suggesting that conversational toxicity may be influenced by a range of factors, including cultural backgrounds, the structure of the language itself, potential biases in toxicity tagging, or a combination of these influences.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7.">RQ5: Quality</head><p>In this section, we assess the quality of the assistant's responses, which we define as how well the assistant meets the user's needs and expectations. To the best of our knowledge, no large-scale multilingual dataset exists with labels for assistant response quality, such as those used in RLHF <ref type="bibr" target="#b15">[16]</ref>. This motivated us to develop a custom architecture tailored for evaluating assistant responses in a multilingual setting.</p><p>Our approach is inspired by the siamese architecture proposed in Sentence-BERT <ref type="bibr" target="#b10">[11]</ref>. While the original architecture encodes two inputs and trains a classification head that is later discarded to retain only the fine-tuned encoder, we focus on training the classification head with three inputs to evaluate the quality of assistant responses, the encoder remaining frozen. The classification task asks "Is the user satisfied with the assistant's response?", with classes "Yes, " "No, " and "N/A" (not applicable).</p><p>We structure the model to process ⟨𝑢 1 , 𝑎, 𝑢 2 ⟩ triples, where 𝑢 1 is the user's initial message, 𝑎 is the  assistant's response, and 𝑢 2 is the user's second message. The goal is to capture the semantics of the user-assistant exchange and assess whether the assistant's response satisfies the user's original query.</p><p>For instance, if the user were to ask about the weather, the embedding of their first message may belong to a subspace "weather"; if the assistant were to respond with the weather forecast, the embedding of the assistant's response could also belong to the same subspace. The user, having received a satisfactory response, may then express their satisfaction in their second message, the embedding of which could encode a positive sentiment. We expect the model to be able to capture the distances between the embeddings of the messages (e.g., the assistant's response addressing the user's first message), as well as the semantics of the messages themselves (e.g., the user's satisfaction with the assistant's response). Formally, our model takes a tokenized input of shape 3 × 𝐿, where 𝐿 is the maximum length of the tokenized input (shorter sequences are padded), the three messages being stacked along a new dimension. The model outputs a pooled representation 𝑒(𝑖) for each message in the triple, where 𝑖 ∈ 𝑢 1 , 𝑎, 𝑢 2 , producing embeddings 𝑒(𝑢 1 ), 𝑒(𝑎), and 𝑒(𝑢 2 ) of dimension 𝑑 𝑒 = 768.</p><formula xml:id="formula_0">sentence representation = 𝑒(𝑖) ∈ R 𝑑𝑒 for 𝑖 ∈ {𝑢 1 , 𝑎, 𝑢 2 }<label>(1)</label></formula><p>We concatenate these embeddings with their absolute differences into a single vector 𝑖 𝑐𝑙𝑠 : We pass 𝑖 𝑐𝑙𝑠 through a fully connected feedforward neural network with four layers, 𝑊 𝑛 and 𝑏 𝑛 being the weights and biases of layer 𝑛, each with a hidden size of 884 and Mish <ref type="bibr" target="#b16">[17]</ref> activation functions to obtain the output logits 𝑜 𝑐𝑙𝑠 :</p><formula xml:id="formula_1">𝑖 𝑐𝑙𝑠 =</formula><formula xml:id="formula_2">𝑜 𝑐𝑙𝑠 = 𝑊 4 • Mish(𝑊 3 • Mish(𝑊 2 • Mish(𝑊 1 • 𝑖 𝑐𝑙𝑠 + 𝑏 1 ) + 𝑏 2 ) + 𝑏 3 ) + 𝑏 4 ∈ R 3<label>(3)</label></formula><p>Finally, we apply an argmax operation on 𝑜 𝑐𝑙𝑠 to determine the class prediction. A diagram of this architecture is shown in Figure <ref type="figure" target="#fig_3">4</ref>. We used Optuna <ref type="bibr" target="#b17">[18]</ref> to optimize the hyperparameters of the classification head and training. Table <ref type="table" target="#tab_5">5</ref> shows the search space and best values found.</p><p>For training, we built a dataset of ⟨𝑢 1 , 𝑎, 𝑢 2 ⟩ triples by sampling up to 1000 conversations per language with at least three messages. We manually annotated 1000 examples using the Argilla platform <ref type="bibr" target="#b18">[19]</ref>. We also generated translations for non-English conversations to assist the annotation process using the EuroLLM-1.7B model <ref type="bibr" target="#b19">[20]</ref>. We did not assess the quality of the translations, as they were utilized solely for contextual purposes, thereby facilitating a general comprehension of the discourse. Furthermore, we established explicit guidelines to facilitate a uniform interpretation of the task. Our annotation process maintained reasonable balance across languages, with an average of 37(33) annotations/language. Specific counts and annotation guidelines are provided in the supplementary material.</p><p>We trained the model using cross-entropy loss, optimized with AdamW <ref type="bibr" target="#b20">[21]</ref> for 10 epochs, with the default hyperparameters in the Hugging Face Transformers library <ref type="bibr" target="#b21">[22]</ref>, 𝛽 1 = 0.9, 𝛽 2 = 0.999, and 𝜖 = 1 × 10 −8 . For the encoder, we utilized the paraphrase-multilingual-mpnet-base-v2 pretrained weights <ref type="bibr" target="#b10">[11]</ref>. We reserved 10% of the data for validation and another 10% for testing, performing 10-fold cross-validation. The model achieved a mean test accuracy of 0.820(45), significantly outperforming both a random baseline (0.339(42)) and a majority class classifier (0.594(44)).</p><p>When applying the model to the full dataset (raw numbers are reported in the supplementary material), the results (Figure <ref type="figure" target="#fig_4">5</ref>) reveal notable discrepancies in user satisfaction across linguistic groups. Disregarding languages with a limited number of examples, where the results may lack representativeness, we observe that the satisfaction of users with the assistant's responses is generally high and tends to be higher for users speaking languages belonging to the Romance language family <ref type="bibr" target="#b22">[23]</ref>, including Italian, Portuguese, Spanish, Catalan, French, and Romanian. In contrast, users speaking languages that originated in Eastern Europe, such as Bulgarian, Greek, and Hungarian, exhibit a lower level of satisfaction with the assistant's responses. English-speaking users demonstrate a satisfaction level that is comparable to the overall mean, similar to that observed in German and Polish, although the latter two languages exhibit a higher percentage of unsatisfied users.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="8.">Discussion</head><p>Our analysis of the WildChat and LMSYS datasets reveals significant differences in user-assistant interaction across European languages. These findings contribute to a more nuanced understanding of multilingual conversational AI and highlight the importance of considering linguistic diversity in the development and evaluation of language models.</p><p>The lack of clear clustering by language in our topic analysis (RQ1) suggests that users across different languages engage with conversational assistants on a wide variety topics. This finding is encouraging, as it indicates that the assistants are capable of handling a wide range of subjects across multiple languages. However, it also emphasizes the need for language models to be equally proficient in diverse topics across all supported languages.</p><p>Our analysis of conversation length (RQ2) revealed a positive correlation between the number of user words and the number of examples in a language. While further investigation is needed to understand the underlying factors driving this correlation, it suggests that users in languages with more examples may be more likely to engage in longer conversations, possibly due to a higher level of comfort or familiarity with the conversational AI system or a better performance of the assistant in those languages.</p><p>The sentiment analysis (RQ3) showed that both user and assistant messages tend to be neutral, with assistant responses generally being either more positive or negative. This is a consistent result across languages, and suggests that the current training approaches are effective in maintaining a coherent tone across different linguistic contexts.</p><p>Perhaps our most striking finding relates to toxicity (RQ4). The significant differences in toxicity levels across languages highlight the need for language-specific approaches to content moderation and toxicity detection. This is particularly important for languages like Dutch, English, German, and Swedish, which also exhibited higher average scores.</p><p>Finally, our quality analysis (RQ5) revealed considerable variations in user satisfaction across different language groups. The higher satisfaction levels among Romance language speakers and lower levels among Eastern European language speakers underscore the importance of tailoring conversational AI to specific linguistic and cultural contexts.</p><p>These findings collectively emphasize the need for a more nuanced, language-specific approach to the development and evaluation of conversational AI. While current models show promise in their ability to engage across multiple languages, there is still significant room for improvement in addressing language-specific challenges and user expectations.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="9.">Conclusion</head><p>This study presents the first comprehensive analysis of differences in user interaction and assistant responses across a wide range of European-origin languages for the WildChat and LMSYS datasets. Our work addresses a critical gap in existing literature, mainly treating these datasets as homogeneous.</p><p>By examining topics, conversation length, sentiment, toxicity, and response quality, we have uncovered significant variations across languages that have important implications for the development and deployment of conversational AI systems. Our findings highlight the need for more nuanced, language-specific approaches in areas such as content moderation and quality assessment.</p><p>The insights gained from this study are crucial for ensuring that the perspectives and needs of non-English speakers are adequately represented in the development of conversational AI. As the use of these systems continues to grow globally, it is imperative that they are designed to provide equitable and high-quality experiences across all languages.</p><p>Future work should focus on developing language-specific strategies for improving conversational AI, particularly in areas where we observed significant differences, such as toxicity levels and user satisfaction. Additionally, expanding this analysis to include non-European languages would provide a more comprehensive global perspective on multilingual conversational AI.</p><p>In conclusion, our work seeks to contribute to a more inclusive and effective approach to conversational AI development, providing insight into the importance of linguistic diversity in creating truly global and user-centric AI systems. We hope that these findings will inform and inspire future research and development efforts in multilingual conversational AI, ultimately leading to more equitable and effective language technologies for users worldwide.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Number of examples per language.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: Scatter plot of the mean number of user words and the number of examples, for each language.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: Dunn's test p-values for the user's mean toxicity score. We only color significant cells.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Figure 4 :</head><label>4</label><figDesc>Figure 4: Architecture of our model to evaluate the quality of assistant responses.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_4"><head>Figure 5 :</head><label>5</label><figDesc>Figure 5: Percentage and total counts of satisfaction labels for each language.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_5"><head></head><label></label><figDesc>concatenate(𝑒(𝑢 1 ), 𝑒(𝑎), 𝑒(𝑢 2 ), |𝑒(𝑎) − 𝑒(𝑢 1 )|, |𝑒(𝑢 2 ) − 𝑒(𝑎)|, |𝑒(𝑢 2 ) − 𝑒(𝑢 1 )|) ∈ R 6×𝑒 (2)</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 1</head><label>1</label><figDesc>Mean number of messages, user words, and assistant words per conversation.</figDesc><table><row><cell cols="4">Language Messages User words Assistant words</cell></row><row><cell>Bulgarian</cell><cell>3.38</cell><cell>92.68</cell><cell>158.92</cell></row><row><cell>Catalan</cell><cell>4.29</cell><cell>98.00</cell><cell>254.60</cell></row><row><cell>Croatian</cell><cell>5.78</cell><cell>103.07</cell><cell>248.80</cell></row><row><cell>Czech</cell><cell>3.76</cell><cell>45.89</cell><cell>176.18</cell></row><row><cell>Danish</cell><cell>3.74</cell><cell>83.94</cell><cell>219.82</cell></row><row><cell>Dutch</cell><cell>3.72</cell><cell>88.88</cell><cell>217.99</cell></row><row><cell>English</cell><cell>3.68</cell><cell>89.28</cell><cell>264.70</cell></row><row><cell>Estonian</cell><cell>3.38</cell><cell>44.26</cell><cell>100.77</cell></row><row><cell>Finnish</cell><cell>3.96</cell><cell>57.18</cell><cell>211.57</cell></row><row><cell>French</cell><cell>4.24</cell><cell>92.68</cell><cell>298.27</cell></row><row><cell>Galician</cell><cell>4.71</cell><cell>44.84</cell><cell>340.51</cell></row><row><cell>German</cell><cell>4.31</cell><cell>87.44</cell><cell>272.59</cell></row><row><cell>Greek</cell><cell>3.44</cell><cell>49.90</cell><cell>121.65</cell></row><row><cell>Hungarian</cell><cell>3.19</cell><cell>46.03</cell><cell>128.22</cell></row><row><cell>Irish</cell><cell>3.25</cell><cell>33.69</cell><cell>193.06</cell></row><row><cell>Italian</cell><cell>4.77</cell><cell>66.66</cell><cell>312.91</cell></row><row><cell>Latvian</cell><cell>4.29</cell><cell>61.05</cell><cell>186.12</cell></row><row><cell>Lithuanian</cell><cell>2.14</cell><cell>7.14</cell><cell>88.07</cell></row><row><cell>Maltese</cell><cell>4.00</cell><cell>16.75</cell><cell>163.88</cell></row><row><cell>Norwegian</cell><cell>3.44</cell><cell>51.00</cell><cell>219.81</cell></row><row><cell>Polish</cell><cell>3.83</cell><cell>59.40</cell><cell>171.79</cell></row><row><cell>Portuguese</cell><cell>4.78</cell><cell>83.19</cell><cell>319.67</cell></row><row><cell>Romanian</cell><cell>3.83</cell><cell>76.29</cell><cell>221.79</cell></row><row><cell>Slovak</cell><cell>4.16</cell><cell>67.23</cell><cell>227.99</cell></row><row><cell>Slovenian</cell><cell>3.07</cell><cell>60.65</cell><cell>153.43</cell></row><row><cell>Spanish</cell><cell>4.59</cell><cell>92.25</cell><cell>307.73</cell></row><row><cell>Swedish</cell><cell>4.89</cell><cell>100.37</cell><cell>253.91</cell></row><row><cell>Ukrainian</cell><cell>5.72</cell><cell>109.19</cell><cell>250.05</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>(A)</head><label></label><figDesc></figDesc><table><row><cell>Bulgarian</cell><cell>0.133(85)</cell><cell>0.74(12)</cell><cell>0.128(93)</cell><cell>0.18(16)</cell><cell>0.63(19)</cell><cell>0.19(17)</cell></row><row><cell>Catalan</cell><cell>0.27(20)</cell><cell>0.60(19)</cell><cell>0.13(11)</cell><cell>0.28(19)</cell><cell>0.51(18)</cell><cell>0.2102(18)</cell></row><row><cell>Croatian</cell><cell>0.30(20)</cell><cell>0.55(20)</cell><cell>0.15(13)</cell><cell>0.26(19)</cell><cell>0.53(19)</cell><cell>0.21(13)</cell></row><row><cell>Czech</cell><cell>0.25(18)</cell><cell>0.61(20)</cell><cell>0.14(11)</cell><cell>0.23(16)</cell><cell>0.57(18)</cell><cell>0.200(15)</cell></row><row><cell>Danish</cell><cell>0.21(16)</cell><cell>0.63(17)</cell><cell>0.16(12)</cell><cell>0.25(18)</cell><cell>0.52(17)</cell><cell>0.22(15)</cell></row><row><cell>Dutch</cell><cell>0.24(21)</cell><cell>0.64(22)</cell><cell>0.13(11)</cell><cell>0.25(20)</cell><cell>0.569(21)</cell><cell>0.18(17)</cell></row><row><cell>English</cell><cell>0.26(20)</cell><cell>0.58(19)</cell><cell>0.16(11)</cell><cell>0.28(20)</cell><cell>0.49(17)</cell><cell>0.22(15)</cell></row><row><cell>Estonian</cell><cell>0.31(14)</cell><cell>0.45(18)</cell><cell>0.24(12)</cell><cell>0.24(11)</cell><cell>0.45(18)</cell><cell>0.306(20)</cell></row><row><cell>Finnish</cell><cell>0.34(25)</cell><cell>0.51(22)</cell><cell>0.16(14)</cell><cell>0.30(20)</cell><cell>0.483(17)</cell><cell>0.22(16)</cell></row><row><cell>French</cell><cell>0.27(21)</cell><cell>0.481(20)</cell><cell>0.25(17)</cell><cell>0.28(19)</cell><cell>0.410(17)</cell><cell>0.311(18)</cell></row><row><cell>Galician</cell><cell>0.20(19)</cell><cell>0.68(19)</cell><cell>0.121(92)</cell><cell>0.23(15)</cell><cell>0.56(16)</cell><cell>0.21(14)</cell></row><row><cell>German</cell><cell>0.22(18)</cell><cell>0.68(20)</cell><cell>0.11(12)</cell><cell>0.2150(17)</cell><cell>0.63(20)</cell><cell>0.151(17)</cell></row><row><cell>Greek</cell><cell>0.26(15)</cell><cell>0.55(16)</cell><cell>0.19(13)</cell><cell>0.24(14)</cell><cell>0.53(14)</cell><cell>0.24(14)</cell></row><row><cell cols="2">Hungarian 0.20(14)</cell><cell>0.66(18)</cell><cell>0.14(12)</cell><cell>0.20(18)</cell><cell>0.60(20)</cell><cell>0.20(16)</cell></row><row><cell>Irish</cell><cell>0.284(18)</cell><cell>0.53(21)</cell><cell>0.19(18)</cell><cell>0.372(31)</cell><cell>0.394(23)</cell><cell>0.23(25)</cell></row><row><cell>Italian</cell><cell>0.28(23)</cell><cell>0.59(23)</cell><cell>0.14(14)</cell><cell>0.28(22)</cell><cell>0.53(21)</cell><cell>0.19(17)</cell></row><row><cell>Latvian</cell><cell>0.22(15)</cell><cell>0.67(14)</cell><cell>0.106(86)</cell><cell>0.28(22)</cell><cell>0.51(21)</cell><cell>0.20(17)</cell></row><row><cell cols="2">Lithuanian 0.22(11)</cell><cell>0.71(11)</cell><cell>0.070(30)</cell><cell>0.244(79)</cell><cell>0.59(13)</cell><cell>0.165(84)</cell></row><row><cell>Maltese</cell><cell>0.26(16)</cell><cell>0.58(22)</cell><cell>0.16(12)</cell><cell>0.16(11)</cell><cell>0.58(17)</cell><cell>0.25(18)</cell></row><row><cell cols="2">Norwegian 0.194(17)</cell><cell>0.65(19)</cell><cell>0.15(14)</cell><cell>0.220(18)</cell><cell>0.51(18)</cell><cell>0.27(19)</cell></row><row><cell>Polish</cell><cell>0.19(17)</cell><cell>0.70(19)</cell><cell>0.116(11)</cell><cell>0.193(18)</cell><cell>0.63(20)</cell><cell>0.18(16)</cell></row><row><cell cols="2">Portuguese 0.190(17)</cell><cell>0.67(18)</cell><cell>0.142(13)</cell><cell>0.20(16)</cell><cell>0.60(17)</cell><cell>0.19(15)</cell></row><row><cell>Romanian</cell><cell>0.21(18)</cell><cell>0.66(19)</cell><cell>0.13(12)</cell><cell>0.196(17)</cell><cell>0.60(19)</cell><cell>0.21(17)</cell></row><row><cell>Slovak</cell><cell>0.25119(19)</cell><cell>0.60(21)</cell><cell>0.15(13)</cell><cell>0.22(15)</cell><cell>0.58(17)</cell><cell>0.20(15)</cell></row><row><cell>Slovenian</cell><cell>0.18(17)</cell><cell>0.67(19)</cell><cell>0.14(10)</cell><cell>0.228(16)</cell><cell>0.55(19)</cell><cell>0.217(14)</cell></row><row><cell>Spanish</cell><cell>0.21(19)</cell><cell>0.65(20)</cell><cell>0.14(13)</cell><cell>0.22(17)</cell><cell>0.56(17)</cell><cell>0.22(16)</cell></row><row><cell>Swedish</cell><cell>0.25(19)</cell><cell>0.58(20)</cell><cell>0.17(16)</cell><cell>0.24(19)</cell><cell>0.497(19)</cell><cell>0.262(20)</cell></row><row><cell>Ukrainian</cell><cell>0.17(12)</cell><cell>0.69(13)</cell><cell>0.138(87)</cell><cell>0.19(14)</cell><cell>0.64(16)</cell><cell>0.17(12)</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 3</head><label>3</label><figDesc>Mean and standard deviation of averaged and maximum toxicity scores of user (U) and assistant (A) messages.</figDesc><table><row><cell cols="3">Language Average (U) Maximum (U)</cell><cell cols="2">Average (A) Maximum (A)</cell></row><row><cell>Bulgarian</cell><cell>0.0009(53)</cell><cell>0.0086(55)</cell><cell cols="2">0.00037(82) 0.0028(73)</cell></row><row><cell>Catalan</cell><cell>0.006(26)</cell><cell>0.05(18)</cell><cell>0.006(25)</cell><cell>0.04(17)</cell></row><row><cell>Croatian</cell><cell>0.0010(75)</cell><cell>0.007(44)</cell><cell>0.0005(20)</cell><cell>0.004(17)</cell></row><row><cell>Czech</cell><cell>0.002(12)</cell><cell>0.0143(74)</cell><cell>0.0015(80)</cell><cell>0.014(77)</cell></row><row><cell>Danish</cell><cell>0.003(16)</cell><cell>0.03(12)</cell><cell>0.003(17)</cell><cell>0.02(11)</cell></row><row><cell>Dutch</cell><cell>0.006(21)</cell><cell>0.05(18)</cell><cell>0.006(26)</cell><cell>0.042(17)</cell></row><row><cell>English</cell><cell>0.006(25)</cell><cell>0.05(18)</cell><cell>0.008(31)</cell><cell>0.06(22)</cell></row><row><cell>Estonian</cell><cell cols="2">0.00042(64) 0.0026(46)</cell><cell>0.0006(12)</cell><cell>0.005(12)</cell></row><row><cell>Finnish</cell><cell cols="2">0.00084(31) 0.007(27)</cell><cell>0.0009(53)</cell><cell>0.007(54)</cell></row><row><cell>French</cell><cell>0.005(19)</cell><cell>0.04(15)</cell><cell>0.005(22)</cell><cell>0.041(17)</cell></row><row><cell>Galician</cell><cell>0.0016(67)</cell><cell>0.014(68)</cell><cell>0.0008(40)</cell><cell>0.0073(35)</cell></row><row><cell>German</cell><cell>0.006(27)</cell><cell>0.04(16)</cell><cell>0.008(32)</cell><cell>0.06(20)</cell></row><row><cell>Greek</cell><cell>0.0007(19)</cell><cell>0.0046(14)</cell><cell>0.0004(11)</cell><cell>0.0030(92)</cell></row><row><cell cols="2">Hungarian 0.0014(58)</cell><cell>0.012(57)</cell><cell>0.0012(98)</cell><cell>0.008(52)</cell></row><row><cell>Irish</cell><cell cols="2">0.00032(45) 0.00222(41)</cell><cell cols="2">0.00042(57) 0.0025(39)</cell></row><row><cell>Italian</cell><cell>0.003(14)</cell><cell>0.02(11)</cell><cell>0.003(17)</cell><cell>0.03(14)</cell></row><row><cell>Latvian</cell><cell>0.002(11)</cell><cell>0.014(85)</cell><cell>0.0004(12)</cell><cell>0.004(13)</cell></row><row><cell cols="2">Lithuanian 0.00010(3)</cell><cell>0.00039(0)</cell><cell>0.00003(3)</cell><cell>0.00029(37)</cell></row><row><cell>Maltese</cell><cell>0.0005(11)</cell><cell>0.005(12)</cell><cell>0.0006(18)</cell><cell>0.006(19)</cell></row><row><cell cols="2">Norwegian 0.0007(24)</cell><cell>0.0066(25)</cell><cell>0.0006(20)</cell><cell>0.006(20)</cell></row><row><cell>Polish</cell><cell>0.002(12)</cell><cell>0.02(10)</cell><cell>0.002(15)</cell><cell>0.02(11)</cell></row><row><cell cols="2">Portuguese 0.002(13)</cell><cell>0.017(91)</cell><cell>0.002(12)</cell><cell>0.0133(85)</cell></row><row><cell>Romanian</cell><cell>0.0012(75)</cell><cell>0.011(68)</cell><cell>0.0008(56)</cell><cell>0.007(54)</cell></row><row><cell>Slovak</cell><cell>0.0020(72)</cell><cell>0.019(74)</cell><cell>0.0011(65)</cell><cell>0.011(72)</cell></row><row><cell>Slovenian</cell><cell>0.003(17)</cell><cell>0.02(14)</cell><cell>0.0008(26)</cell><cell>0.007(23)</cell></row><row><cell>Spanish</cell><cell>0.003(14)</cell><cell>0.02(11)</cell><cell>0.003(18)</cell><cell>0.02(12)</cell></row><row><cell>Swedish</cell><cell>0.005(15)</cell><cell>0.05(15)</cell><cell>0.003(10)</cell><cell>0.0262(11)</cell></row><row><cell>Ukrainian</cell><cell>0.0007(64)</cell><cell>0.006(45)</cell><cell>0.0005(44)</cell><cell>0.004(30)</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Table 4</head><label>4</label><figDesc>Results of the Kruskal-Wallis test for user messages.</figDesc><table><row><cell>Toxicity category</cell><cell>H-statistic</cell><cell>p-value</cell></row><row><cell>harassment</cell><cell>3091.9</cell><cell>0.0</cell></row><row><cell>harassment/threatening</cell><cell>163.4</cell><cell>1.8×10 −21</cell></row><row><cell>hate</cell><cell>8424.2</cell><cell>0.0</cell></row><row><cell>hate/threatening</cell><cell>314.2</cell><cell>1.1×10 −50</cell></row><row><cell>self-harm</cell><cell>811.6</cell><cell>1.4×10 −153</cell></row><row><cell>self-harm/instructions</cell><cell>30.2</cell><cell>3.1×10 −1</cell></row><row><cell>self-harm/intent</cell><cell>144.1</cell><cell>5.9×10 −18</cell></row><row><cell>sexual</cell><cell>717.7</cell><cell>7.3×10 −134</cell></row><row><cell>sexual/minors</cell><cell>730.9</cell><cell>1.2×10 −136</cell></row><row><cell>violence</cell><cell>1774.8</cell><cell>0.0</cell></row><row><cell>violence/graphic</cell><cell>666.0</cell><cell>5.0×10 −123</cell></row><row><cell>max_toxicity_score</cell><cell>3511.0</cell><cell>0.0</cell></row><row><cell>avg_toxicity_score</cell><cell>2307.0</cell><cell>0.0</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_5"><head>Table 5</head><label>5</label><figDesc>Search space for the hyperparameters of the classification head, and best values found. , 𝑎, 𝑢2 , |𝑎 − 𝑢 1 |, |𝑢 2 − 𝑎|, |𝑢 2 − 𝑢 1 |⟩ = 𝑖 𝑐𝑙𝑠</figDesc><table><row><cell>Hyperparameter</cell><cell>Search space</cell><cell></cell><cell>Best value</cell></row><row><cell>Learning rate</cell><cell cols="2">[1 × 10 −6 , 1 × 10 −3 ] ∈ R</cell><cell>4.7 × 10 −4</cell></row><row><cell>Batch size</cell><cell cols="2">{1, 4, 8, 16, 32, 64, 128}</cell><cell>1</cell></row><row><cell>Hidden size</cell><cell>[10, 3000] ∈ N</cell><cell></cell><cell>884</cell></row><row><cell>Number of layers</cell><cell>[1, 5] ∈ N</cell><cell></cell><cell>4</cell></row><row><cell>Use bias</cell><cell>{True, False}</cell><cell></cell><cell>True</cell></row><row><cell>Apply weighted loss</cell><cell>{True, False}</cell><cell></cell><cell>False</cell></row><row><cell cols="2">Loss label smoothing [0.0, 0.3] ∈ R</cell><cell></cell><cell>0.2</cell></row><row><cell>Activation function</cell><cell cols="3">{Tanh, ReLU, LeakyReLU, Sigmoid, GELU, SiLU, Mish} Mish</cell></row><row><cell></cell><cell></cell><cell>⟨𝑦𝑒𝑠, 𝑛𝑜, 𝑛/𝑎⟩</cell></row><row><cell></cell><cell cols="2">⟨𝑢 1 Classification head</cell></row><row><cell></cell><cell>𝑒(𝑢 1 )</cell><cell>𝑒(𝑎)</cell><cell>𝑒(𝑢 2 )</cell></row><row><cell></cell><cell>Pooling</cell><cell>Pooling</cell><cell>Pooling</cell></row><row><cell></cell><cell>Encoder</cell><cell>Encoder</cell><cell>Encoder</cell></row><row><cell></cell><cell>𝑢 1</cell><cell>𝑎</cell><cell>𝑢 2</cell></row><row><cell></cell><cell></cell><cell>Tokenizer</cell></row><row><cell></cell><cell></cell><cell>⟨𝑢 1 , 𝑎, 𝑢 2 ⟩</cell></row></table></figure>
		</body>
		<back>
			<div type="annex">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Supplementary materials</head><p>Supplementary materials are available at https://github.com/ACMCMC/eur-langs-convs-analysis.</p></div>			</div>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Attention is All you Need</title>
		<author>
			<persName><forename type="first">A</forename><surname>Vaswani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Shazeer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Parmar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Uszkoreit</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Jones</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">N</forename><surname>Gomez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Kaiser</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Polosukhin</surname></persName>
		</author>
		<ptr target="https://papers.nips.cc/paper_files/paper/2017/hash/3f5ee243547dee91fbd053c1c4a845aa-Abstract.html" />
	</analytic>
	<monogr>
		<title level="m">Advances in Neural Information Processing Systems</title>
				<imprint>
			<publisher>Curran Associates, Inc</publisher>
			<date type="published" when="2017">2017</date>
			<biblScope unit="volume">30</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">A contemporary review on chatbots, AI-powered virtual conversational agents, ChatGPT: Applications, open challenges and future research directions</title>
		<author>
			<persName><forename type="first">A</forename><surname>Casheekar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Lahiri</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Rath</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">S</forename><surname>Prabhakar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Srinivasan</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.cosrev.2024.100632</idno>
		<ptr target="https://www.sciencedirect.com/science/article/pii/S1574013724000169.doi:10.1016/j.cosrev.2024.100632" />
	</analytic>
	<monogr>
		<title level="j">Computer Science Review</title>
		<imprint>
			<biblScope unit="volume">52</biblScope>
			<biblScope unit="page">100632</biblScope>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Automating Customer Support: A Study on the Efficacy of Machine Learning-Driven Chatbots and Virtual Assistants</title>
		<author>
			<persName><forename type="first">V</forename><surname>Katragadda</surname></persName>
		</author>
		<ptr target="https://www.irejournals.com/" />
	</analytic>
	<monogr>
		<title level="j">IRE Journals</title>
		<imprint>
			<biblScope unit="volume">7</biblScope>
			<biblScope unit="page" from="600" to="610" />
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">ChatGPT for Language Teaching and Learning</title>
		<author>
			<persName><forename type="first">L</forename><surname>Kohnke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><forename type="middle">L</forename><surname>Moorhouse</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Zou</surname></persName>
		</author>
		<idno type="DOI">10.1177/00336882231162868</idno>
		<ptr target="http://journals.sagepub.com/doi/10.1177/00336882231162868.doi:10.1177/00336882231162868" />
	</analytic>
	<monogr>
		<title level="j">RELC Journal</title>
		<imprint>
			<biblScope unit="volume">54</biblScope>
			<biblScope unit="page" from="537" to="550" />
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">A Brief Overview of ChatGPT: The History, Status Quo and Potential Future Development</title>
		<author>
			<persName><forename type="first">T</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>He</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Sun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q.-L</forename><surname>Han</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Tang</surname></persName>
		</author>
		<idno type="DOI">10.1109/JAS.2023.123618</idno>
		<ptr target="https://www.ieee-jas.net/en/article/doi/10.1109/JAS.2023.123618.doi:10.1109/JAS.2023.123618" />
	</analytic>
	<monogr>
		<title level="j">IEEE/CAA Journal of Automatica Sinica</title>
		<imprint>
			<biblScope unit="volume">10</biblScope>
			<biblScope unit="page" from="1122" to="1136" />
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Crossing the Conversational Chasm: A Primer on Natural Language Processing for Multilingual Task-Oriented Dialogue Systems</title>
		<author>
			<persName><forename type="first">E</forename><surname>Razumovskaia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Glavas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Majewska</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">M</forename><surname>Ponti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Korhonen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Vulic</surname></persName>
		</author>
		<idno type="DOI">10.1613/jair.1.13083</idno>
		<ptr target="http://www.jair.org/index.php/jair/article/view/13083.doi:10.1613/jair.1.13083" />
	</analytic>
	<monogr>
		<title level="j">Journal of Artificial Intelligence Research</title>
		<imprint>
			<biblScope unit="volume">74</biblScope>
			<biblScope unit="page" from="1351" to="1402" />
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">AI Chatbots and Linguistic Injustice</title>
		<author>
			<persName><forename type="first">S</forename><surname>Park</surname></persName>
		</author>
		<idno type="DOI">10.22425/jul.2024.25.1.99</idno>
		<idno>doi:</idno>
		<ptr target="10.22425/jul.2024.25.1." />
	</analytic>
	<monogr>
		<title level="j">Journal of Universal Language</title>
		<imprint>
			<biblScope unit="volume">25</biblScope>
			<biblScope unit="page" from="99" to="119" />
			<date type="published" when="1999">2024. 99</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<monogr>
		<author>
			<persName><forename type="first">W</forename><surname>Zhao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Ren</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Hessel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Cardie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Choi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Deng</surname></persName>
		</author>
		<ptr target="https://openreview.net/forum?id=Bl8u7ZRlbM" />
		<title level="m">WildChat: 1M ChatGPT Interaction Logs in the Wild</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<monogr>
		<author>
			<persName><forename type="first">L</forename><surname>Zheng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W.-L</forename><surname>Chiang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Sheng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Zhuang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Zhuang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Lin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Xing</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">E</forename><surname>Gonzalez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Stoica</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Zhang</surname></persName>
		</author>
		<ptr target="https://openreview.net/forum?id=BOfDKxfwt0" />
		<title level="m">LMSYS-Chat-1M: A Large-Scale Real-World LLM Conversation Dataset</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<author>
			<persName><forename type="first">Y</forename><surname>Deng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Zhao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Hessel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Ren</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Cardie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Choi</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.2409.03753</idno>
		<idno type="arXiv">arXiv:2409.03753</idno>
		<ptr target="http://arxiv.org/abs/2409.03753.doi:10.48550/arXiv.2409.03753" />
		<title level="m">WildVis: Open Source Visualizer for Million-Scale Chat Logs in the Wild</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<monogr>
		<title level="m" type="main">Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks</title>
		<author>
			<persName><forename type="first">N</forename><surname>Reimers</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Gurevych</surname></persName>
		</author>
		<idno>arXiv:</idno>
		<ptr target="1908.10084[cs" />
		<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Cluster Quality Analysis Using Silhouette Score</title>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">R</forename><surname>Shahapure</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Nicholas</surname></persName>
		</author>
		<idno type="DOI">10.1109/DSAA49011.2020.00096</idno>
		<ptr target="https://ieeexplore.ieee.org/document/9260048.doi:10.1109/DSAA49011.2020.00096" />
	</analytic>
	<monogr>
		<title level="m">IEEE 7th International Conference on Data Science and Advanced Analytics (DSAA)</title>
				<imprint>
			<date type="published" when="2020">2020. 2020</date>
			<biblScope unit="page" from="747" to="748" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">XLM-T: Multilingual Language Models in Twitter for Sentiment Analysis and Beyond</title>
		<author>
			<persName><forename type="first">F</forename><surname>Barbieri</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Espinosa Anke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Camacho-Collados</surname></persName>
		</author>
		<ptr target="https://aclanthology.org/2022.lrec-1.27" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Thirteenth Language Resources and Evaluation Conference, European Language Resources Association</title>
				<editor>
			<persName><forename type="first">N</forename><surname>Calzolari</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">F</forename><surname>Béchet</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Blache</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">K</forename><surname>Choukri</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">C</forename><surname>Cieri</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">T</forename><surname>Declerck</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Goggi</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">H</forename><surname>Isahara</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">B</forename><surname>Maegaard</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Mariani</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">H</forename><surname>Mazo</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Odijk</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Piperidis</surname></persName>
		</editor>
		<meeting>the Thirteenth Language Resources and Evaluation Conference, European Language Resources Association<address><addrLine>Marseille, France</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2022">2022</date>
			<biblScope unit="page" from="258" to="266" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<monogr>
		<author>
			<persName><forename type="first">Y</forename><surname>Bai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Jones</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Ndousse</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Askell</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Dassarma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Drain</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Fort</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Ganguli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Henighan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Joseph</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Kadavath</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Kernion</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Conerly</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>El-Showk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Elhage</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Hatfield-Dodds</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Hernandez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Hume</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Johnston</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Kravec</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Lovitt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Nanda</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Olsson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Amodei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Brown</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Clark</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Mccandlish</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Olah</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Mann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Kaplan</surname></persName>
		</author>
		<ptr target="https://arxiv.org/abs/2204.05862v1" />
		<title level="m">Training a Helpful and Harmless Assistant with Reinforcement Learning from Human Feedback</title>
				<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<monogr>
		<title level="m" type="main">Know Your Limits: A Survey of Abstention in Large Language Models</title>
		<author>
			<persName><forename type="first">B</forename><surname>Wen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Yao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Feng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Tsvetkov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Howe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">L</forename><surname>Wang</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.2407.18418</idno>
		<idno type="arXiv">arXiv:2407.18418</idno>
		<ptr target="http://arxiv.org/abs/2407.18418.doi:10.48550/arXiv.2407.18418" />
		<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<monogr>
		<title level="m" type="main">Training language models to follow instructions with human feedback</title>
		<author>
			<persName><forename type="first">L</forename><surname>Ouyang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Jiang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Almeida</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">L</forename><surname>Wainwright</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Mishkin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Agarwal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Slama</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Ray</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Schulman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Hilton</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Kelton</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Miller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Simens</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Askell</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Welinder</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Christiano</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Leike</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Lowe</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.2203.02155</idno>
		<idno type="arXiv">arXiv:2203.02155</idno>
		<ptr target="http://arxiv.org/abs/2203.02155.doi:10.48550/arXiv.2203.02155" />
		<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<monogr>
		<title level="m" type="main">Mish: A Self Regularized Non-Monotonic Activation Function</title>
		<author>
			<persName><forename type="first">D</forename><surname>Misra</surname></persName>
		</author>
		<ptr target="https://www.bmvc2020-conference.com/conference/papers/paper_0928.html" />
		<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<monogr>
		<author>
			<persName><forename type="first">T</forename><surname>Akiba</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Sano</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Yanase</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Ohta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Koyama</surname></persName>
		</author>
		<idno>arXiv:</idno>
		<ptr target="1907.10902[cs,stat" />
		<title level="m">Optuna: A Next-generation Hyperparameter Optimization Framework</title>
				<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<monogr>
		<title level="m" type="main">Argilla -Open-source framework for data-centric NLP</title>
		<author>
			<persName><forename type="first">D</forename><surname>Vila-Suero</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Aranda</surname></persName>
		</author>
		<ptr target="https://github.com/argilla-io/argilla" />
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<monogr>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">H</forename><surname>Martins</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Fernandes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Alves</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><forename type="middle">M</forename><surname>Guerreiro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Rei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">M</forename><surname>Alves</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Pombal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Farajian</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Faysse</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Klimaszewski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Colombo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Haddow</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">G C</forename><surname>De Souza</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Birch</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">F T</forename><surname>Martins</surname></persName>
		</author>
		<ptr target="https://arxiv.org/abs/2409.16235v1" />
		<title level="m">Eu-roLLM: Multilingual Language Models for Europe</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<monogr>
		<title level="m" type="main">Decoupled Weight Decay Regularization</title>
		<author>
			<persName><forename type="first">I</forename><surname>Loshchilov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Hutter</surname></persName>
		</author>
		<ptr target="https://arxiv.org/abs/1711.05101v3" />
		<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<monogr>
		<title level="m" type="main">HuggingFace&apos;s Transformers: State-of-the-art Natural Language Processing</title>
		<author>
			<persName><forename type="first">T</forename><surname>Wolf</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Debut</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Sanh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Chaumond</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Delangue</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Moi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Cistac</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Rault</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Louf</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Funtowicz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Davison</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Shleifer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Von Platen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Ma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Jernite</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Plu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">L</forename><surname>Scao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Gugger</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Drame</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Lhoest</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">M</forename><surname>Rush</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.1910.03771</idno>
		<idno type="arXiv">arXiv:1910.03771</idno>
		<ptr target="http://arxiv.org/abs/1910.03771.doi:10.48550/arXiv.1910.03771" />
		<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">Comparing Germanic, Romance and Slavic: Relationships among linguistic distances</title>
		<author>
			<persName><forename type="first">W</forename><surname>Heeringa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Gooskens</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><forename type="middle">J</forename><surname>Van Heuven</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.lingua.2023.103512</idno>
		<ptr target="https://www.sciencedirect.com/science/article/pii/S0024384123000360.doi:10.1016/j.lingua.2023.103512" />
	</analytic>
	<monogr>
		<title level="j">Lingua</title>
		<imprint>
			<biblScope unit="volume">287</biblScope>
			<biblScope unit="page">103512</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
