<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Testing the Syntactic Competence of Large Language Models with a Translation Task Dative Ambiguity in Russian</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Edyta</forename><surname>Jurkiewicz-Rohrbacher</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">Universität Hamburg</orgName>
								<address>
									<addrLine>Mittelweg 177</addrLine>
									<postCode>22222</postCode>
									<settlement>Hamburg</settlement>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
							<affiliation key="aff1">
								<orgName type="institution">Universität Regensburg</orgName>
								<address>
									<addrLine>Universitätsstr. 34</addrLine>
									<postCode>93333</postCode>
									<settlement>Regensburg</settlement>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Testing the Syntactic Competence of Large Language Models with a Translation Task Dative Ambiguity in Russian</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">7519BAD1E55B7E1A428FBD4CC4D9B0B8</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T18:16+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>syntax</term>
					<term>ambiguity</term>
					<term>translation task</term>
					<term>linguistics</term>
					<term>linguistic competence of large language models</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>The paper explores opportunities for using a translation task to obtain knowledge about the syntactic competence of large language models. It reports the accuracy achieved in a Russian-English translation task on Russian sentences containing highly ambiguous structures with two dative personal pronouns. Seven tools (systems and agents) based on pre-trained generative models were tested in their function as machine translators on a data set obtained from several web corpora. The study shows that the principles of reference assignment relevant to the syntax of human language users (referential prominence and linear order of pronouns and predicates) are also statistically relevant for pre-trained generative models.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>The rapid developments in generative pre-trained language models have resulted in agents that deliver relatively well-formed texts in various natural languages. The economic result of this process is a large number of cheaply produced but relatively well-written machine-generated texts (MGT) freely circulating and spreading online. For linguists this means that language users are being exposed to automatically generated content on an equal footing with human-generated content. The language varieties emerging from MGTs are thus quite naturally becoming an object of linguistic research next to human varieties such as slangs, dialects, idiolects, etc. Consequently, linguistics as a discipline is facing new challenges pertaining to the methods through which knowledge about artificially emerging lects can be obtained. The new question before the linguistic community is: Can the established corpus-, psycho-and neurolinguistic methods be applied in research on rapidly emerging LLMs? In general, tasks intuitively formulated as instructions, where the input has a similar structure to the output are better processed in zero-shot prompts than tasks presented in other ways, for example, as finishing an incomplete sentence <ref type="bibr" target="#b0">[1]</ref>.This study aims to explore to what extent using translation, a method well-known from typological questionnaires, can be applied to explore the syntactic competence of LLMs. In the subsections that follow, translation as a task and a selected phenomenon of dative case ambiguity in Russian are described. Section 2 presents the study design. The central quantitative results are provided in Section 3, while minor results, which might feed into future studies, are presented in Section 4.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.1.">Translation task</head><p>Translation has been used as a data elicitation task in typological and psycholinguistic research in various ways. In linguistic fieldwork, translational questionnaires are frequently constructed to examine how a particular area of grammar with a known representation in language A is represented by its 4th Workshop on Humanities-Centred Artificial Intelligence 2024 <ref type="bibr">(CHAI 2024)</ref> edyta.jurkiewicz-rohrbacher@uni-hamburg.de (E. Jurkiewicz-Rohrbacher) https://www.slm.uni-hamburg.de/slavistik/personen/jurkiewicz-rohrbacher.html (E. Jurkiewicz-Rohrbacher) 0000-0001-6737-7847 (E. Jurkiewicz-Rohrbacher)</p><p>native speakers in language B [cf. 2]. In psycholinguistics, translation is in itself an object of study as a cognitive process <ref type="bibr" target="#b2">[3]</ref>, but it is also used as a method for accessing the linguistic performance and competence of multilingual speakers, e.g., for exploring their multilingual lexicons <ref type="bibr" target="#b3">[4]</ref>.</p><p>Previous studies <ref type="bibr" target="#b4">[5,</ref><ref type="bibr" target="#b5">6]</ref> suggest that pre-trained generative models do capture syntactic information. However, accessing this information seems computationally demanding, and due to various practical reasons, impossible in the case of very large, commercially developed models. To address this, the present study employs a translation task to access knowledge of the principles governing syntactic parsing by having various types of pre-trained systems or agents perform a translation task.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.2.">Test Case: Dative Ambiguity in Russian</head><p>Recent reports show that neural machine translation (NMT) systems still have shortcomings in the area of co-reference resolution and lexical cohesiveness, which results in inaccurate translation of pronouns <ref type="bibr" target="#b6">[7]</ref>. Syntactically ambiguous structures pose another type of problem <ref type="bibr" target="#b7">[8]</ref>, which I assume is challenging for the correlates of syntactic constituency parsing that might be found in generative pre-trained language models. A typical example of such structure is the prepositional phrase attachment, as in the often-cited sentence A man saw a woman with a telescope, where the phrase with a telescope can be parsed as an attribute of a woman or as an adjunct to the predicate saw. Nevertheless, ambiguity is an inherent feature of natural languages. Some scholars propose that it is a desirable quality because it facilitates efficient, that is, short and simple communication <ref type="bibr" target="#b8">[9]</ref>. To explore the feasibility of using translation as a task in research on the syntactic competence of large language models, I examined ambiguous Russian structures containing two personal pronouns in the dative case placed adjacently in a complex sentence.</p><p>Although Slavic languages have extremely flexible word order, ambiguity in syntactic role assignment is rather rare because of their rich morphology. However, when two arguments have identical lexicogrammatical properties and the same morphological marking on the sentence surface, ambiguity is possible even in the case of two full NPs, as shown in example (1).</p><p>(1) Miškata mouse.f.sg.def vižda see.ipfv.prs.3sg kotkata. cat.f.sg.def 'The mouse sees the cat / The cat sees the mouse' (Bg, <ref type="bibr" target="#b9">[10]</ref>)</p><p>Studies on the Russian dative case with infinitive structures <ref type="bibr" target="#b10">[11,</ref><ref type="bibr" target="#b11">12,</ref><ref type="bibr" target="#b12">13,</ref><ref type="bibr" target="#b13">14,</ref><ref type="bibr" target="#b14">15]</ref> mention in passing that the co-occurrence of two dative arguments in one sentence is possible, being predominantly observed in sentences with a free infinitive. <ref type="foot" target="#foot_0">1</ref> Such sentences are ambiguous because in several types of Russian clauses the dative case is not assigned solely to the syntactic role of indirect object (third argument), but also to the so-called 'logical subject',<ref type="foot" target="#foot_1">2</ref> as shown in ( <ref type="formula">2</ref> It is suggested <ref type="bibr" target="#b12">[13]</ref> that such structures might generally be avoided in language use. However, where they occurred, semantic-syntactic role assignment would follow the linear principle, correlating with the syntactic hierarchy of arguments (Agent over Recipient or other Participant). Others <ref type="bibr" target="#b13">[14]</ref> claim that a word order of dative arguments which is at variance with the syntactic hierarchy is marked only prosodically. Therefore, such structures could pose a challenge for large language models that are not trained on acoustic data.</p><p>Another work <ref type="bibr" target="#b14">[15]</ref> argues that the context clarifies role assignment. For example, in (3), it is the negative personal pronoun nekomu 'to/for nobody', and not the linearly first dative mne 'to/for me' which is higher in the syntactic structure, and therefore more subject-like.</p><p>(3) Mne me.dat</p><formula xml:id="formula_0">zvonit' call.inf nekomu nobody.dat -ja I i foc ne neg slušaju.</formula><p>listen.1sg 'Nobody calls me so I'm not listening (for the telephone).' [I. Grekova. Letom v gorode (1962)] (after <ref type="bibr" target="#b14">[15]</ref>) Finally, scholars have yet to provide an overview of structures in which two dative arguments interact in one sentence in Russian, and limit themselves to at least overtly one-predicate infinitive structures. <ref type="foot" target="#foot_2">3</ref>Hence, the order of predicates governing dative arguments is usually neglected as a factor.</p><p>A study on adjacent dative pronouns in Russian natural data originating from written text corpora <ref type="bibr" target="#b20">[21]</ref> establishes that such structures do occur in language use, albeit mostly in overtly bipredicative structures, in combination with embedded infinitival complements, as shown in example (4). <ref type="foot" target="#foot_3">4</ref>( The source also mentions some interesting speculations regarding the plans of Intel and NVIDIA, but we would like to dedicate a separate article to them. '</p><p>The linearly first dative pronoun im 'them' is governed by the infinitival complement posvjatit' 'dedicate', while the linearly second and adjacent pronoun nam 'us' is governed by the complementtaking matrix predicate chotelos' 'wish.refl'. Note that this sentence does not contain an explicit subject in the nominative.</p><p>According to the analysis in question <ref type="bibr" target="#b20">[21]</ref>, two factors significantly impact the probability of obtaining different word orders of arguments: the order of the main and embedded predicate, and the type of referential prominence that the pronouns represent, locuphoric pronouns (first/second person) being more likely to be assigned the agentive role than aliophoric ones (third person) <ref type="bibr" target="#b21">[22]</ref>.</p><p>For better comprehension the model is shown in Figure <ref type="figure" target="#fig_0">1</ref>. Considerable variation is observed in sentences with an infinitival complement preceding the matrix verb in the linear order of the sentence (CM type marked on the abscissa). In such environments, two locuphoric pronouns are more likely to comply with the deep syntactic order of the predicates rather than with the shallow order suggested by the surface.</p><p>In combinations with at least one aliophoric pronoun, the picture is more complex (marked with red circle). Sentences where the referential prominence hierarchy is retained show the highest variation in pronoun order, as the probabilities of the two word orders occurring are nearly equal. When the referential prominence hierarchy is violated (an aliphoric pronoun takes a high position in syntax), the pronoun order corresponding to the order of the predicates on the surface is preferred, and so is significantly more likely to occur. Nonetheless, without context it is impossible to distinguish between these two conditions, which is a source of ambiguity.</p><p>It may be predicted that in a translation task, adjacent dative pronouns in Russian structures with embedded infinitives would be a source of error for LLMs. A particularly high error rate is to be expected for sentences with an infinitival complement preceding the matrix predicate in the surface linear order of a sentence, and sentences with a combination of a locuphoric and an aliophoric pronoun.  <ref type="bibr" target="#b20">[21]</ref>. Red marked items are predicted to be source for error in interpretation for LLMs.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Study Design</head><p>In this section I describe the translation task conducted in the study. The primary sources of data were the Russian Timestamped JSI web corpus 2014-2021 <ref type="bibr" target="#b22">[23]</ref> and the ruTenTen17 corpus <ref type="bibr" target="#b23">[24]</ref>, from which I extracted 74 stimuli excerpts of 200-1100 characters each. <ref type="foot" target="#foot_4">5</ref> Every excerpt contained a sentence with a two-predicate structure, <ref type="foot" target="#foot_5">6</ref> in which the embedded predicate (complement) was placed earlier in the linear structure of the sentence than the embedding predicate (matrix), and which contained two adjacent personal pronouns in the dative case, one locuphoric and the other aliophoric (see example 5</p><p>Further in the paper I use the following notation: M stands for matrix predicate (syntactically higher predicate), C for complement predicate (syntactically lower predicate, embedded by M), D1 for dative pronoun governed by M, D2 for dative pronoun governed by C.</p><p>( nothing.gen 'They asked for plain water, but we had nothing to give them. '</p><p>The obtained data set was used to test the performance of three specialized translation tools based on neural network architectures, DeepL<ref type="foot" target="#foot_6">7</ref> , Google Translate<ref type="foot" target="#foot_7">8</ref> , Yandex <ref type="foot" target="#foot_8">9</ref> , and four chatbots with similar architectures: Google Gemini<ref type="foot" target="#foot_9">10</ref> , Perplexity AI <ref type="foot" target="#foot_10">11</ref> , ChatGPT Turbo and ChatGPT Omni <ref type="foot" target="#foot_11">12</ref> .</p><p>The choice of commercial tools has clear drawbacks. First, the exact details of commercial models' architectures and the structure of the training data are not disclosed. Second, the computations performed by the models cannot be controlled, nor can the models be fine-tuned to improve accuracy of performance. However, training methods are beyond the scope of this study. My objective was to verify to what extent errors in MGTs can be predicted on the basis of statistically significant regularities detected in the behavior of language users. The selected pre-trained generative agents are all based on encoder-decoder architectures. This paper treats them similarly to human agents in usage-based theories of language acquisition <ref type="bibr" target="#b24">[25]</ref>. In these theories, language acquisition is possible not thanks to universal grammar but is based on cognitive skills, in particular intention-reading and pattern-finding. Since the latter is clearly relevant to pre-trained generative models, I assume that their linguistic competence is emergent <ref type="bibr" target="#b25">[26]</ref>.</p><p>The training data represents the performance of multiple competent language users and is fed during the training process, from which the linguistic competence emerges, with the difference that each model has been exposed to a much larger amount of linguistic (written) data than any human agent can ever be. Knowledge about the linguistic competence of LLMs is accessed indirectly in this study, by evaluating performance in a specially developed translation task, just as it is done when the linguistic competence of human beings is studied. Therefore, for the selection of tools high competence had a greater priority than control over a language model. Another important argument in favor of commercial models was that typologically interesting Slavic languages, characterized by rich morphology and very flexible word order, are still rarely available in open-source multilingual models such as the Llama family. <ref type="foot" target="#foot_12">13</ref> Although ambiguity per se is not rare in language, keeping as many factors as possible constant, and thus focusing on only one type of ambiguity, leads to considerable data reduction. The construction examined in this study is rather complex and relatively rare. Therefore, correct performance requires big computational capacities and state-of-art technologies.</p><p>In the period 05-18.06.24, the sentences in their authentic contexts were fed into the translation systems as chunks of 200-800 characters. The size of each chunk depended on the place in the text where the context necessary for disambiguation was located. It was mainly found either before or after the tested sentence. In rare cases, both the pre-and post-context were necessary for an unambiguous interpretation of pronouns. The chatbots were zero-shot prompted with the command "Translate the following passage from Russian to English: [passage]". <ref type="foot" target="#foot_13">14</ref> For each stimulus, the chat was restarted and no feedback regarding performance was given. In this way, 518 observations were obtained.</p><p>In the study, I focused only on the correct assignment of syntactic roles to the dative pronouns, as rendered in the process of translation. Other types of translation errors were disregarded.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Results</head><p>Table <ref type="table" target="#tab_3">1</ref> demonstrates that the dative pronoun disambiguation task is not straightforward and that error rates vary primarily between the specialized translation systems and the agents. The best-performing ChatGPT Omni achieved an accuracy of nearly 0.95, while all the other chatbots had a strikingly similar accuracy of 0.89. The best translation system performed under this rate, reaching an accuracy of 0.85. The other two systems showed far poorer accuracy: 0.74 in the case of Google Translate and 0.67, of Yandex.</p><p>Interestingly, single instances of misclassification were observed only for translation systems, but not for generative agents. In other words, if an agent misclassified, there existed a system that misclassified too. In order to find out whether the word order and the prominence hierarchy principle were considerable factors, the data were annotated for these two features. The distributions of errors across them are presented in Table <ref type="table">2</ref>.</p><p>I observe that D2D1 word order with kept prominence hierarchy is clearly the easiest to classify causing barely any errors.</p><p>It may be observed that the D2D1 word order which preserved the prominence hierarchy was clearly the easiest to classify, causing barely any errors. Violation of one of the principles led to a considerable deterioration in the models' performance. I observed a decrease in accuracy when either the surface word order of the pronouns did not replicate the surface order of the predicates or the pronouns did not comply with the prominence hierarchy, i.e., when an aliophoric pronoun was higher in the syntactic hierarchy than a locuphoric one. However, if both of these principles were violated, the language models seemed to act quite randomly: performance dropped to 0.51.</p><p>A logistic-regression model with mixed-effects<ref type="foot" target="#foot_14">15</ref>  <ref type="bibr" target="#b26">[27]</ref> performed in R Environment <ref type="bibr" target="#b27">[28]</ref>, where stimuli and translator were treated as random effects, confirmed the intuition formulated above. The model shows that both factors (word order and prominence hierarchy) play a significant role in modeling the performance of the studied LLMs (cf. Table <ref type="table" target="#tab_4">3</ref>), and they have positive impact on performing the reference assignment in the translation task. and have a positive impact on correct reference assignment in the translation task. According to the studied model, the probability that sentences violating both principles will be accurately classified is 0.55. Sentences complying with both principles have a probability of 0.99 of being classified correctly. For sentences where only the hierarchy prominence is violated, the model predicts correct reference assignment with a probability of 0.93, while for sentences violating only the surface word order correspondence between governors and pronouns the respective number is 0.94.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Discussion and Future Prospects</head><p>The results obtained in the study are in line with prior predictions that "sentences with reversed order of predicates (CM), where two dative pronouns represent different levels of the prominence hierarchy can pose interpretation problems for NMT systems and other tools for NLP" <ref type="bibr" target="#b20">[21]</ref>. It appears that in such structures, contextually available information might not be sufficient for correct disambiguation by a machine; for example, the key features might not be identified, as in sentence (6) <ref type="foot" target="#foot_15">16</ref> which was misclassified by six out of seven translators: Both hierarchical prominence and concordance of the word order of the governors and pronouns turned out to be relevant factors that might facilitate or hinder the task of disambiguation for the purpose of role assignment, for instance if there are no contextual cues. It should be pointed out that typically, sentences with two dative pronouns do not contain a nominative phrase, which in traditional syntax would be interpreted as the canonical subject. Consequently, such sentences most likely place a greater burden on processing, assuming that a correlate of syntactic parsing emerges in LLMs <ref type="bibr" target="#b4">[5,</ref><ref type="bibr" target="#b5">6]</ref>. In other words, phenomena studied in theoretical linguistics and typology seem to be relevant and retrievable from the linguistic behavior of large language models, also through established linguistic methods. Although the way pre-trained language models process language is still comparable to a black box, I argue that methods used to study the linguistic behavior of the human species can be adjusted to studying the linguistic behavior of machines. If not human natural language users can be considered a black box too, since linguistic knowledge is never directly accessible. Another interesting finding of this study is that models pre-trained for performing various tasks communicated within a conversation performed better than specially trained machine translators. This result should by no means suggest that agents are in general better than translation systems at performing translation tasks, as only one particular aspect was evaluated. Nonetheless, in the future it should be examined whether this observation holds for other phenomena and what the reason might be. I cannot rule out that it is related to the number of parameters, the size of the context window, or the type of training data. Nevertheless, it is important to repeat that in this study, agents misclassified only structures which were misclassified by systems, that is, a subset of stimuli misclassified by systems and not a disjoint set of stimuli. This suggests the systematicity of errors made by the agents. Note that for all stimuli, disambiguation was always potentially possible due to the available context. Presumably, agents use context better than translation systems do. This could be due to the fact that agents are trained to be multifunctional. Contextually given information is necessary to perform other types of tasks and therefore better used, also in translation. Multilingualism is in a sense a byproduct. Verifying this claim would certainly require further studies on context processing in reference resolution tasks.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Limitations</head><p>Gicen that ambiguity is an intrinsic feature of natural language, this phenomenon is pertinent to the processing of any natural language by machines. This paper focuses on syntactic ambiguity which is associated with flexible word order and languages where a single morpheme can be used to encode various syntactic arguments. The Slavic branch is an illustrative example, where the scope of roles encoded by the dative is particularly extensive. This case study demonstrates that translation tasks can be employed to evaluate the capabilities of LLMs in a systemic manner and can serve as a foundation for future research.</p><p>The present study was limited to commercial products, which does not allow for evaluation of improvements on the training set. Moreover, the current tasks might permit improvement of the studied tools and thus the obtained results might not be replicable in the future.</p><p>To an extent, the study is limited by the small set of test sentences (which will be enlarged in the future) and neglecting of the contextual factors. However, the point of the study was to ascertain whether translation tasks can bring insights into the linguistic competences of LLMs. Furthermore, the same problem might be relevant for automatic dependency annotation.</p><p>Finally, the study was limited only to the automatic text processing tools only. Although it would be possible to perform a similar study with human language users would be possible, I do not expect that the results obtained in the same or similar task would surpass the best performing ChatGPT Omni. The set of stimuli is cognitively quite demanding. Therefore, I assume that in an artificial experimental setting, language users would base their choices on the two, linguistic principles discussed in this paper, rather than spending time on re-analyzing the full context because it is cognitively more demanding. However, it is precisely for this reason that I postulate that some notional linguistic rules, such as those observed in the theoretical and general linguistics, might be common to humans and machines,notwithstanding the fact, that one would expect the latter group to make fewer mistakes and to follow logic (provided by the context).</p><p>In addition, human users, unlike chatbots (agents), could make mistakes in different stimuli than the erroneous translation systems, which is an interesting result of this study worth further investigation.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure1: Probability of obtaining a reversed order of pronominal arguments in bipredicative structures with two dative arguments<ref type="bibr" target="#b20">[21]</ref>. Red marked items are predicted to be source for error in interpretation for LLMs.</figDesc><graphic coords="4,138.26,65.61,318.75,300.00" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 1</head><label>1</label><figDesc>Accuracy in the translation task.</figDesc><table><row><cell cols="2">Translation Systems</cell><cell>DL</cell><cell cols="2">Google Yandex</cell></row><row><cell>Accuracy</cell><cell></cell><cell>0.85</cell><cell>0.74</cell><cell>0.66</cell></row><row><cell>Chatbots</cell><cell></cell><cell cols="4">Omni Turbo Gemini Perplexity</cell></row><row><cell>Accuracy</cell><cell></cell><cell>0.95</cell><cell>0.89</cell><cell>0.89</cell><cell>0.89</cell></row><row><cell>Table 2</cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>Distribution of the studied factors.</cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>Incorrectly classified</cell><cell></cell><cell></cell><cell></cell><cell cols="2">Correctly classified</cell></row><row><cell cols="2">Prominence hierarchy no</cell><cell></cell><cell>yes</cell><cell cols="2">Prominence hierarchy no yes</cell></row><row><cell>Word order</cell><cell></cell><cell></cell><cell></cell><cell>word order</cell></row><row><cell>D1D2</cell><cell cols="3">24 (0.49) 43(0.17)</cell><cell>D1D2</cell><cell>25 216</cell></row><row><cell>D2D1</cell><cell cols="3">14 (0.15) 2(0.02)</cell><cell>D2D1</cell><cell>77 117</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Table 3</head><label>3</label><figDesc>Results of the performed regression model m not shouting or calling for help -it's pointless: the guys upstairs are far enough away from me, they have nothing to throw to me (and even if they threw a rope, how would I grab it?). '</figDesc><table><row><cell cols="8">Random effects:</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell cols="3">Groups</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell cols="2">Name</cell><cell>Variance Std.Dev.</cell></row><row><cell cols="10">SentID (Intercept)</cell><cell></cell><cell cols="3">3.5459</cell><cell>1.8831</cell></row><row><cell cols="11">Translator (Intercept)</cell><cell cols="3">0.9743</cell><cell>0.9871</cell></row><row><cell cols="6">Fixed effects:</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell cols="3">Estimate</cell><cell>Std. Error</cell><cell>z value Pr(&gt;|z|)</cell></row><row><cell cols="5">(Intercept)</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell cols="3">0.2053</cell><cell>0.9090</cell><cell>0.226</cell><cell>0.82133</cell></row><row><cell cols="2">D2D1</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell cols="3">2.3771</cell><cell>1.0624</cell><cell>2.237</cell><cell>0.02525*</cell></row><row><cell cols="6">D1 locuphoric</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell cols="3">2.5861</cell><cell>0.9511</cell><cell>2.719</cell><cell>0.00655*</cell></row><row><cell cols="10">D2D1:D1 locuphoric</cell><cell></cell><cell cols="3">0.9296</cell><cell>1.5092</cell><cell>0.616</cell><cell>0.53791</cell></row><row><cell>(6) Ja</cell><cell>ne</cell><cell cols="3">kriču</cell><cell></cell><cell>i</cell><cell></cell><cell></cell><cell>ne</cell><cell cols="2">zovu</cell><cell>na</cell><cell>pomošč'</cell><cell>-ėto</cell><cell>bessmyslenno:</cell><cell>rebjata</cell><cell>naverchu,</cell><cell>dostatočno</cell></row><row><cell>I</cell><cell>neg</cell><cell cols="4">shout.1sg</cell><cell cols="3">and</cell><cell>neg</cell><cell cols="2">call</cell><cell>on</cell><cell>help</cell><cell>it</cell><cell>pointless</cell><cell>kids</cell><cell>top</cell><cell>enough</cell></row><row><cell cols="2">daleko</cell><cell cols="2">ot</cell><cell cols="4">menja,</cell><cell cols="3">kinut' 𝐶</cell><cell cols="3">im 𝐷1</cell><cell>mne 𝐷2</cell><cell>nečego 𝑀</cell><cell>(a</cell><cell>esli</cell><cell>by</cell><cell>i</cell><cell>kinuli</cell></row><row><cell>far</cell><cell></cell><cell cols="2">from</cell><cell cols="2">me</cell><cell></cell><cell></cell><cell cols="3">throw.inf</cell><cell cols="3">them.dat</cell><cell>me.dat</cell><cell>nothing</cell><cell>conj</cell><cell>if</cell><cell>cond</cell><cell>foc</cell><cell>throw.pst.3pl</cell></row><row><cell cols="3">verëvku,</cell><cell>to</cell><cell></cell><cell cols="2">čem</cell><cell cols="2">za</cell><cell>neë</cell><cell></cell><cell cols="3">uchvatit'sja?).</cell></row><row><cell cols="2">rope</cell><cell></cell><cell cols="2">then</cell><cell>ins</cell><cell></cell><cell cols="2">for</cell><cell cols="2">she.acc</cell><cell cols="3">catch.inf.refl</cell></row><row><cell cols="2">'I'</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">A sentence where the main predicate is expressed as an infinitive.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_1">There is no general agreement as to which syntactic role should be assigned to such datives; for a recent review of the topic see<ref type="bibr" target="#b15">[16]</ref>. I refrain from generalization on this matter here, as the present study also involves object control structures where the syntactically highest dative argument carries the syntactic role of indirect object in the matrix predicate, but at the same time also assigns the semantic role of the non-overt subject in the complement clause.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_2">It is unclear whether sentences with a free infinitive are monoclausal<ref type="bibr" target="#b16">[17,</ref><ref type="bibr" target="#b17">18]</ref> or biclausal<ref type="bibr" target="#b18">[19,</ref><ref type="bibr" target="#b19">20]</ref>. In the latter case, scholars assume the existence of a copula, which is not marked overtly in present-tense sentences.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_3">https://overclockers.ru/hardnews/show/93720/kazhdyj-kvartal-sledujuschego-goda-budet-prinosit-novye-graficheskie-resheniya-amd</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_4">I applied a CQL query for two adjacent pronominal lowercase word forms using the Sketch Engine corpus manager. The obtained data were manually controlled by a native speaker annotator and controlled for error by a second native speaker annotator. Because the context was always available, the task was usually straightforward. Still, for the current task only such sentences were chosen that did not raise any doubts in either of the annotators.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="6" xml:id="foot_5">For instance subject or object control constructions, predicatives, or modal-existential wh-predicates.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="7" xml:id="foot_6">https://www.deepl.com, licenced account</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="8" xml:id="foot_7">https://translate.google.com, free user account</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="9" xml:id="foot_8">https://translate.yandex.com, free user account</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="10" xml:id="foot_9">https://gemini.google.com/app , free user account</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="11" xml:id="foot_10">https://www.perplexity.ai/, free user account</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="12" xml:id="foot_11">Access to both models via https://uhhgpt.uni-hamburg.de provided via the Universität Hamburg's license.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="13" xml:id="foot_12">The newest Llama 3.2 supports officially only English, German, French, Italian, Portuguese, Hindi, Spanish, and Thai.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="14" xml:id="foot_13">Some prompts had to be modified for Google Gemini, which sometimes failed to perform the task for various reasons.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="15" xml:id="foot_14">Formula: Correct ∼ Order + hierarchy + Order*hierarchy + (1 | SentID) + (1|Translator)</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="16" xml:id="foot_15">https://viktorkotl.livejournal.com/167122.html</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>The research has been partly supported by the Representative for Equal Opportunities and Academic Research Sabbatical Fund of the University of Regensburg. I thank Roman Fisun, Konstanzia Lüke and Irina Maykova for help with the preparation of the data set.</p></div>
			</div>

			<div type="annex">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Ethics Statement</head><p>This work complies with the ACL Ethics Policy. Prior to the current study, I had not taken any actions to pretrain the systems for the needs of the current task.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>A. Online Resources</head><p>The full list of stimuli and their translations is available via GitHub.</p></div>			</div>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<title level="m" type="main">Finetuned language models are zero-shot learners</title>
		<author>
			<persName><forename type="first">J</forename><surname>Wei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Bosma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><forename type="middle">Y</forename><surname>Zhao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Guu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">W</forename><surname>Yu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Lester</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Du</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">M</forename><surname>Dai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><forename type="middle">V</forename><surname>Le</surname></persName>
		</author>
		<ptr target="https://arxiv.org/abs/2109.01652.arXiv:2109.01652" />
		<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">From questionnaires to parallel corpora in typology</title>
		<author>
			<persName><forename type="first">Ö</forename><surname>Dahl</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Sprachtypologie und Universalienforschung</title>
		<imprint>
			<biblScope unit="volume">60</biblScope>
			<biblScope unit="page" from="172" to="181" />
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<monogr>
		<title level="m" type="main">Psycholinguistic and Cognitive Inquiries into Translation and Interpreting</title>
		<editor>A. Ferreira, J. W. Schwieter</editor>
		<imprint>
			<date type="published" when="2015">2015</date>
			<publisher>John Benjamins</publisher>
			<pubPlace>Amsterdam</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Some applications of translation to psycholinguistic research</title>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">M</forename><surname>Włosowicz</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Linguistica Silesiana</title>
		<imprint>
			<biblScope unit="volume">33</biblScope>
			<biblScope unit="page" from="127" to="145" />
			<date type="published" when="2012">2012</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Syntax representation in word embeddings and neural networksa survey</title>
		<author>
			<persName><forename type="first">T</forename><surname>Limisiewicz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Mareček</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 20th Conference Information Technologies -Applications and Theory (ITAT 2020)</title>
		<title level="s">CEUR-Workshop Proceedings</title>
		<editor>
			<persName><forename type="first">M</forename><surname>Holeňa</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">T</forename><surname>Horváth</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Kelemenová</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">F</forename><surname>Mráz</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">D</forename><surname>Pardubská</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Plátek</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Sosík</surname></persName>
		</editor>
		<meeting>the 20th Conference Information Technologies -Applications and Theory (ITAT 2020)<address><addrLine>Košice, Slovakia</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="38" to="48" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Do transformers parse while predicting the masked word?</title>
		<author>
			<persName><forename type="first">H</forename><surname>Zhao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Panigrahi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Ge</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Arora</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/2023.emnlp-main.1029</idno>
		<ptr target="https://aclanthology.org/2023.emnlp-main.1029.doi:10.18653/v1/2023.emnlp-main.1029" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, Association for Computational Linguistics</title>
				<editor>
			<persName><forename type="first">H</forename><surname>Bouamor</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Pino</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">K</forename><surname>Bali</surname></persName>
		</editor>
		<meeting>the 2023 Conference on Empirical Methods in Natural Language Processing, Association for Computational Linguistics<address><addrLine>Singapore</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="16513" to="16542" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Document-level neural MT: A systematic comparison</title>
		<author>
			<persName><forename type="first">A</forename><surname>Lopes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">A</forename><surname>Farajian</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Bawden</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">F T</forename><surname>Martins</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 22nd Annual Conference of the European Association for Machine Translation, European Association for Machine Translation</title>
				<editor>
			<persName><forename type="first">A</forename><surname>Martins</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">H</forename><surname>Moniz</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Fumega</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">B</forename><surname>Martins</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">F</forename><surname>Batista</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">L</forename><surname>Coheur</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">C</forename><surname>Parra</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">I</forename><surname>Trancoso</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Turchi</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Bisazza</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Moorkens</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Guerberof</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Nurminen</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">L</forename><surname>Marg</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><forename type="middle">L</forename><surname>Forcada</surname></persName>
		</editor>
		<meeting>the 22nd Annual Conference of the European Association for Machine Translation, European Association for Machine Translation<address><addrLine>Lisboa, Portugal</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="225" to="234" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<monogr>
		<title level="m" type="main">Going beyond the sentence: Contextual Machine Translation of Dialogue</title>
		<author>
			<persName><forename type="first">R</forename><surname>Bawden</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2018">2018</date>
			<pubPlace>Orsay, France</pubPlace>
		</imprint>
		<respStmt>
			<orgName>LIMSI, CNRS, Université Paris-Sud, Université Paris-Saclay</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">Ph.D. thesis</note>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">The communicative function of ambiguity in language</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">T</forename><surname>Piantadosi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Tily</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Gibson</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.cognition.2011.10.004</idno>
		<idno>doi:</idno>
		<ptr target="https://doi.org/10.1016/j.cognition.2011.10.004" />
	</analytic>
	<monogr>
		<title level="j">Cognition</title>
		<imprint>
			<biblScope unit="volume">122</biblScope>
			<biblScope unit="page" from="280" to="291" />
			<date type="published" when="2012">2012</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<author>
			<persName><forename type="first">M</forename><surname>Korytkowska</surname></persName>
		</author>
		<title level="m">Gramatyka konfrontatywna bułgarsko-polska, volume V, Slavistyczny Ośrodek Wydawniczy</title>
				<meeting><address><addrLine>Warsaw</addrLine></address></meeting>
		<imprint>
			<date type="published" when="1992">1992</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Konstrukcija tipa negde spat&apos;: sintaksis, semantika, leksikografija</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">D</forename><surname>Apresjan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">L</forename><surname>Iomdin</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Semiotika i informatika</title>
		<imprint>
			<biblScope unit="page" from="34" to="92" />
			<date type="published" when="1989">1989</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Infinitif et datif en polonais contemporain: un couple malheureux?</title>
		<author>
			<persName><forename type="first">D</forename><surname>Weiss</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Complétude et incomplétude dans les langues romanes et slaves. Actes du VI Colloque international de linguistique romane et slave</title>
				<editor>
			<persName><forename type="first">S</forename><surname>Karolak</surname></persName>
		</editor>
		<meeting><address><addrLine>Cracovie; Cracow</addrLine></address></meeting>
		<imprint>
			<date type="published" when="1991-10-03">29 sept.-3 oct. 1991. 1993</date>
			<biblScope unit="page" from="443" to="487" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<monogr>
		<author>
			<persName><forename type="first">F</forename><surname>Maurice</surname></persName>
		</author>
		<title level="m">Der modale Infinitiv in der modernen russischen Standardsprache</title>
				<meeting><address><addrLine>Munich</addrLine></address></meeting>
		<imprint>
			<publisher>Peter Lang</publisher>
			<date type="published" when="1996">1996</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<monogr>
		<author>
			<persName><forename type="first">A</forename><surname>Bonč-Osmolovskaja</surname></persName>
		</author>
		<title level="m">Konstrukcii s dativnym subjektom v russkom jazyke</title>
				<meeting><address><addrLine>Moscow</addrLine></address></meeting>
		<imprint>
			<publisher>MGU</publisher>
			<date type="published" when="2003">2003</date>
		</imprint>
	</monogr>
	<note type="report_type">PhD thesis</note>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Otricatel&apos;nye mestoimenija-predikativy (na ne-)</title>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">V</forename><surname>Padučeva</surname></persName>
		</author>
		<ptr target="http://rusgram.ru/new/chapter/pos/pronoun/#label_otritsateljnye_mestoimeniya-predikativy__na_ne-_" />
	</analytic>
	<monogr>
		<title level="m">Russkaja korpusnaja grammatika</title>
				<imprint>
			<date type="published" when="2015">26.02.2022. 2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Subject</title>
		<author>
			<persName><forename type="first">B</forename><surname>Hansen</surname></persName>
		</author>
		<idno type="DOI">10.1163/2589-6229_ESLO_COM_032471</idno>
		<ptr target="http://dx.doi.org/10.1163/2589-6229_ESLO_COM_032471" />
	</analytic>
	<monogr>
		<title level="m">Encyclopedia of Slavic Languages and Linguistics Online</title>
				<editor>
			<persName><forename type="first">M</forename><forename type="middle">L</forename><surname>Greenberg</surname></persName>
		</editor>
		<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Dva tipa dativnych predloženij v russkom jazyke</title>
		<author>
			<persName><forename type="first">A</forename><surname>Zimmerling</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Slovo -čistoe vesel&apos;e: Sbornik statej v čest</title>
				<editor>
			<persName><forename type="first">A</forename><forename type="middle">B</forename><surname>Pen'kovskogo</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename></persName>
		</editor>
		<meeting><address><addrLine>Moscow</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2009">2009</date>
			<biblScope unit="page" from="471" to="489" />
		</imprint>
	</monogr>
	<note>Jazyki slavjanskich kul&apos;tur</note>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Dative infinitive constructions in Russian: Are they really biclausal</title>
		<author>
			<persName><forename type="first">E</forename><surname>Tsedryk</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Formal approaches to Slavic Linguistics 25 : The third Cornell meeting 2016</title>
				<editor>
			<persName><forename type="first">B</forename><surname>Wayles</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">D</forename><surname>Miloje</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Enzinna</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Harmath De Lemos</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">R</forename><surname>Karlin</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">D</forename><surname>Zec</surname></persName>
		</editor>
		<meeting><address><addrLine>Ann Arbor</addrLine></address></meeting>
		<imprint>
			<publisher>Michigan Slavic Publications</publisher>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="298" to="317" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Modal possessive constructions: Evidence from Russian</title>
		<author>
			<persName><forename type="first">I</forename><surname>Livitz</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Lingua</title>
		<imprint>
			<biblScope unit="volume">122</biblScope>
			<biblScope unit="page" from="714" to="747" />
			<date type="published" when="2012">2012</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Quantificational properties of neg-wh items in Russian</title>
		<author>
			<persName><forename type="first">N</forename><surname>Kondrashova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Šimík</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of North East Linguistics Society 40</title>
				<meeting>North East Linguistics Society 40<address><addrLine>Amherst</addrLine></address></meeting>
		<imprint>
			<publisher>Graduate Linguistic Student Association</publisher>
			<date type="published" when="2013">2013</date>
			<biblScope unit="page" from="15" to="28" />
		</imprint>
		<respStmt>
			<orgName>University of Massachusetts</orgName>
		</respStmt>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">Dative ambiguity in Russian: A corpus induced study</title>
		<author>
			<persName><forename type="first">E</forename><surname>Jurkiewicz-Rohrbacher</surname></persName>
		</author>
		<idno type="DOI">10.2478/jazcas-2023-0025</idno>
	</analytic>
	<monogr>
		<title level="j">Journal of Linguistics/Jazykovedný casopis</title>
		<imprint>
			<biblScope unit="volume">74</biblScope>
			<biblScope unit="page" from="70" to="80" />
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">Role-reference associations and the explanation of argument coding splits</title>
		<author>
			<persName><forename type="first">M</forename><surname>Haspelmath</surname></persName>
		</author>
		<idno type="DOI">10.1515/ling-2020-0252</idno>
		<idno>doi:</idno>
		<ptr target="10.1515/ling-2020-0252" />
	</analytic>
	<monogr>
		<title level="j">Linguistics</title>
		<imprint>
			<biblScope unit="volume">59</biblScope>
			<biblScope unit="page" from="123" to="174" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">The internals of an aggregated web news feed</title>
		<author>
			<persName><forename type="first">M</forename><surname>Trampuš</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Novak</surname></persName>
		</author>
		<ptr target="http://ailab.ijs.si/dunja/SiKDD2012/Papers/Trampus_Newsfeed.pdf" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of 15th Multiconference on Information Society 2012 (IS-2012)</title>
				<meeting>15th Multiconference on Information Society 2012 (IS-2012)</meeting>
		<imprint>
			<date type="published" when="2012">2012</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">The TenTen corpus family</title>
		<author>
			<persName><forename type="first">M</forename><surname>Jakubíček</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Kilgarriff</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Kovář</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Rychlý</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Suchomel</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">7th International Corpus Linguistics Conference CL</title>
				<imprint>
			<date type="published" when="2013">2013</date>
			<biblScope unit="page" from="125" to="127" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<monogr>
		<title level="m" type="main">Constructing a Language: A Usage-Based Theory of Language Acquisition</title>
		<author>
			<persName><forename type="first">M</forename><surname>Tomasello</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2003">2003</date>
			<publisher>Harvard University Press</publisher>
			<pubPlace>Harvard</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<analytic>
		<title level="a" type="main">Emergent grammar</title>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">J</forename><surname>Hopper</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Berkeley Linguistics Society</title>
		<imprint>
			<biblScope unit="page" from="139" to="157" />
			<date type="published" when="1987">1987</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<analytic>
		<title level="a" type="main">Fitting linear mixed-effects models using lme4</title>
		<author>
			<persName><forename type="first">D</forename><surname>Bates</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Mächler</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Bolker</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Walker</surname></persName>
		</author>
		<idno type="DOI">10.18637/jss.v067.i01</idno>
	</analytic>
	<monogr>
		<title level="j">Journal of Statistical Software</title>
		<imprint>
			<biblScope unit="volume">67</biblScope>
			<biblScope unit="page" from="1" to="48" />
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b27">
	<monogr>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">C</forename><surname>Team</surname></persName>
		</author>
		<ptr target="https://www.R-project.org/" />
		<title level="m">R: A language and environment for statistical computing</title>
				<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
