<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Integrating TEI/XML Text with Semantic Lexicographic Data</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author role="corresp">
							<persName><forename type="first">Tarrin</forename><surname>Wills</surname></persName>
							<email>tarrin@hum.ku.dk</email>
							<affiliation key="aff0">
								<orgName type="department">Dictionary of Old Norse Prose</orgName>
								<orgName type="institution">University of Copenhagen</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Thor</forename><surname>Johannsson</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Dictionary of Old Norse Prose</orgName>
								<orgName type="institution">University of Copenhagen</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Simonetta</forename><surname>Battista</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Dictionary of Old Norse Prose</orgName>
								<orgName type="institution">University of Copenhagen</orgName>
							</affiliation>
						</author>
						<title level="a" type="main">Integrating TEI/XML Text with Semantic Lexicographic Data</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">E5CF235C24F89D80831339B38C601CBF</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-23T20:56+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Lexicography</term>
					<term>Textual editing</term>
					<term>TEI</term>
					<term>XML</term>
					<term>Databases</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Traditional excerption-based historical dictionaries often provide a very detailed semantic analysis of a high proportion of words in the corpora they cover. The Dictionary of Old Norse Prose will have analyzed and defined around 7% of all words in a 11 million word corpus, for example. Linking the semantic analysis of excerpted citations to new digital texts of the works in the corpus offers the potential to give much more detailed context for the citations in the dictionary and at the same time contextual semantic information (definitions) for a high proportion of specific words in the corpus. The task is nontrivial as it involves linking separately-formed datasets consisting of tens of thousands of tokens. This paper describes a process by which a very high proportion of citations in the dictionary are linked to individual words in new digital editions, using sorting and lexical information. The result is that users of the dictionary can view the citations in their full textual context, and readers of the text can access definitions specific to individual words in the text.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Background</head><p>The Dictionary of Old Norse Prose (ONP: onp.ku.dk) is an extensive digital resource which links the semantic analysis of the lexicon of Old Norse with its material record (manuscripts and charters). ONP started in 1939 with a traditional process of manual excerption of words from manuscript-based editions. In this process, dictionary staff selected words from editions that were significant for lexicographic treatment. ONP works on the principle that all words are identifiable in their material context, that is, every word comes from a reading of the original manuscripts. The digital resource encodes this complex tripartite structure of words, texts and manuscripts. An example of the excerption process for a print edition is shown in Fig. <ref type="figure" target="#fig_0">1</ref>. Words selected for excerption were underlined and later copied onto paper slips. These have since been digitized along with metadata covering the original manuscript of the edition as well as the page and line references for each word cited.</p><p>The process of excerpting and supplementing citations has resulted in an index of around 800,000 citations from a corpus we estimate at around 11 million words. The actual number of words in original documents is much greater given the proliferation of copies of Old Norse manuscripts, but this figure represents non-repeated text. The dictionary is still in progress, with about 50% of the citations analyzed as part of finished dictionary entries. By the time of completion, the dictionary will therefore have a semantic analysis of around 7% of the words in the entire corpus. The activity of excerption means that the focus is largely on rarer words and usages, making this potentially a very rich resource for readers of these texts. Already readers can access a large number of out-of-copyright editions and view the accompanying information about each word that has been excerpted <ref type="bibr" target="#b0">[1]</ref>.</p><p>The methods and data structures for the dictionary, which began in 1939, were developed before digital corpus linguistics was possible. The dictionary's methods are based on manual excerption of words and surrounding text and are not in themselves automatically compatible with corpus-based approaches. The dictionary uses preferably diplomatic editions where orthography is not normalized, thus showing something close to the original manuscript form, and is heavily reliant on print editions. The project digitizes the citations manually, which involves some adjustment to the orthography and mark-up.</p><p>Other projects, however, have been developing manuscript-based Old Norse digital texts that belong to the same corpus covered by ONP and use a compatible manuscriptbased approach.  <ref type="bibr" target="#b1">[2]</ref> showing the words excerpted for later inclusion in the dictionary. All these words can be used to generate a contextual glossary once they are analysed.</p><p>The most extensive of these is the Menota project (menota.org <ref type="bibr" target="#b2">[3]</ref>) which includes a catalogue of TEI/XML manuscript texts encoded according to a specified subset of TEI. Central to Menota's guidelines <ref type="bibr" target="#b3">[4]</ref> is the principle that every word in the manuscript should be marked up with a TEI &lt;w&gt; (word) element. This can be extended to include up to three stages of normalization representing something very close to the manuscript orthography including abbreviations, an intermediate stage with abbreviations expanded and non-phonetic orthographic variation normalized slightly, and a fully normalized level corresponding to a kind of Classical Old Norse orthography. The markup can be further extended to include dictionary lemmas and a morphosyntactic analysis.</p><p>Menota's online catalogue <ref type="bibr" target="#b4">[5]</ref> incorporates these features into its interface, so that users can access the information about each word (see Fig. <ref type="figure" target="#fig_1">2</ref>). Where the text includes lemmas, the user can use them to look up the word in a dictionary, eliminating some difficulties with morphological (grammatical) or orthographic (spelling) variation. The inserted lemma, however, does not provide a concrete link to the equivalent word in a dictionary, which may use different orthography or contain different homographs. Old Norse manuscripts are notoriously variable in their orthography and use abbreviations extensively. There is a sufficiently large discrepancy in transcription practices that it is extremely difficult to use simple string comparison to match the same text in different digital versions. This is particularly the case between earlier print editions and state-of-the-art digital editions. Even among the relatively standard Menota editions there are differences in transcription practice and in the inclusion of the diplomatic and/or normalized 'levels' of the text.</p><p>A previous paper by the authors <ref type="bibr" target="#b5">[6]</ref> describes a fast and user-friendly workflow whereby Menota texts can be linked at the lexical level to dictionary headwords in ONP using a combination of automated and manual stages. This workflow is designed to achieve very high levels of accuracy (close to 99.9%) for the automated stages. The workflow demonstrates an interoperable method whereby TEI/XML encoded texts can be integrated and linked into relational data models such as dictionaries. These methods are designed to maintain a link between the two external data sources at the level of the word so that they can be edited and maintained separately.</p><p>Linking lemmas at the lexical level means that users can access the dictionary directly by interacting digitally with the words in the text: clicking on a word, for example, can bring up a full dictionary entry regardless of homographs, and regardless of the normalisation or lemmatisation used in the particular text edition. It also means that full concordances for a particular lemma can be generated automatically. In the reverse direction, a user of the dictionary can be directed to the exact page and line of the original document. Where images are available it means that the material source of the lexicographic data can be instantly consulted. These capabilities have already been built into the web application ONP Online at onp.ku.dk for texts that have been treated in this way.</p><p>Where this process falls short is in the semantic analysis of the word in the text: a dictionary entry can tell the user all the meanings of a word, but not what the word means in that particular context. This limitation applies even when a dictionary like ONP contains that information for a large proportion of words in any given text: for almost all texts between 5% and 15% of words have been excerpted and form part of the dictionary's analysis. In some cases close to 100% have been excerpted. The excerpted words focus on rarer lemmas and usages, meaning that there is a higher proportion still of words treated by the dictionary that might either cause difficulty or be of interest to a reader. From the potential user's point of view, the words treated by the dictionary are therefore quite likely to be a high proportion of those that an intermediate user may want further information on. The potential exists in combining the two datasets to retrieve semantic information about individual words from the manuscript text itself.</p><p>The current research builds on these resources and processes to link the words of the corpus deeper into the dictionary's semantic structure. A dictionary aims for not just a lexical but also a semantic overview of the corpus. This is done in traditional dictionaries such as ONP by excerpting relevant words from the corpus and analysing every citation excerpted, building a semantic tree of how the headword is used in the texts. Every node in that tree contains a sense and a definition of that sense, forming the structure of the dictionary entry. With such dense excerption of examples in a dictionary such as ONP, it is technically possible to link a high proportion of words in a given text to a particular semantic analysis as assigned by dictionary editors. That is, a high proportion of words can be potentially specifically linked to the individual senses and definitions of the structured dictionary entry. For the user this would mean that they can find a particular sense or usage of a word in a digital edition of a manuscript in the specific context they are reading, according to the dictionary's analysis.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Method</head><p>This goal of this project is the digital linking of individual words in a text not only to the dictionary headword but to the particular citation in the semantic tree of the dictionary entry. This is a nontrivial task. For a longer text of 100,000 words where the dictionary has excerpted some 10,000 citations, any system designed to align the two datasets must potentially deal with the Cartesian product of the two sets, namely one billion possible combinations. This processing task is simplified by using additional information to reduce the number of possible combinations, including the ordering of words and a common reference point in the lemma list. The processing is done by a system suited to the alignment of large datasets.</p><p>The references in the dictionary for citations are in almost all cases to the physical page and line of the published edition. For Menota-style TEI texts the words can normally be identified by the page and line of the manuscript version of the text. The two sets of references are in largely the same order (although this cannot always be determined for two words on the same line in the dictionary) but are not otherwise compatible. Not all words are excerpted by the dictionary, leaving no simple way of aligning and linking the two types of reference. There are nevertheless a number of principles that allow us to establish a largely automated process of linking:</p><p>1. The two datasets are in exactly the same order, despite some differences in readings.</p><p>If a single word is matched between the two datasets then any word before (or after) it in the one dataset will correspond to a word before (or after) it in the other dataset. 2. All the words in one dataset (the manuscript-based editions) can be accurately linked to the dictionary headwords using an existing workflow. 3. The other dataset (dictionary citations) may be incomplete but all of the incomplete subset of words are linked to the dictionary headwords, providing the same point of reference for a subset of both corpora. 4. For lower frequency words it is very likely that the dictionary has excerpted all examples in a given text.</p><p>Therefore, if a citation in the dictionary is the only example of a particular headword excerpted from the text, and if the edition has only one word linked to the same headword then it is very likely the same word in the same context in the same text. That is, the word in the text can be linked to the corresponding citation in the dictionary (and vice-versa) in the case of most low-frequency words in the edition. The same principle applies to smaller sections of text where the two resources are known to contain the same text. Once some words are linked, the same method can be applied to the words in both datasets that occur between the linked words.</p><p>The methodology employed here has as its first stage to identify (by database queries) lemmas that appear only once in the TEI/XML text and which also appear only once in citations from the same text for that lemma in the dictionary. Because there is little chance for ambiguity that the word in such a case corresponds to the citation in the dictionary, the word and citation can be fairly reliably linked automatically. Accuracy is around 95% and so these links require manual checking that the citation is of the same word in the same context as the word in the text.</p><p>The two datasets are currently housed on different database systems, with the imported Menota data in a MariaDB server and ONP on an Oracle system. The two systems are fairly compatible, allowing for the following process:</p><p>1. A subset of the word data for the text including the lemma references, ordering of the words and manuscript page/line references is exported to the ONP database. 2. A complex SQL query is run to align the words: a. A virtual table is built with the sections of text that are framed by the aligned words (initially the whole text), including the references to the ordering in both datasets. b. A virtual table is built with all the citations in ONP for the same text, including the page/line references to the edition, which forms the basis of the ordering of the cited words in ONP. c. These two tables are joined, and lemmas that are only found once in each section of the text among ONP's citations then form a third virtual table. d. A fourth virtual table links the unique words from ONP to the imported word table with the same lemmas. This table is then filtered to include lemmas that only appear once in each section of the imported text. e. The final table assembles then the references to the original words/citations, exporting the cross-references so that both datasets can be updated with a reference for each matched word to the corresponding word in the other dataset. 3. The exported cross-references are imported into the original data tables for each word/citation. 4. The process is then repeated so that the words matched in the previous stage can then frame the searches in more specific sections of text. In practice this is only of use for about three iterations, after which almost no further words can be matched.</p><p>A more technical description of these steps and the SQL code used can be found at https://goo.gl/ncdWAC ( §9.4). The initial links between the words and the citation index therefore provide a framework by which the same method can be applied to the smaller sections of text between the linked words. This again involves identifying lemmas unique to each section of text in both the manuscript-based edition and the print-based edition used by the dictionary. The process uses page and line references in each case to frame the extent of the section searched. Links are inserted in both data structures: in the dictionary to the word in the text, and in the text to the citation in the dictionary. This method is repeated, with decreasing gaps between the identified words, until no further automatic linking is possible. The remaining words tend to either be unlinked by the previous lemmatising processes, or are ambiguous in some way within this process and are therefore excluded.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Results</head><p>The process is extremely quick. Modern relational database management systems are optimized to align very large amounts of data and rapidly build virtual tables. The SQL query that performs the main work of this operation contains 33 lines of code and takes about 1-2 minutes to execute for each iteration on a single-processor system, allowing this process to be repeated and refined many times.  This process was tested for Barlaams saga, a text of some 76,000 words for which ONP has some 10,000 citations. In the first iteration 966 citations (9%) were matched. Fig. <ref type="figure" target="#fig_2">3</ref> shows the results with a page taken from the original non-digital edition used by the dictionary together with the matched (grey) and unmatched (white) citations excerpted from the page by the dictionary. As the first stage has few points of alignment, the resulting matches required manual checking, aided by queries retrieving surrounding text. The second iteration, using these initial matches, produced 4,318 further matches. Subsequent iterations matched a further 430 and 23 words respectively. The result is that 56% of all citations could be matched with a very high level of accuracy by this method. Fig. <ref type="figure" target="#fig_3">4</ref> shows the resulting matches after three iterations of the query.</p><p>The same method has also been applied to Strengleikar (the Old Norse version of the Lais of Marie de France) in Uppsala manuscript DG 4-7 (onp.ku.dk/r10468). The method described above automatically and accurately linked 3168 of the 4065 citations in ONP to the Menota edition, representing 8% of the whole text and 78% of citations. The reasons for the much higher success rate with this text compared with Barlaams saga require further investigation, but it seems at this stage that a shorter text with fewer variants can be processed more successfully.</p><p>These links (as URIs and/or database keys) represent the minimal information needed to connect the words in each resource and are maintained even as the texts and the dictionary continue to be edited and developed separately.</p><p>The method is conservative and produces very few false positive matches. This requires further verification, but the authors have found very few instances where the citation is not correctly matched to the word in the manuscript text. The method, however, produces a large number of false negatives, with 20-40% of citations not matched to the corresponding words in the text. For a project such as ONP, traditional users expect extremely high levels of accuracy. Reducing the number of false negatives would likely increase the number of false positives, which would be an unacceptable trade-off in such a project: users are more likely to lose faith in the quality of the dictionary if they find information that is incorrect compared with finding information that is lacking.</p><p>The unmatched citations (false negatives) are owing to various causes. Differences in lemmatization, normalization and analytical practices mean that sometimes the same word can appear under different headwords, making them impossible to match by this method alone. In many cases a citation of a single word may appear more than once in ONP's citation index because it also records manuscript variation or other features. The alignment algorithm needs therefore to be adjusted so that it does not assume that such cases are instances of the same lemma appearing more than once, thus preventing alignment based on a unique lemma. With such refinements we expect this process to produce 80-90% matches in the future, with fewer false positives in the initial stage.</p><p>For the user the linking of words in a text to analyzed citations means that when they access the text, the individual words are not only linked to the dictionary entry, but in a good proportion of instances they are linked to the individual definition and/or phrasal-grammatical context of the word as defined in the dictionary. A userfor example a student or researchercan pull up a section of text and click on any word to get the dictionary entry, if available. Words linked at the citation level can be highlighted to indicate that further information is linked and when clicked will show the individual definition for the word, if available, and other information that the dictionary may record about that particular citation, such as the citation slip and edition information. Users of the dictionary can find specific examples for usages and can access the full text where that usage occurs, rather than the minimal surrounding text normally provided for each citation. (For an example see Fig. <ref type="figure">5</ref>, also at https://onp.ku.dk/c475521 and click on the Menota button. The red coloured words are linked to other citations in the dictionary, many of which have been defined.) Fig. <ref type="figure">5</ref>. Citations are linked to the full manuscript text, which in turn can link all matched words back to the dictionary. Here a word in the text links to the specific definition for the citation in the dictionary (see https://onp.ku.dk/c185746 -'Menota' button).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Discussion</head><p>The advantages of the linking method described here include the automated generation of integrated glossaries for the text edition. Such glossaries can be used as a reading aid, for those less familiar with the language, and as a language learning tool. Glosses assist in language acquisition by improving text comprehension and aiding in vocabulary acquisition <ref type="bibr" target="#b6">[7]</ref>. This applies also to digital glosses and 'authentic' texts <ref type="bibr" target="#b7">[8]</ref>, such as those used in this project. While the research in this area generally focuses on living languages, it is likely that such tools would also be of benefit to students wishing to improve their reading knowledge of a historical language such as Old Norse.</p><p>The semantic analysis of significant portions of the text can be further developed if the dictionary at a later point, as is hoped, integrates a digital thesaurus. The thesaurus, when linked to particular senses in the dictionary, can be integrated into the text itself, potentially creating a semantic map of the text as a whole and helping users to find semantically similar material in the corpus. Lastly, the majority of words, those which are not analysed by the dictionary project, can be semantically analysed according to statistical or other digital methods so that the particular meanings of non-manually analysed words can potentially be predicted from those in similar contexts.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Fig. 1 .</head><label>1</label><figDesc>Fig. 1. Pages from the Saga of St Thomas<ref type="bibr" target="#b1">[2]</ref> showing the words excerpted for later inclusion in the dictionary. All these words can be used to generate a contextual glossary once they are analysed.</figDesc><graphic coords="2,124.70,375.40,345.87,256.30" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Fig. 2 .</head><label>2</label><figDesc>Fig. 2. Information linked to digital texts in the Menota archive includes lexical and morphosyntactic information, but does not easily connect to the semantic and lexicological information provided by dictionaries.</figDesc><graphic coords="3,124.70,339.40,345.83,202.75" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Fig. 3 .</head><label>3</label><figDesc>Fig. 3. First iteration of query matches around 10% of citations (highlighted grey here) with the corresponding manuscript text.</figDesc><graphic coords="7,124.70,183.40,345.86,187.80" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Fig. 4 .</head><label>4</label><figDesc>Fig. 4. Third iteration of the matching query showing a majority of citations linked to the corresponding word in the manuscript text (see https://onp.ku.dk/b2352-11a).</figDesc><graphic coords="7,124.70,411.20,345.87,188.50" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0"><head></head><label></label><figDesc></figDesc><graphic coords="9,124.70,231.40,345.82,173.35" type="bitmap" /></figure>
		</body>
		<back>
			<div type="annex">
<div xmlns="http://www.tei-c.org/ns/1.0"><p>The method applied here can be potentially applied for any project where a subset of words needs to be realigned with a separate digital corpus with differing orthography. This includes similar excerption-based historical dictionaries that have been digitized, and smaller projects such as glossaries of individual texts or subcorpora. The key foundation to this approach is an accurate process to completely lemmatize the words in the text.</p></div>			</div>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Reengineering an Online Historical Dictionary for Readers of Specific Texts</title>
		<author>
			<persName><forename type="first">T</forename><surname>Wills</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Johannsson</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Electronic lexicography in the 21 st century: Smart lexicography: Proceedings of the eLex 2019 conference</title>
				<editor>
			<persName><forename type="first">I</forename><surname>Kosem</surname></persName>
		</editor>
		<meeting><address><addrLine>Brno</addrLine></address></meeting>
		<imprint>
			<publisher>Lexical Computing</publisher>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="116" to="129" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<monogr>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">R</forename><surname>Unger</surname></persName>
		</author>
		<title level="m">Thomas Saga Erkibyskups: Fortaelling om Thomas Becket Erkebiskop af Canterbury: To Bearbeidelser samt Fragmenter af en tredie</title>
				<meeting><address><addrLine>Kristiania</addrLine></address></meeting>
		<imprint>
			<date type="published" when="1869">1869</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<monogr>
		<ptr target="https://menota.org" />
		<title level="m">Medieval Nordic Text Archive (Menota)</title>
				<imprint>
			<date type="published" when="2020-12-02">2020/12/02</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">The Menota handbook: Guidelines for the electronic encoding of Medieval Nordic primary sources</title>
		<author>
			<persName><forename type="first">O</forename><forename type="middle">E</forename><surname>Haugen</surname></persName>
		</author>
		<ptr target="http://www.menota.org/handbook.xml" />
	</analytic>
	<monogr>
		<title level="m">Version 3.0. Medieval Nordic Text Archive</title>
				<meeting><address><addrLine>Bergen</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<monogr>
		<ptr target="https://clarino.uib.no/menota/catalogue" />
		<title level="m">Menota Public Catalogue</title>
				<imprint>
			<date type="published" when="2020-12-02">2020/12/02</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Linking Corpus Data to an Excerpt-based Historical Dictionary</title>
		<author>
			<persName><forename type="first">T</forename><surname>Wills</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Johannsson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Battista</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the XVIII EURALEX International Congress: Lexicography in Global Contexts</title>
				<editor>
			<persName><forename type="first">J</forename><surname>Čibej</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">V</forename><surname>Gorjanc</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">I</forename><surname>Kosem</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Krek</surname></persName>
		</editor>
		<meeting>the XVIII EURALEX International Congress: Lexicography in Global Contexts<address><addrLine>Ljubljana; Ljubljana</addrLine></address></meeting>
		<imprint>
			<publisher>Ljubljana University Press</publisher>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="979" to="987" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">To gloss or not to gloss: An investigation of reading comprehension online</title>
		<author>
			<persName><forename type="first">L</forename><surname>Lomicka</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Language learning &amp; technology</title>
		<imprint>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="41" to="50" />
			<date type="published" when="1998">1998</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Computer-mediated glosses in second language reading comprehension and vocabulary learning: A meta-analysis</title>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">B</forename><surname>Abraham</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Computer Assisted Language Learning</title>
		<imprint>
			<biblScope unit="volume">21</biblScope>
			<biblScope unit="issue">3</biblScope>
			<biblScope unit="page" from="199" to="226" />
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
