<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">TALP at eHealth-KD Challenge 2020: Multi-Level Recurrent and Convolutional Neural Networks for Joint Classification of Key-Phrases and Relations</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Salvador</forename><surname>Medina</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">Universitat Politècnica de Catalunya</orgName>
								<address>
									<addrLine>Campus Nord, Carrer de Jordi Girona, 1, 3</addrLine>
									<postCode>08034</postCode>
									<settlement>Barcelona</settlement>
									<country key="ES">Spain</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Jordi</forename><surname>Turmo</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">Universitat Politècnica de Catalunya</orgName>
								<address>
									<addrLine>Campus Nord, Carrer de Jordi Girona, 1, 3</addrLine>
									<postCode>08034</postCode>
									<settlement>Barcelona</settlement>
									<country key="ES">Spain</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">TALP at eHealth-KD Challenge 2020: Multi-Level Recurrent and Convolutional Neural Networks for Joint Classification of Key-Phrases and Relations</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">CAD6B06F1F251B907A6910570CF132F7</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T04:19+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>NERC</term>
					<term>Relation Extraction</term>
					<term>eHealth NLP</term>
					<term>Contextual Embeddings</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>This article describes the model presented by the TALP Team to IberLEF's eHealth Knowledge Discovery 2020 shared task <ref type="bibr" target="#b0">[1]</ref>. The model iterates over the idea of using a single model for simultaneously identify key-phrases and their relationships. Taking into account the new transfer-learning sub-task presented for 2020's edition of eHealthKD, our model does not rely on any domain-specific knowledge nor handcrafted features. Our model was competitive in all four sub-tasks, ranking in 2nd, 3rd, 4th and 1st position respectively.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>This article describes the design choices and training strategy behind the model presented by the TALP team for IberLEF's eHealth Knowledge Discovery 2020 shared task <ref type="bibr" target="#b0">[1]</ref>. This shared task consists of identifying relevant key-phrases and relationships among them in Electronic Health documents from Spanish Medline. eHealthKD 2020's edition includes two significant additions respect to previous editions: an ensemble dataset created by combining predictions from previous 2019 edition's model outputs when applied to an unlabelled dataset, and a new transfer-learning sub-task.</p><p>Our model iterates over our team's 2019 model <ref type="bibr" target="#b1">[2]</ref> by leveraging several pre-trained word-level text representation models, as well as taking advantage of the automatically labelled corpus in a pre-training step. In particular, we make use of the pre-trained Word2Vec and FastText Medical Word Embedding for Spanish models <ref type="bibr" target="#b2">[3]</ref> from Barcelona Super-computing Center, which were trained using the SciELO database and a health-related subset of the Wikipedia. We use these two models to add context-specific knowledge to our model, which we believe was one of the shortcomings of the superseded model. However, the results suggest using these word representations does not represent an appreciable improvement over the general-purpose ones for Scenarios 1 to 3 and may even be detrimental for Scenario 4.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Proceedings of the Iberian Languages Evaluation Forum (IberLEF 2020)</head><p>email: smedina@cs.upc.edu (S. Medina); turmo@cs.upc.edu (J. Turmo) orcid: 0000-0003-2473-8571 (S. Medina); 0000-0002-7521-1115 (J. Turmo)</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">System Description</head><p>Our model expects a document and a source token index as input and generates a sequence of labels for each key-phrase and relation class. Input documents are parsed using FreeLing's dependency parser and each one of their tokens are encoded using either a BERT, a Word2Vec or a FastText pre-trained word-embedding model. The model then applies convolution filters to the encoded tokens of the input documents, combines the word-level filter's outputs of each input token and the specified source token with sentence-level embeddings of the documents, and outputs the boundaries of each key-phrase containing the source token as well as the likelihoods that every other token is the target of a relation having the specified source token's key-phrases as a source.</p><p>In order to generate all possible relations, the model should be run for every input token and have the all raw likelihoods combined across every one of them. This approach of looking at a single input token at a time is inspired by attention-based translation models such as the Transformer, in which the model comes up with the most likely output token one at a time, conditioned to the previously generated tokens and the whole untranslated document.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1.">Internal structure of the model</head><p>A visual representation of the model's structure is shown in Figure <ref type="figure" target="#fig_0">1</ref>. The network is composed of a set of shared intermediate layers and two independent output layers. The intermediate layers include a Bidirectional Gated Recurrent Unit layer followed by a set of convolution filters. The recurrent units' and convolution outputs are finally concatenated and fed to a fully connected layer. The output layers consist of a fully connected layer followed by a Conditional Random Field layer.</p><p>This structure lets the model look at both the local and global contexts of each of the input tokens. Particularly, the local context is captured by the recurrent units' output and the nonpooled convolution layer's output, while the global context is captured by the max-pooled convolution layer's output. Additional global context information is added when the BERT-based model is used by concatenating the encoding of the auxiliary CLS token.</p><p>The global context information and the target token's local context information are added to all time-steps before being fed to the fully connected shared layer. The final outputs are then generated by a Conditional Random Field (CRF) layer. Output CRF layers have proven to improve the capabilities of GRU and LSTM networks in low-resource sequence tagging tasks <ref type="bibr" target="#b3">[4]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.">Output generation and decoding</head><p>As described in Section 2, our system receives the sequence of tokens of a document and a token's index and outputs the bounds of the innermost key-phrase to which the token belongs. These bounds are encoded and decoded by assigning a Begin, Inside, Unitary and End tag to each token included in that key-phrase and Out to every other token (BIOUE-tag). One limitation of this approach is the fact that just one key-phrase is decoded for each token index, but this is not an issue in our case, as key-phrases may subsume but not overlap other key-phrases.</p><p>For each input token, our model outputs the list of relations' probabilities between the innermost entity to which the token belongs and each one of the tokens in the document is</p><formula xml:id="formula_0">LSTM LSTM R n X n LSTM LSTM R n+1 X n+1 LSTM LSTM R n+2 X n+2 LSTM LSTM R n+3 X n+3 C n C n+1 C n+2 C n+3 + M + + + + O n O n+1 O n+2 O n+3 FC FC FC FC Y c n Y c n+1 Y c n+2 Y c n+3</formula><p>Max-pooling predicted. Note that for the source token, we only consider the innermost entity whereas for the target tokens we consider all parent entities. Consequently, our method does not allow for overlapping relations from the same source token. This restriction is imposed so that the encoded sequence is not ambiguous. A visual representation of relations' probability predictions is shown in Figure <ref type="figure" target="#fig_1">2</ref>. Relations are predicted from the target key-phrase if the aggregated score  inside a key-phrase span surpasses a threshold. Only the key-phrase with the highest score is selected if multiple key-phrases overlap.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.3.">Input features</head><p>As previously mentioned, our model process the documents at the token level. We represent each token by a vector, which results from the concatenation of the features listed below:</p><p>• One-Hot encoding of the category and type fields of the token's Part-of-Speech Tag from FreeLing's tag-set.</p><p>• Normalized vector encoding the dependencies found in the path between the token and the target token (the one that is being decoded). It is computed by adding the onehot encoding representation of the dependency class for each hop in the dependency path and normalizing the resulting vector, not considering its direction. For instance, the representation of the token "I" in "I eat fish" when the target token is "fish" would be a vector with √ 2 in the positions corresponding to "subj" (subject) and "cd" (direct complement); whereas for "eat" it would be a vector with a single 1 in the "cd" position.</p><p>• One-Hot encoding of the distance between the token and the target token.</p><p>• Word-embedding of the token. We consider 4 alternative pre-trained word embedding models:</p><p>-Concatenation of the last output layers of a multi-language general-purpose BERT <ref type="bibr" target="#b4">[5]</ref> model <ref type="foot" target="#foot_0">1</ref> with no fine tuning.</p><p>-Word2Vec and FastText Medical Word Embedding for Spanish models from Barcelona Super-computing Center<ref type="foot" target="#foot_1">2</ref>  <ref type="bibr" target="#b2">[3]</ref>.</p><p>-FastText Spanish Unannotated Corpora from SUC 3 [6]</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.4.">Pre-training with the ensemble corpus</head><p>Due to the comparatively large number of parameters in our model respect to the size of the training dataset, overfitting can be an issue. We prevent this by using the relatively larger but inaccurate ensemble in a pre-training phase. In order not to let our model's variables fall into local minima that would make our model mimic previous' years models, we randomly add documents from IberLEF 2020's training corpus. Furthermore, we increase dropout and gradually decrease the learning rate for the training and fine-tuning training steps.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.5.">Single-scenario training and fine-tuning</head><p>In the general evaluation scenario, the loss function has to balance accuracy for both the keyphrase recognition and relation extraction tasks. This may be problematic, as the parameter updates made by the optimizer to improve one task might be detrimental for the other task. However, in evaluation scenarios 2 and 3, that is, independent key-phrase recognition and relation extraction tasks, the model does not have to generate both outputs. Consequently, on the one hand, we can use an uncompromising loss function. On the other hand, this means not being able to exploit the correlation between tasks, so it might as well lead to worse performance.</p><p>To study this effect, we suggest thee different single-scenario training strategies: using the general model with no alteration whatsoever, fine-tuning the general model's outputs with independent loss function for a few epochs, or training the specific model from scratch. Note that in the case of scenario 3, we decode the key-phrases using the gold truth rather than the model's output for all three strategies; and concatenate a one-hot-encoding of the key-phrase labels to the input for the from-scratch strategy. Table <ref type="table">2</ref> shows the results for all three single-scenario training strategies.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.6.">Trainable parameters and computational resources</head><p>All models were trained using the TensorFlow® 1.15 framework for Python® 3.6 in an 8 core Intel® Xeon® E5-2620 v4 CPU at 2.10GHz, 16GB of DDR4 RAM, a GeForce® GTX 1070 GPU  <ref type="table">1</ref> Final evaluation results of IberLEF2020's eHealth-KD scenarios 1 and 4. We also include the evaluation for the context-specific corpora. BERT FT is fine-tuned using the the transfer-learning development corpus.</p><p>and a 7200rpm 1TB Seagate® HDD.</p><p>BERT-based and Word2Vec/FastText-based models were trained for a total of 128 and 96 epochs respectively, divided among the pre-training, training and fine-tuning steps. Training epochs were evenly distributed between pre-training and training steps for models with no finetuning. When fine-tuning was applied (transfer-learning or single-task scenarios), pre-training was shortened by 16 epochs.</p><p>For each word representation model, independent models were trained with 8, 32 and 64 convolution filters of sizes 3 and 5; and 8, 32 and 64 single-layer recurrent units.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Results</head><p>Tables <ref type="table">1 and 2</ref> show the results for all four scenarios of IberLEF's 2020 eHealth Knowledge Discovery shared task for the best scoring submissions. Our model is competitive in all the scenarios, landing in second, third, fourth and first position respectively.</p><p>The scores obtained when using the context-specific Word2Vec (W2V Health), FastText (FastText Health) and general-purpose FastText (FastText General) models are shown in the bottom half of Table <ref type="table">1</ref>. With these models, see a drop of 0.04 in 𝐹 1 score in scenario 1 respect to the BERT-based model and a 0.13 in scenario 4. The difference in score for scenario 1 between the context-specific and general-purpose models is negligible, whilst we see a drop of 0.01 in 𝐹 1 for scenario 4.</p><p>Table <ref type="table">1</ref> also shows the independent results of our BERT-based model when fine-tuned with the transfer-learning development corpus. The fine-tuned model sees an increase of 0.08 in 𝐹 1 score for Scenario 4 while only seeing a 0.03 drop in Scenario 1.</p><p>As it can be seen in Table <ref type="table">2</ref>, our model is least competitive in scenario 3, in which it is outperformed by IXA-NER-RE's model by a margin of almost 0.06 in 𝐹 1 score. However, this score is matched the unsubmitted from-scratch model described in Section 2.5. Similarly, the general model strategy is comparable to SINAI's model in scenario 2.  <ref type="table">2</ref> Final evaluation results of IberLEF2020's eHealth-KD Scenarios 2 and 3. We also include the evaluation for the two additional single-scenario training strategies, which were not submitted.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Discussion</head><p>The joint key-phrase classification and relation extraction model presented by our team for the previous edition of IberLEF's eHealth Knowledge Discovery shared task outperformed every other participant model by a wide margin. This confirmed our belief that a joint model has the potential to exploit the mutual information between the two tasks and provide better evaluation results than traditional step-by-step architecture. The improvement was, however, less appreciable for the key-phrase classification task.</p><p>After comparing our model to the rest of the participant's submissions, we hypothesised that one of the main shortcomings of ours was the absolute lack of context-specific knowledge. For this year's edition, we decided to explore different alternatives to tackle this. But since a new transfer learning scenario was added, whose evaluation score would probably be compromised if the source model relied too heavily upon context-specific features, we opted for adding this context-specific information in a way that would not significantly alter the model's structure nor make it less general with handcrafted rules. Particularly, we opted for swapping the general-purpose word representation model by a health-specific one.</p><p>Unfortunately, the results show that the use of context-specific word embeddings does not substantially improve upon general-purpose embeddings and even leads to worse results in the transfer-learning scenario. Not only that, but we have also shown that contextual word embeddings such as BERT and XLNet significantly outperform predictive word embedding models such as Word2Vec and FastText. Moreover, the concatenation of this second word representation does not seem to provide any additional information over the original, whilst it makes the model more complex in terms of the number of trainable parameters.</p><p>Several hypotheses may explain these unsatisfactory results. First, we argue that although the documents' language register is formal, the use of technical terms is limited. Similarly, relation classes and specially key-phrase categories are arguably general, as pointed out by the results obtained in Scenario 4. Secondly, predictive word embedding models may not be able to capture the medical terms' semantic information to a degree that can be used by our model, but rather more explicit features may be preferable.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Conclusions</head><p>In this article, we have described the main characteristics of the model that we have developed for TALP team's submission to IberLEF's 2020 eHealth Knowledge Discovery shared task. Our model follows the trend started by our team's 2018's model, which consists of using a single network with shared weights that jointly performs the key-phrase recognition and relation extraction tasks to leverage the mutual information between the two. It has proven to be competitive against the other participants model's, especially in the general and transferlearning scenarios, ranking in second and first position respectively. The transfer-learning scenario particularly highlights the adaptability and context-independence of our model.</p><p>Three main improvements were made over the previous year's model: adaptive learning-rate for pre-training, single scenario fine-tuning and context-specific word vector representations. The last of which has been rather underwhelming though, and we conclude that adding contextspecific information to our model is still an unresolved issue.</p><p>Besides the aforementioned limitation, we see other shortcomings to our model that still need to be tackled to more accurately capture the mutual information between the two knowledge discovery tasks. Among these improvements, we would like to point out two that we believe are more promising:</p><p>• Use a trainable combination function for the outputs generated by the model for different source tokens in a document. Our current model, on the other hand, uses a simple union operation to join the predictions for the different tokens of single key-phrase.</p><p>• Use of fine-tuned context-specific contextual word embedding model. The use of contextspecific predictive word embeddings have proven not successful for our model, but general-purpose contextual word embeddings can be fine-tuned with context-specific unlabelled corpora.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Schematic architecture of the identification artificial neural network</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: Visual representation of how relations and key-phrases are encoded by the network.</figDesc><graphic coords="4,98.65,73.91,925.96,375.40" type="bitmap" /></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">We used the BERT-Base, Multilingual Cased model (104 languages, 12-layer, 768-hidden, 12-heads, 110M parameters) from the authors' repository (https://github.com/google-research/bert)</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_1"><ref type="bibr" target="#b1">2</ref> We used April 15, 2020's SciELO +</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="300" xml:id="foot_2">Wikipedia, 300 dimensions version of Medical Word Embedding for Spanish, which can be downloaded from https://zenodo.org/record/3744326<ref type="bibr" target="#b2">3</ref> We used the 300 dimensions sub-word binary model from https://github.com/dccuchile/ spanish-word-embeddings/blob/master/emb-from-suc.md</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>This contribution has been partially funded by the Spanish Ministry of Economy (MINECO) and the European Union (TIN2016-77820-C3-3-R and AEI/ FEDER,UE).</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Overview of the eHealth Knowledge Discovery Challenge at IberLEF</title>
		<author>
			<persName><forename type="first">A</forename><surname>Piad-Morffis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Gutiérrez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Cañizares-Diaz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Estevez-Velarde</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Almeida-Cruz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Muñoz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Montoyo</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Iberian Languages Evaluation Forum co-located with 36th Conference of the Spanish Society for Natural Language Processing, IberLEF@SEPLN 2020</title>
				<meeting>the Iberian Languages Evaluation Forum co-located with 36th Conference of the Spanish Society for Natural Language Processing, IberLEF@SEPLN 2020<address><addrLine>Spain</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2020-09">2020. September, 2020. 2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Talp-upc at ehealth-kd challenge 2019: A joint model with contextual embeddings for clinical information extraction</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">Medina</forename><surname>Herrera</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">Turmo</forename><surname>Borras</surname></persName>
		</author>
		<idno>CEUR-WS. org</idno>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Iberian Languages Evaluation Forum (IberLEF 2019): co-located with 35th Conference of the Spanish Society for Natural Language Processing (SEPLN 2019)</title>
				<meeting>the Iberian Languages Evaluation Forum (IberLEF 2019): co-located with 35th Conference of the Spanish Society for Natural Language Processing (SEPLN 2019)<address><addrLine>Bilbao, Spain</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2019-09-24">September 24th, 2019. 2019</date>
			<biblScope unit="page" from="78" to="84" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Medical word embeddings for Spanish: Development and evaluation</title>
		<author>
			<persName><forename type="first">F</forename><surname>Soares</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Villegas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Gonzalez-Agirre</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Krallinger</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Armengol-Estapé</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/W19-1916</idno>
		<ptr target="https://www.aclweb.org/anthology/W19-1916.doi:10.18653/v1/W19-1916" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2nd Clinical Natural Language Processing Workshop, Association for Computational Linguistics</title>
				<meeting>the 2nd Clinical Natural Language Processing Workshop, Association for Computational Linguistics<address><addrLine>Minneapolis, Minnesota, USA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="124" to="133" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<monogr>
		<author>
			<persName><forename type="first">Z</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Yu</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1508.01991</idno>
		<title level="m">Bidirectional lstm-crf models for sequence tagging</title>
				<imprint>
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b4">
	<monogr>
		<author>
			<persName><forename type="first">J</forename><surname>Devlin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M.-W</forename><surname>Chang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Toutanova</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1810.04805</idno>
		<title level="m">Bert: Pre-training of deep bidirectional transformers for language understanding</title>
				<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Enriching word vectors with subword information</title>
		<author>
			<persName><forename type="first">P</forename><surname>Bojanowski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Grave</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Joulin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Mikolov</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Transactions of the Association for Computational Linguistics</title>
		<imprint>
			<biblScope unit="volume">5</biblScope>
			<biblScope unit="page" from="135" to="146" />
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
