<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">LIP6@CLEF2017: Multi-Modal Spatial Role Labeling using Word Embeddings Working notes</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Éloi</forename><surname>Zablocki</surname></persName>
							<email>eloi.zablocki@lip6.fr</email>
						</author>
						<author>
							<persName><forename type="first">Patrick</forename><surname>Bordes</surname></persName>
							<email>patrick.bordes@lip6.fr</email>
						</author>
						<author>
							<persName><forename type="first">Laure</forename><surname>Soulier</surname></persName>
							<email>laure.soulier@lip6.fr</email>
						</author>
						<author>
							<persName><forename type="first">Benjamin</forename><surname>Piwowarski</surname></persName>
							<email>benjamin.piwowarski@lip6.fr</email>
						</author>
						<author>
							<persName><forename type="first">Patrick</forename><surname>Gallinari</surname></persName>
							<email>patrick.gallinari@lip6.fr</email>
						</author>
						<author>
							<affiliation key="aff0">
								<orgName type="department">Sorbonne Universités</orgName>
								<orgName type="laboratory">UMR 7606</orgName>
								<orgName type="institution">UPMC Univ Paris</orgName>
								<address>
									<addrLine>06</addrLine>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff1">
								<orgName type="institution">CNRS</orgName>
								<address>
									<addrLine>LIP6</addrLine>
									<postCode>F-75005</postCode>
									<settlement>Paris</settlement>
									<country key="FR">France</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">LIP6@CLEF2017: Multi-Modal Spatial Role Labeling using Word Embeddings Working notes</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">897CC83A00840AEFC1EF2F557FBB626F</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T20:27+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>multi-modal spatial role labeling</term>
					<term>linear SVM</term>
					<term>multi-label classification</term>
					<term>spatial indicator</term>
					<term>landmark</term>
					<term>trajector</term>
					<term>word embedding</term>
					<term>RCC8 regions</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>We report our participation to the multi-modal Spatial Role Labeling (mSpRL) lab at CLEF 2017. The task consists in extracting and classifying spatial relationships from textual data and associated images. Our approach focuses on the classification part as we use a baseline system for the extraction of the relations: we train a linear Support Vector Machine (SVM) model to classify hand-crafted vectors representing spatial relations. We present the obtained experiments and discuss also the effect of model parameters. Finally, we conclude the paper and introduce ideas for future developments.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>In this paper, we report our participation to the multi-modal Spatial Role Labeling lab <ref type="bibr" target="#b7">[8]</ref> at CLEF 2017 <ref type="bibr" target="#b3">[4]</ref>. The task consists in extracting and classifying spatial relationships from textual data and associated images.</p><p>The mSpRL goal is composed of three successive sub-tasks. The first one aims at extracting spatially-related entities and annotating the text using the following labels: Trajector, Spatial Indicator and Landmark. The second one consists in associating the previously found entities into spatial relation triplets r = (trajector, spatial indicator, landmark). The goal of the third sub-task is to classify those relations. While the two first sub-tasks could be seen as a linguistic conceptual representation (spatial role labeling), the third sub-task rather refers to a formal semantic representation of relations (spatial qualitative labeling). Possible labels for the relation classification are divided in three general types:</p><p>-Region RCC8 <ref type="bibr" target="#b9">[10]</ref> (8 possible values): disconnected (DC), externally connected (EC), equal (EQ), partially overlapping (PO), tangential proper part (TPP), tangential proper part inverse (TPPi), non-tangential proper part (NTPP), non-tangential proper part inverse (NTPPi).</p><p>-Direction (6 possible values): left, right, above, below, behind, front -Distance (5 possible values): middle, fast, close, far, near</p><p>The mSpRL task of CLEF 2017 is built upon SemEval 2012 task 3 <ref type="bibr" target="#b4">[5]</ref> which was proposed several years ago. That task has been augmented with images paired with text as additional inputs and sub-task 3 (relation classification).</p><p>The training set consists in 275 images, 600 associated sentences (several sentences can be linked to a single image) and a total of 761 relations. Figure <ref type="figure" target="#fig_0">1</ref> shows an example of the task. In qualitative spatial representation and reasoning, spatial relations can be classified precisely (for example, RCC8 is a set of topological relations between regions). Identifying spatial relations using text only is a difficult task, due to the variety of meaning and interpretations words and sentences can have. Exploiting visual data could be paramount to recognize the spatial objects and their relations but gives rise to multi-modal alignment issues. In the dataset, images are segmented into annotated bounded boxes, and the spatial relations between these boxes is given. This spatial information enables us to enrich the textual data.</p><p>In our contribution, the extraction of spatial roles and relationships (subtasks 1 and 2) was done using the winning system of previous years <ref type="bibr" target="#b10">[11]</ref>, an implementation of this model being in a software called Saul <ref type="bibr" target="#b6">[7]</ref>. It considers text solely as input (ignoring images) and returns a set of trajector, spatial indicator and landmark grouped into relations. Our contribution focuses on two aspects: using provided images as a complementary source of information and the relation classification sub-task. To do so, we handcraft a representation for spatial relations as a vector built using multi-modal inputs: the textual triplet and features from the associated image. We then train a SVM to classify the spatial relation. The SVM is trained to predict both general types (region, direction and distance labels) and specific values (EC, front, close, ...). Note that the multiple labels can be associated with a single relation (as it is the case in the example of Figure <ref type="figure" target="#fig_0">1</ref>).</p><p>The rest of this document is organized as follows: we describe our contribution including the classification pipeline and the design of relation embeddings. Finally, we present our experiments and the associated results.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Model</head><p>Our contribution mainly focuses on sub-task 3 since we used the previous state of the art model of <ref type="bibr" target="#b10">[11]</ref> -re-implemented in Saul <ref type="bibr" target="#b6">[7]</ref> -to perform sub-tasks 1 and 2. Sub-task 3 is a supervised classification problem in which the available input data is composed of 3 elements: the relation triplet, the original sentence from which the triplet was extracted, and an associated image. We convert that input data into a multi-modal embedding e relation that we describe in Section 2.1. We then use a linear SVM to classify the general types and specific values of the relations, the classification part being described in Section 2.2.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1">Relation Embedding</head><p>A relation is defined by visual (an image) and textual data (a triplet and the sentence from which it was extracted). We build our embedding by concatenating a textual embedding e text and an image embedding e image .</p><p>e relation = e text ⊕ e image Textual embedding. In our model, the text embedding contains information from the triplet only, as we drop the original sentence. Indeed, we assume that the useful information for the classification sub-task is contained in the extracted triplet, and that using the surrounding context of the sentence would lead to over-fitting of the model, given the small size of the training data.</p><p>We construct e text as follows:</p><formula xml:id="formula_0">e text = u trajector ⊕ u landmark ⊕ 1 spatial indicator</formula><p>where u is the average of the pre-trained embedding of the words that compose . In our experiments, we consider both Glove embeddings <ref type="bibr" target="#b8">[9]</ref> and multi-modal word embedding described in <ref type="bibr" target="#b0">[1]</ref>. 1 is a one-hot encoding of the spatial-indicator of the relation ; we use a fixed lexicon of 77 spatial-indicators as they are limited in number. ⊕ denotes the concatenation operator. Given the small amount of training data, ideally, e text has a small dimension to prevent over-fitting. With that objective in mind, we project word embeddings in a space of reduced dimensions ; we consider both random projections and Principal Component Analysis (PCA) techniques conducted on the training database.</p><p>Visual embedding. Segmented images and pre-computed visual features are provided in the dataset. A label is provided for each region of the segmented image. The given visual feature of a region is a 27-dimensional vector containing lowlevel information such as region area, width and height of the region, mean and standard deviation of height and width in the x and y axis respectively, the boundary/area ratio, convexity, average, standard deviation and skewness in RGB and CIE-Lab color spaces. Segmenting the images was done manually and visual features of all the regions were computed in <ref type="bibr" target="#b2">[3]</ref> and are included in the dataset. We construct e image as follows:</p><formula xml:id="formula_1">e image = r trajector ⊕ r landmark ⊕ r spatial</formula><p>where r is the visual embedding of the region ; we find the matching region of a landmark or a trajector by taking the region annotated with the most similar word (i.e. we compute cosine similarity scores on word embeddings). In r spatial , we encode in a one-hot vector the connectivity relations between the landmark and trajector regions: adjacent/disjoint, beside/x-aligned, above/below/y-aligned. This information is also provided as input data. As with textual embedding, to avoid over-fitting of the SVM classifier, we project the r vectors in a space of smaller dimension.</p><p>Note that the relation embeddings are hand-crafted and they remain fixed during training. The main reason for that is to reduce over-fitting of the model given the small size of the training dataset.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2">Classification</head><p>Once the embeddings of spatial relations are built (as explained in section 2.1), we use linear Support Vector Machines for classification <ref type="bibr" target="#b1">[2]</ref>, according to three strategies as shown in Figure <ref type="figure" target="#fig_1">2:</ref> -Mono-label: predicts a single label corresponding to specific values. Multilabels are considered as distinct classes. This gives a total of 28 classes (we remove multi-label classes that do not occur in the training set). The general type is simply deduced from the specific values. -Multi-label: predicts multiple labels corresponding to the specific values.</p><p>We use the One-vs-Rest (OvR) strategy to do so, which gives a total of 19 classes. The general type is simply deduced from the specific values. -Hierarchical Multi-label: first predicts multiple labels corresponding to the general types, then uses appropriate classifiers (each one trained on a particular general type) to predict multiple labels corresponding to the specific values for each of the predicted general type. That gives us a total of 4 classifiers (one to determine the general types and one for each possible general type).   <ref type="table">1</ref>. Results of Saul's baseline on test data for spatial roles prediction. LCount stands for the count of labels in the gold data and PCount stands for the number of predicted labels Sub-Task 2. Identification of spatial relations. In order to classify spatial relations, two binary classifiers are trained on pairs of phrases: one takes as input Trajector-Spatial Indicator pairs, the other on considers Spatial Indicator-Landmark pairs. With the perceptron assigned to Spatial Indicators, the indicator candidates are found, and all possible role-indicator pairs are possible candidates for the binary classifiers trained earlier. At the end, pairs with common indicators are the final triplets. Results are presented in Table <ref type="table" target="#tab_1">2</ref>. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">Sub-task 3: Spatial Relation Classification</head><p>This sub-task is the main focus of our contribution since we were interested in using multi-modal embedding for classifying spatial relations. For this purpose, we run two different scenarios:</p><p>-Our submitted best model (no image) which uses the mono-label classification strategy. We have e relation = e text as e image is ignored. Word embeddings are projected in a space of dimension 25 with PCA (outperforming random projection). Glove embeddings are used as they outperformed multi-modal embeddings of <ref type="bibr" target="#b0">[1]</ref>. -Our submitted model with image is the same than the model without image but the relation embedding includes r spatial in addition: e relation = e text ⊕ r spatial</p><p>Overall results. Table <ref type="table">3</ref> presents the obtained results for sub-task 3 for our scenarios with respect to two baselines:</p><p>-Organizer's baseline: The features for the type classifiers are the concatenation of the features from sub-task 1 for each argument of the triplet. -Best model which is the same model as the one submitted without image but the word embeddings are not projected in a lower-dimensional space and stay unchanged.  <ref type="table">3</ref>. Global scores on test data for the specific value prediction task Please, note that all hyper-parameters are chosen with a 5-fold cross-validation computation on the training set. Our model, with its different settings reaches better scores (precision, recall and F1) than the baseline by a large margin. <ref type="bibr" target="#b5">[6]</ref> reports comparable and slightly better results with a 10-fold cross-validation. For more detailed results, we refer the reader to Appendix A.</p><p>Classification strategies. To refine these results, we also compare the different classification strategies, as shown in Table <ref type="table" target="#tab_3">4</ref>. Mono-label classification appears to work better than other strategies. Interestingly, hierarchical strategy gives a higher precision but worse recall (and overall worse F1 score). A joint model, coupling mono-label and hierarchical models, might lead to even better performance by taking advantage of both models (best precision for Hierarchical, best recall for Mono-label). Influence of the components of the embeddings. At a finer level, we also measure the influence of each component of the global relation embedding vector e relation with an ablation study. Instead of using full relation embeddings with all of their components, we remove one or several parts of them and we report in Table <ref type="table" target="#tab_4">5</ref> obtained results of classification models trained on those partial relation embeddings. Each line of the table contains the results of a 5-fold cross-validation training on e relation without the ablated part, namely image, text, spatial indicator, visual region embeddings. This experimentally highlights the importance of which spatial indicator is being used and the textual embeddings for the trajector and then landmark. Visual parts of the relation embedding are useless or even harmful for the overall performance. While it is unclear why using the visual embeddings slightly degrades the overall performance, we note that many images do not contain areas for the entities found by sub-task 1 in the sentence (for example "bench" is found as trajector in the sentence "a bench in a park" but the associated image might not contain any region labeled "bench"). Moreover, sometimes our algorithm misses the good region: for example sub-task 1 gives the entity "head" in the text but there is no "head" region in the image but rather "face-of-person". Despite some handcrafted rules that we have added to account for this problem, lots of regions are not considered. Also, even though high-level features from the images are provided, we assume that there are not enough images in the training set to learn something complementary from the text. Eventually, since the classification labels of sub-task 3 are only gold labels prone to annotation subjectivity, a human annotator would not get 100% f1-score. It would then be interesting to know about human performance on that sub-task for a comparison with our results.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Classif. type Precision</head><p>Word embedding dimension influence. Since textual embeddings are proved to be major components of the spatial relation representation, we evaluate the impact of the choice of the dimension of the space in which word embeddings are projected with PCA. We make that parameter vary while others are kept fixed, in the experiment reported in Table <ref type="table">3</ref>. We can see that increasing the word embedding dimension improves the effectiveness of our approach and our best performing model does not project word embeddings and keep them unchanged. Intuitively, having too high-dimension embeddings leads to more parameters and a higher risk in over-fitting the model. For a good trade-off between performance and limited size for relation embeddings, 50 is also a suitable choice. In this work, we focused on sub-task 3 of the mSpRL lab of CLEF 2017: predicting general types and specific values for relations. Our system relies on a baseline to extract spatial roles and relations from raw textual data. We build fixed embeddings for spatial triplets, and a linear SVM classifies relations. Unfortunately, we were not able to use provided visual inputs in a profitable way as our best model ignores images. These results highlight that considering multi-modal data for enhancing natural language processing is a difficult task and requires more efforts in terms of model design.</p><p>As future work, we have two objectives. First, we want to use the image data for sub-tasks 1 and 2 in an end-to-end fashion, as visual information might be useful to disambiguate between several candidate relations. Our other goal aims at addressing the problem of the limited quantity of training data: we wish to explore transfer learning techniques to train spatial word embeddings on auxiliary tasks.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Fig. 1 .</head><label>1</label><figDesc>Fig. 1. Overview of the mSpRL task</figDesc><graphic coords="2,169.35,246.52,276.67,131.98" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Fig. 2 .</head><label>2</label><figDesc>Fig. 2. Overview of the considered classification strategies</figDesc><graphic coords="5,186.64,115.83,242.08,127.39" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Fig. 3 .</head><label>3</label><figDesc>Fig. 3. Influence of the word embedding dimension on the F1 score. Vertical bars are the standard deviation of the results on 10 experiments, where each experiment is a 5-fold cross-validation</figDesc><graphic coords="8,169.35,398.40,276.65,138.33" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2 .</head><label>2</label><figDesc>Results of Saul's baseline on test data for spatial relations extraction</figDesc><table><row><cell cols="2">Precision Recall F1 LCount PCount</cell></row><row><cell>Overall 68.33 48.03 56.41 939</cell><cell>660</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 4 .</head><label>4</label><figDesc>Global score on test data for the specific value prediction task, using several classification techniques</figDesc><table><row><cell></cell><cell>Recall F1</cell></row><row><cell>Mono-label</cell><cell>58.66 41.39 46.29</cell></row><row><cell>Multi-label</cell><cell>62.14 39.49 45.32</cell></row><row><cell cols="2">Hierarchical 62.17 40.05 45.88</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Table 5 .</head><label>5</label><figDesc>Ablation study of the influence of the different parts of the relation embedding</figDesc><table><row><cell cols="2">Ablated part Precision Recall F1</cell></row><row><cell>None</cell><cell>87.26 87.32 85.57</cell></row><row><cell>eimage</cell><cell>87.60 88.49 86.50</cell></row><row><cell>etext</cell><cell>71.67 45.95 28.85</cell></row><row><cell>r spatial</cell><cell>87.04 87.62 85.95</cell></row><row><cell>r</cell><cell>85.76 86.51 84.70</cell></row><row><cell>1 indicator</cell><cell>75.80 75.69 72.16</cell></row><row><cell>u</cell><cell>74.72 73.03 70.34</cell></row></table></figure>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgements</head><p>This work is partially supported by the CHIST-ERA EU project MUSTER (http://www.chistera.eu/projects/muster) and the Labex SMART. We additionally thank the task organizers for their help in using the baseline for sub-tasks 1 and 2.</p></div>
			</div>

			<div type="annex">
<div xmlns="http://www.tei-c.org/ns/1.0" />			</div>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Imagined visual representations as multimodal embeddings</title>
		<author>
			<persName><forename type="first">G</forename><surname>Collell Talleda</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">F</forename><surname>Moens</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence (AAAI-17)</title>
				<meeting>the Thirty-First AAAI Conference on Artificial Intelligence (AAAI-17)</meeting>
		<imprint>
			<publisher>AAAI</publisher>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Support-vector networks</title>
		<author>
			<persName><forename type="first">C</forename><surname>Cortes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Vapnik</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Machine learning</title>
		<imprint>
			<biblScope unit="volume">20</biblScope>
			<biblScope unit="issue">3</biblScope>
			<biblScope unit="page" from="273" to="297" />
			<date type="published" when="1995">1995</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">The segmented and annotated iapr tc-12 benchmark</title>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">J</forename><surname>Escalante</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">A</forename><surname>Hernández</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">A</forename><surname>Gonzalez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>López-López</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Montes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">F</forename><surname>Morales</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">E</forename><surname>Sucar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Villaseñor</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Grubinger</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Computer Vision and Image Understanding</title>
		<imprint>
			<biblScope unit="volume">114</biblScope>
			<biblScope unit="issue">4</biblScope>
			<biblScope unit="page" from="419" to="428" />
			<date type="published" when="2010">2010</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
	</analytic>
	<monogr>
		<title level="m">Experimental IR Meets Multilinguality, Multimodality, and Interaction. 8th International Conference of the CLEF Association, CLEF 2017</title>
		<title level="s">Proceedings</title>
		<editor>
			<persName><forename type="first">J</forename><forename type="middle">F</forename><surname>Gareth</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Séamus</forename><surname>Jones</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><forename type="middle">G L K L G T M L C N F</forename><surname>Lawless</surname></persName>
		</editor>
		<meeting><address><addrLine>Dublin, Ireland</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2017">September 11-14, 2017</date>
			<biblScope unit="volume">10456</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Semeval-2012 task 3: Spatial role labeling</title>
		<author>
			<persName><forename type="first">P</forename><surname>Kordjamshidi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Bethard</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">F</forename><surname>Moens</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the First Joint Conference on Lexical and Computational Semantics-Volume 1: Proceedings of the main conference and the shared task</title>
				<meeting>the First Joint Conference on Lexical and Computational Semantics-Volume 1: the main conference and the shared task</meeting>
		<imprint>
			<date type="published" when="2012">2012</date>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="page" from="365" to="373" />
		</imprint>
	</monogr>
	<note>Association for Computational Linguistics</note>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Global machine learning for spatial ontology population</title>
		<author>
			<persName><forename type="first">P</forename><surname>Kordjamshidi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">F</forename><surname>Moens</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Web Semantics: Science, Services and Agents on the World Wide Web</title>
		<imprint>
			<biblScope unit="volume">30</biblScope>
			<biblScope unit="page" from="3" to="21" />
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Saul: Towards declarative learning based programming</title>
		<author>
			<persName><forename type="first">P</forename><surname>Kordjamshidi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Roth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Wu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IJCAI: proceedings of the conference/sponsored by the International Joint Conferences on Artificial Intelligence</title>
				<imprint>
			<publisher>NIH Public Access</publisher>
			<date type="published" when="2015">2015</date>
			<biblScope unit="volume">2015</biblScope>
			<biblScope unit="page">1844</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<monogr>
		<title level="m">CLEF 2017 Labs Working Notes</title>
				<editor>
			<persName><forename type="first">Linda</forename><surname>Cappellato</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Nicola</forename><surname>Ferro</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">L</forename><forename type="middle">G</forename><surname>Mandl</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">T</forename></persName>
		</editor>
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Glove: Global vectors for word representation</title>
		<author>
			<persName><forename type="first">J</forename><surname>Pennington</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Socher</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">D</forename><surname>Manning</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">EMNLP</title>
		<imprint>
			<biblScope unit="volume">14</biblScope>
			<biblScope unit="page" from="1532" to="1543" />
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">A spatial logic based on regions and connection</title>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">A</forename><surname>Randell</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Cui</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">G</forename><surname>Cohn</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">KR</title>
		<imprint>
			<biblScope unit="volume">92</biblScope>
			<biblScope unit="page" from="165" to="176" />
			<date type="published" when="1992">1992</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Utd-sprl: A joint approach to spatial role labeling</title>
		<author>
			<persName><forename type="first">K</forename><surname>Roberts</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">M</forename><surname>Harabagiu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the First Joint Conference on Lexical and Computational Semantics-Volume 1: Proceedings of the main conference and the shared task</title>
				<meeting>the First Joint Conference on Lexical and Computational Semantics-Volume 1: the main conference and the shared task</meeting>
		<imprint>
			<date type="published" when="2012">2012</date>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="page" from="419" to="424" />
		</imprint>
	</monogr>
	<note>Association for Computational Linguistics</note>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
