<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Monolingual and Cross-lingual Ontology Matching with CIDER-CL: evaluation report for OAEI 2013</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Jorge</forename><surname>Gracia</surname></persName>
							<email>jgracia@fi.upm.es</email>
							<affiliation key="aff0">
								<orgName type="laboratory">Ontology Engineering Group</orgName>
								<orgName type="institution">Universidad Politécnica de Madrid</orgName>
								<address>
									<country key="ES">Spain</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Kartik</forename><surname>Asooja</surname></persName>
							<email>kartik.asooja@deri.org</email>
							<affiliation key="aff0">
								<orgName type="laboratory">Ontology Engineering Group</orgName>
								<orgName type="institution">Universidad Politécnica de Madrid</orgName>
								<address>
									<country key="ES">Spain</country>
								</address>
							</affiliation>
							<affiliation key="aff1">
								<orgName type="department">Digital Enterprise Research Institute</orgName>
								<orgName type="institution">National University of Ireland</orgName>
								<address>
									<settlement>Galway</settlement>
									<country key="IE">Ireland</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Monolingual and Cross-lingual Ontology Matching with CIDER-CL: evaluation report for OAEI 2013</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">8C1626BC4ECDD438E1E81CD6B445B9DB</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T07:43+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>CIDER-CL is the evolution of CIDER, a schema-based ontology alignment system. Its algorithm compares each pair of ontology entities by analysing their similarity at different levels of their ontological context (linguistic description, superterms, subterms, related terms, etc.). Then, such elementary similarities are combined by means of artificial neural networks. In its current version, CIDER-CL uses SoftTFIDF for monolingual comparisons and Cross-Lingual Explicit Semantic Analysis for comparisons between entities documented in different natural languages. In this paper we briefly describe CIDER-CL and comment its results at the Ontology Alignment Evaluation Initiative 2013 campaign (OAEI'13).</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>1 Presentation of the system CIDER-CL is the evolution of CIDER (Context and Inference baseD alignER) <ref type="bibr" target="#b7">[7]</ref>, now incorporating cross-lingual capabilities. In order to match ontology entities, CIDER-CL extracts their ontological context and enriches it by applying lightweight inference rules. Then, elementary similarity comparisons are performed to compare different features of the ontological contexts. Such elementary comparisons are combined by means of artificial neural networks (ANNs) <ref type="bibr" target="#b9">[9]</ref> to produce a final similarity value between the compared entities. The use of ANNs saves a lot of effort on manual tuning and allows to quickly adapt the system into different domains (as far as there are reference alignments available for them).</p><p>In its current version, the aligner has been re-implemented to include more features in the comparisons and to add new metrics for similarity computation. In particular, cross-lingual capabilities have been added by including the use of Cross-Lingual Explicit Semantic Analysis (CL-ESA) <ref type="bibr" target="#b10">[10]</ref> between entities documented in different natural languages. Further, the previous metrics for monolingual comparison (based on Vector Space Modelling <ref type="bibr" target="#b8">[8]</ref>) have been changed by the SoftTFIDF metric <ref type="bibr" target="#b2">[3]</ref>. CIDER-CL is not intended to be used with large ontologies, particularly in the cross-lingual case (CL-ESA computation is quite costly in terms of time).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.1">State, purpose, general statement</head><p>According to the high level classification given in <ref type="bibr" target="#b5">[5]</ref>, our method is a schema-based system (opposite to others which are instance-based, or mixed), because it relies mostly on schema-level input information for performing ontology matching. CIDER-CL can operate in two modes: (i) as an ontology aligner, taking two ontologies as input and giving their alignment as output, and (ii) as a similarity service, taking two ontology entities as input and giving the similarity value between them as output. In the first case the input to CIDER-CL are two OWL ontologies and a threshold value and the output is an RDF file expressed in the alignment format <ref type="foot" target="#foot_0">3</ref> , although it can be easily translated into another formats such as EDOAL <ref type="foot" target="#foot_1">4</ref> .</p><p>The type of alignment that CIDER-CL obtains is semantic equivalence. In its current implementation the following languages are covered: English (EN), Spanish (ES), German (DE), and Dutch (NL).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.2">Specific techniques used</head><p>In this section we briefly introduce the monolingual and cross-lingual metrics used by CIDER-CL, as well as the overall architecture of the ontology aligner.</p><p>SoftTFIDF. SoftTFIDF <ref type="bibr" target="#b2">[3]</ref> is a hybrid string similarity measure that combines TF-IDF, a token-based similarity widely used in information retrieval <ref type="bibr" target="#b8">[8]</ref>, with an edit-based similarity such as Jaro-Winkler <ref type="bibr" target="#b11">[11]</ref> (although any other could be used instead).</p><p>Typically, string comparisons to compute TF-IDF weights are based on exact matching (after some normalisation or tokenisation step). The idea of SoftTFIDF is to use an edit distance instead to support a higher degree of variation between the terms. In particular, we use Jaro-Winkler similarity with a 0.9 threshold, above which two strings are consider equal. SoftTFIDF measure has proved to be very effective when comparing short strings <ref type="bibr" target="#b2">[3]</ref>. In our case, the corpus used by SoftTFIDF is dynamically created with the lexical information coming from the two compared ontologies (extracting their labels, comments, and URI fragments).</p><p>CL-ESA. For cross-lingual ontology matching we propose the use of CL-ESA <ref type="bibr" target="#b10">[10]</ref>, a cross-lingual extension of an approach called Explicit Semantic Analysis <ref type="bibr" target="#b6">[6]</ref> (ESA). ESA allows comparing two texts semantically with the help of explicitly defined concepts. This method uses the co-occurrence information of the words from the textual definitions of the concepts using, for instance, the Wikipedia articles. In short, ESA extends a simple bag of words model to a bag of concepts model. Some reports <ref type="bibr" target="#b1">[2]</ref> have demonstrated the good behaviour of CL-ESA for certain tasks such as cross lingual information retrieval.</p><p>To compare two texts in different languages semantically, Wikipedia-based CL-ESA represents the two texts as vectors in a vector space that has the Wikipedia titles (articles) as dimensions, each vector in its own language specific Wikipedia. The magnitude of each title/dimension is the associativity weight of the text to that title. To quantify this associativity, the textual content of the Wikipedia article is utilized. This weight can be calculated by using different methods, for instance, TF-IDF score.</p><p>For implementing CL-ESA, we followed an information retrieval-based approach by creating a Lucene inverted index of the Wikipedia extended abstracts that exist in all the considered languages i.e., EN, ES, NL, and DE. To create the weighted vector of concepts, the term is searched over the index of the respective languages to retrieve the top associated Wikipedia concepts and the Lucene ranking scores are taken as the associativity weights of the concepts to the term. We used DBpedia URIs <ref type="bibr" target="#b0">[1]</ref> as the pivot between cross-lingual Wikipedia spaces and to identify a Wikipedia concept no matter the language.</p><p>Scheme of the Aligner. Briefly explained, the alignment process is as follows (see Figure <ref type="figure" target="#fig_0">1</ref>): 1. First, the ontological context of each ontology term is extracted. This process is enriched by applying a lightweight inference mechanism <ref type="foot" target="#foot_2">5</ref> , in order to add more semantic information that is not explicit in the asserted ontologies. 2. Second, similarities are computed between different parts of the ontological context. In particular, ten different features are considered: labels, comments, equivalent terms, subterms, superterms, direct subterms, direct superterms (both for classes and properties) and properties, direct properties, and related classes (for classes) or domains, direct domains, and ranges (for properties). 3. Third, the different similarities are combined within an ANN to provide a final similarity degree. CIDER-CL uses four different neural networks (multilayer perceptrons in particular) for computing monolingual and cross-lingual similarities between classes and properties, respectively. 4. Finally, a matrix (M in Figure <ref type="figure" target="#fig_0">1</ref>) with all similarities is obtained. The final alignment (A) is then extracted from it, finding the highest rated one-to-one relationships among terms and filtering out the ones below the given threshold.</p><p>Implementation. Some datasets used in OAEI campaigns are open and the reference alignments available for download. We have used part of such data to train our system.</p><p>In particular, we chose a subset of the OAEI'11 benchmark track to train our neural networks for the monolingual case. We used the whole dataset but excluding cases 202 and 248-266, which present a total absence or randomization of labels and comments (however their variations, 248-2, 248-4, etc., were not excluded). Also the reference alignments of the conference track, which are also open, were added to the training data set. The use of the benchmark track for adjusting the ANNs is motivated by the fact that it covers many possible situations and variations well, such as presence or absence of certain ingredients (labels, comments, etc.) or the effect of aligning at different granularity levels (flattened/expanded hierarchies), etc. Further, we add also data of the conference track to include training data coming from "real world" ontologies.</p><p>For the cross-lingual case, we trained the neural networks with a subset of the ontologies of the OAEI'13 Multifarm track (in EN, ES, DE, and NL): cmt, conference, confOf, and sigkdd. Comparisons were run among the different ontologies in the different languages, excluding comparisons between the same ontologies. Due to the slow performance of CL-ESA, we decided to perform an attribute selection analysis to discover which features have more predictive power. As result, we limited the system to compute these features for classes: labels, subterms, direct superterms, direct subterms, and properties; while for properties they were limited to: labels, subterms, and ranges.</p><p>CIDER-CL has been developed in Java, extending the Alignment API <ref type="bibr" target="#b4">[4]</ref>. To create and manipulate neural networks we use Weka<ref type="foot" target="#foot_3">6</ref> data mining framework. For SoftTFIDF we use SecondString <ref type="foot" target="#foot_4">7</ref> and for CL-ESA we use the implementation developed by the Monnet project <ref type="foot" target="#foot_5">8</ref> , which is available in GitHub as open source <ref type="foot" target="#foot_6">9</ref> .</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.3">Adaptations made for the evaluation</head><p>The weights and the configuration of the neural networks remained constant for all the tests and tracks of OAEI'13, as well as the threshold. In particular we selected a threshold of 0.0025. The intention of such a small value was to promote recall over precision (while filtering out some extremely low values). Therefore, later filtering can be made to perform a threshold analysis as the organisers of some OAEI tracks do (e.g., conference track). Some minor technical adaptations were needed to integrate the system into the Seals platform, like solving compatibility issues with the libraries used by the Seals wrapper.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.4">Link to the system and parameters file</head><p>The version of CIDER-CL used for this evaluation (v1.1) was uploaded to the Seals platform: http://www.seals-project.eu/ . More information can be found at CIDER-CL's website http://www.oeg-upm.net/files/cider-cl .</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.5">Link to the set of provided alignments (in align format)</head><p>The resultant alignments will be provided by the Seals platform: http://www.sealsproject.eu/</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Results</head><p>For OAEI'13 campaign, CIDER-CL participated in all the Seals-based tracks <ref type="foot" target="#foot_7">10</ref> . In the following, we report the results of CIDER-CL for benchmark, conference, anatomy, and multifarm tracks. For the other tracks, the system was not fit for the type of evaluation (e.g., interactive track) or could not complete the task (e.g., library). Details about the test ontologies, the evaluation process, and the complete results for all tracks can be found at the OAEI'13 website <ref type="foot" target="#foot_8">11</ref> .</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1">Benchmark</head><p>This year, a blind test set was generated based on a seed ontology of the bibliographic domain. Out of the 21 systems participating in this track, CIDER-CL was within the three best systems in terms of F-measure. In particular, the obtained results were: Precision(P)=0.85, Recall(R)=0.67 and F-Measure(F)=0.75</p><p>Compare to the F=0.41 of edna, a simple edit distance-based baseline. In addition, confidence-weighted measures were also computed for those systems that provided a confidence value. In almost all cases the results were worse, as it was also the case of CIDER-CL: P=0.84, R=0.55, and F=0.66 Also the time spent in the evaluation was calculated. CIDER-CL took 844±19 seconds, which was slower than most of the systems (the median value was 173 sec) although still far from the slowest one (10241±347 sec).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2">Conference</head><p>In this track, several ontologies from the conference domain were matched, resulting in 21 alignments. In this case the organisers explored different thresholds and selected the best achievable results. This test is not blind and the participants have the reference alignments at their disposal before the evaluation phase.</p><p>Two reference alignments were used in this track: the original reference alignment (ra1) and its transitive closure (ra2). Two baselines (edna and string equivalence) were computed for comparison. Notice that the results for CIDER-CL in this track are merely illustrative and should not be taken as a proper test, due to the fact that part of the training data of its neural networks came from the conference track reference alignments (i.e., training and test data coincide partially).</p><p>Out of the 25 systems participating in this track (some of them were variations of the same system), CIDER-CL performance was close to the average. The results were: test ra1 (original): P = 0.75, R = 0.47, and F = 0.58 with threshold= 0.14 test ra2 (entailed): P = 0.72, R = 0.44, and F = 0.55 with threshold= 0.08 CIDER-CL was in the group of systems that performed better than the two baselines for ra2 and between the two baselines for ra1. The results for ra1 illustrates an improvement with respect to the results obtained by its previous version (CIDER v0.4) for the same test at OAEI'11 (F=0.53). The runtime was also registered: CIDER-CL took less than 10 minutes for computing the 21 alignments. The other systems ranged from 1 minute to more than 40.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.3">Anatomy</head><p>This year, the current version of CIDER-CL completed the task and gave results for the first time. In fact, in previous editions of OAEI, CIDER gave time-outs and the tool did not finish the task, due to the big size of the involved ontologies. The results are: P = 0.65, R = 0.73, F = 0.69, R+ = 0.31 These results are below the average of the overall results (F-Measure ranging from 0.41 to 0.94, with a median value of 0.81). An "extended recall" (R+) was also computed, that is, the amount of detected non-trivial correspondences (that do not have the same normalized label). For this metric CIDER-CL behaved better than the median value (0.23). In terms of running time, CIDER-CL was the third slowest system (12308 sec) in this track, after discarding those that gave time-out.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.4">Multifarm</head><p>This track is based on the alignment of ontologies in nine different languages: EN, DE, ES, NL, CZ, RU, PT, FR, and CN. All pairs of languages (36 pairs) were considered in the evaluation. A total of 900 matching tasks were performed. There were 21 participants in this track, 7 of them implementing specific cross-lingual modules as it was the case of CIDER-CL.</p><p>The organisers divided the results in two types: comparisons between different ontologies (type i) and comparisons between the same ontologies (type ii). The result summary published by the organisers aggregates the individual results for all the language pairs. In the case of CIDER-CL this hampers direct comparisons with other systems, owing to the fact that CIDER-CL only covers a subset of languages (EN, DE, ES, NL) and non produced alignments in other languages penalised the overall results. For this reason we have filtered the language specific results to consider only such subset of languages. The averaged results for CIDER-CL are: type i (different ontologies): P = 0.16, R = 0.19, F = 0.17 type ii (same ontologies): P = 0.82 , R = 0.16, F = 0.26 For type ii, CIDER-CL got the 4th best result overall in terms of F-Measure and the 3rd best result in the set of systems implementing specific cross-lingual techniques (the results for such systems ranged from F = 0.12 to F = 0.44 for the referred subset of languages). On the other hand, for type i CIDER-CL was in 8th position out of the 21 participants, although in the last place among the set of systems implementing crosslingual techniques (F-measure of the other techniques ranged from 0.17 to 0.35).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">General comments</head><p>The following subsections contain some remarks and comments about the results obtained and the evaluation process.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">Comments on the results</head><p>CIDER-CL obtained good results for the benchmark track (third place out of 21 participants). This shows that our system performs well for domains in which the system could be trained with available reference data. Also that SoftTFIDF is suitable for ontology matching. In contrast, the results for the anatomy track were relatively poor. This shows that creating a general purpose aligner based on our technique is not immediate. Adding more training data from other domains would help to solve this.</p><p>The results from the multilingual track are rather modest, but the fact that even the best systems scored low illustrates the difficulty of the problem. We consider that the use of CL-ESA is promising for cross-lingual matching, but it will require more study and adaptation to achieve better results.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">Discussions on the way to improve the proposed system</head><p>More reference alignments from "real world" ontologies will be used in the future for training the ANNs, in order to cover more domains and different types of ontologies. Regarding the cross-lingual matching, there is still room for continuing improving the use of CL-ESA to that end. We plan also to combine this novel technique with other ones such as machine translation.</p><p>Time response in CIDER-CL is still an issue and has to be further improved. In fact CIDER-CL works well with small and medium sized ontologies but not with large ones. Partitioning and other related techniques will be explored in order to solve this.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3">Comments on the OAEI 2013 test cases</head><p>The variety of tracks and the improvements introduced along the years makes the campaign very useful to test the performance of ontology aligners and analyse their strengths and weaknesses. Nevertheless, we miss blind tests cases in more tracks, which would allow a fair comparison between systems.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Conclusion</head><p>CIDER-CL is a schema-based alignment system that compares the ontological context of each pair of terms in the aligned ontologies. Several elementary comparisons are computed and combined by means of artificial neural networks. Monolingual and crosslingual metrics are used in the matching.</p><p>We have presented here some results of the participation of CIDER-CL at OAEI'13 campaign. The results vary depending on the track, from the good results in the benchmark track to the relatively limited behaviour in anatomy, for instance. We confirmed that the proposed technique, based on ANNs, is suitable in conjunction with SoftTFIDF metric for monolingual ontology matching. The use of CL-ESA metric for cross-lingual matching is promising but requires more study.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Fig. 1 .</head><label>1</label><figDesc>Fig. 1. Scheme of the matching process.</figDesc><graphic coords="3,134.77,275.47,345.84,126.29" type="bitmap" /></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_0">http://alignapi.gforge.inria.fr/format.html</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_1">http://alignapi.gforge.inria.fr/edoal.html</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_2">Typically transitive inference, although RDFS or more complex rules can be also applied, at the cost of processing time.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="6" xml:id="foot_3">http://www.cs.waikato.ac.nz/ml/weka/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="7" xml:id="foot_4">http://secondstring.sourceforge.net/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="8" xml:id="foot_5">http://www.monnet-project.eu/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="9" xml:id="foot_6">https://github.com/kasooja/clesa</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="10" xml:id="foot_7">http://oaei.ontologymatching.org/2013/seals-eval.html</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="11" xml:id="foot_8">http://oaei.ontologymatching.org/2013</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Acknowledgments. This work is supported by the Spanish national project BabeLData (TIN2010-17550) and the Spanish Ministry of Economy and Competitiveness within the Juan de la Cierva program.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">DBpedia -a crystallization point for the web of data</title>
		<author>
			<persName><forename type="first">C</forename><surname>Bizer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Lehmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Kobilarov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Auer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Becker</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Cyganiak</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Hellmann</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Web Semantics: Science, Services and Agents on the World Wide Web</title>
		<imprint>
			<biblScope unit="volume">7</biblScope>
			<biblScope unit="issue">3</biblScope>
			<biblScope unit="page" from="154" to="165" />
			<date type="published" when="2009-09">Sept. 2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Explicit versus latent concept models for cross-language information retrieval</title>
		<author>
			<persName><forename type="first">P</forename><surname>Cimiano</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Schultz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Sizov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Sorg</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Staab</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 21st international jont conference on Artifical intelligence, IJCAI&apos;09</title>
				<meeting>the 21st international jont conference on Artifical intelligence, IJCAI&apos;09<address><addrLine>San Francisco, CA, USA</addrLine></address></meeting>
		<imprint>
			<publisher>Morgan Kaufmann Publishers Inc</publisher>
			<date type="published" when="2009">2009</date>
			<biblScope unit="page" from="1513" to="1518" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">A comparison of string distance metrics for name-matching tasks</title>
		<author>
			<persName><forename type="first">W</forename><forename type="middle">W</forename><surname>Cohen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Ravikumar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">E</forename><surname>Fienberg</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. Workshop on Information Integration on the Web (IIWeb-03)</title>
				<meeting>Workshop on Information Integration on the Web (IIWeb-03)</meeting>
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<monogr>
		<title level="m">@ IJCAI-03</title>
				<meeting><address><addrLine>Acapulco, Mexico</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2003-08">Aug. 2003</date>
			<biblScope unit="page" from="73" to="78" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">An API for ontology alignment</title>
		<author>
			<persName><forename type="first">J</forename><surname>Euzenat</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">3rd International Semantic Web Conference (ISWC&apos;04)</title>
				<meeting><address><addrLine>Hiroshima (Japan</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2004-11">November 2004</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<monogr>
		<title level="m" type="main">Ontology matching</title>
		<author>
			<persName><forename type="first">J</forename><surname>Euzenat</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Shvaiko</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2007">2007</date>
			<publisher>Springer-Verlag</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Computing semantic relatedness using wikipedia-based explicit semantic analysis</title>
		<author>
			<persName><forename type="first">E</forename><surname>Gabrilovich</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Markovitch</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 20th International Joint Conference on Artificial Intelligence</title>
				<meeting>the 20th International Joint Conference on Artificial Intelligence</meeting>
		<imprint>
			<date type="published" when="2007">2007</date>
			<biblScope unit="page" from="1606" to="1611" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Ontology matching with CIDER: Evaluation report for OAEI</title>
		<author>
			<persName><forename type="first">J</forename><surname>Gracia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Bernad</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Mena</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of 6th Ontology Matching Workshop (OM&apos;11), at 10th International Semantic Web Conference (ISWC&apos;11)</title>
				<meeting>of 6th Ontology Matching Workshop (OM&apos;11), at 10th International Semantic Web Conference (ISWC&apos;11)<address><addrLine>Bonn (Germany</addrLine></address></meeting>
		<imprint>
			<publisher>CEUR-WS</publisher>
			<date type="published" when="2011-10">2011. Oct. 2011</date>
			<biblScope unit="volume">814</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">A critical analysis of vector space model for information retrieval</title>
		<author>
			<persName><forename type="first">V</forename><forename type="middle">V</forename><surname>Raghavan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">S K</forename><surname>Wong</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of the American Society for Information Science</title>
		<imprint>
			<biblScope unit="volume">37</biblScope>
			<biblScope unit="issue">5</biblScope>
			<biblScope unit="page" from="279" to="287" />
			<date type="published" when="1986">1986</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<title level="m" type="main">Neural Networks for Statistical Modeling</title>
		<author>
			<persName><forename type="first">M</forename><surname>Smith</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1993">1993</date>
			<publisher>John Wiley &amp; Sons, Inc</publisher>
			<pubPlace>New York, NY, USA</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Exploiting wikipedia for cross-lingual and multilingual information retrieval</title>
		<author>
			<persName><forename type="first">P</forename><surname>Sorg</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Cimiano</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Data Knowl. Eng</title>
		<imprint>
			<biblScope unit="volume">74</biblScope>
			<biblScope unit="page" from="26" to="45" />
			<date type="published" when="2012-04">Apr. 2012</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">String comparator metrics and enhanced decision rules in the Fellegi-Sunter model of record linkage</title>
		<author>
			<persName><forename type="first">W</forename><forename type="middle">E</forename><surname>Winkler</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Section on Survey Research</title>
				<meeting>the Section on Survey Research</meeting>
		<imprint>
			<date type="published" when="1990">1990</date>
			<biblScope unit="page" from="354" to="359" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
