<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">I. INTRODUCTION FFECTIVE acquisition, organization, processing, use and sharing of the knowledge embedded in textual and multimedia content play a major role for competitiveness in the modern information society and for the emerging knowledge economy. However, this wealth of knowledge implicitly conveyed in the vast amount of available digital content is nowadays only accessible, if considerable manual effort has been invested into its interpretation and semantic annotation, which is possible only for a small fraction of the available content</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Roberto</forename><surname>Bartolini</surname></persName>
							<email>roberto.bartolini@ilc.cnr.it</email>
						</author>
						<author>
							<persName><forename type="first">Emiliano</forename><surname>Giovannetti</surname></persName>
							<email>emiliano.giovannetti@ilc.cnr.it</email>
						</author>
						<author>
							<persName><forename type="first">Simone</forename><surname>Marchi</surname></persName>
							<email>simone.marchi@ilc.cnr.it</email>
						</author>
						<author>
							<persName><forename type="first">Claudio</forename><surname>Andreatta</surname></persName>
							<email>andreatta@itc.it</email>
						</author>
						<author>
							<persName><forename type="first">Roberto</forename><surname>Brunelli</surname></persName>
							<email>brunelli@itc.it</email>
						</author>
						<author>
							<affiliation key="aff0">
								<orgName type="department" key="dep1">Montemagni work</orgName>
								<orgName type="department" key="dep2">Istituto di Linguistica Computazionale (ILC-CNR)</orgName>
								<address>
									<addrLine>Pisa, via Moruzzi 1</addrLine>
									<postCode>56124</postCode>
									<settlement>Pisa</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff1">
								<orgName type="department">Istituto per la Ricerca Scientifica e Tecnologica ITC-irst in Trento</orgName>
								<address>
									<addrLine>via Sommarive 18</addrLine>
									<postCode>38050</postCode>
									<settlement>Trento</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff2">
								<orgName type="department">L3S Research Center</orgName>
								<address>
									<addrLine>Appelstrasse 9a</addrLine>
									<settlement>Hannover</settlement>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff3">
								<orgName type="department">Department of Information and Communication Technologies</orgName>
								<orgName type="institution">University of Trento</orgName>
								<address>
									<addrLine>via Sommarive 14</addrLine>
									<postCode>38050</postCode>
									<settlement>Trento</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">I. INTRODUCTION FFECTIVE acquisition, organization, processing, use and sharing of the knowledge embedded in textual and multimedia content play a major role for competitiveness in the modern information society and for the emerging knowledge economy. However, this wealth of knowledge implicitly conveyed in the vast amount of available digital content is nowadays only accessible, if considerable manual effort has been invested into its interpretation and semantic annotation, which is possible only for a small fraction of the available content</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">30266E6E924BBC7A1BD94874B8733491</idno>
					<note type="submission">received October 27, 2006.</note>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-25T02:55+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Semantic Web Technologies</term>
					<term>Ontology Creation</term>
					<term>Ontology Extraction</term>
					<term>Ontology Evolution</term>
					<term>Semantic Annotation of Multimedia Content</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>The demand for efficient methods for extracting knowledge from multimedia content has led to a growing research community investigating the convergence of multimedia and knowledge technologies. In this paper we describe a methodology for extracting multimedia information from product catalogues empowered by the synergetic use and extension of a domain ontology. The methodology was implemented in the Trade Fair Advanced Semantic Annotation Pipeline of the VIKE-framework.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>The field of semi-automatic information extraction from multimedia corpora is central for overcoming the so-called "knowledge acquisition bottleneck". Multimedia sources of information, such as product catalogues, contain text (captions) and images (pictures of the products) thus requiring information extraction approaches combining several different techniques, ranging from Natural Language Processing to Image Analysis and Understanding. In our approach we have three main aspects to consider: 1) the information extraction per se, 2) the ontology, its use and creation, and 3) the usage of the ontology in the information extraction process and the synergy between different kinds of extraction processes.</p><p>The development of adequate ontologies itself is one of the knowledge acquisition bottlenecks: the use of (semi-) automatic tools for semantic information extraction from multimedia corpora is very promising but, to be efficiently exploited, must have access to a formal representation of a given domain, i.e., an ontology. We support the ontology creation process in two different and complementary ways, ontology learning and reuse of existing ontologies. The ontology learning approach takes advantage of the results of the extraction to enrich the ontology, and the reuse support provides methods and tools to re-use already existing ontologies which capture the target domain under a similar modelling perspective as the one of interest for the extraction task. This (apparent) vicious circle (between the need of having the domain represented in the ontology for an extraction process and the enrichment of the ontology based on the results obtained from the extraction) can be turned to a virtuous circle if the necessary conditions are set to let the evolving ontology and the information extraction tool interact in a synergetic way.</p><p>After a brief introduction to the Vike-Framework the general methodology is described in section III, including specific details about the four different components of the system pipeline. Some conclusions will be presented in section </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>II. THE VIKE-FRAMEWORK</head><p>The methodology we present is developed inside the Vikef project (Virtual Information and Knowledge Environment Framework, IST-2002-507173 -http://www.vikef.net/), which creates an advanced software framework for enabling the integrated development of semantic-based Information, Content, and Knowledge (ICK) management systems. Apart from the scientific and academic interest related to these fields of research we have also registered a growing need from industrial parties for automated knowledge elicitation tools to be applied to their commercial resources, such as product catalogues.</p><p>VIKEF bridges the gap between the partly implicit knowledge and information conveyed in scientific and business content resources (e.g. text, speech, images) and the explicit representation of knowledge required for a targeted and effective access, dissemination, sharing, use, and annotation of ICK resources by scientific and business communities and their information-and knowledge-based work processes.</p><p>R&amp;D within VIKEF builds on and significantly extends the current Semantic Web efforts by addressing crucial operationalisation and application challenges in building up real-world semantically enriched virtual information and knowledge environments.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>III. THE METHODOLOGY</head><p>The task of (semi) automatically annotating content objects with semantic information requires a multi-phased process, where multimedia entities discovered within a content object are coupled with domain knowledge represented by an ontology. For effective semantic annotation support, linguistic, image-related and knowledge representation aspects, approaches, and formats, have to be combined in a synergetic way. The proposed methodology can be presented as a pipeline (together with the employed representation formats within the pipeline), which supports semantic annotation in a flexible and pragmatic way.</p><p>The pipeline has been implemented as a prototype developed as part of the VIKEF project and evaluated for content from the Trade Fair domain.</p><p>The main components of the pipeline are four, and can be functionally summarized in this way: The approach has been thought and implemented to provide the possibility of triggering a "virtuous circle": once information extracted in the annotation steps are integrated inside the ontology, the whole process can be restarted, thus allowing the textual and image annotators to exploit the novel information added to the ontology during the previous run.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>A. Annotation of Text</head><p>Semantic annotation of content is a crucial task (probably the most important) for processing documents to be accessed inside the Semantic Web. To semantically annotate a text it is necessary to develop (semi) automatic Information Extraction techniques capable of overcoming the so-called "knowledge acquisition bottleneck" typical of Semantic Web related applications.</p><p>Semantic annotation of product catalogues poses different challenges at different levels. Concerning the textual part, relative to product descriptions, catalogues do not contain linguistically sound text: very often, sentences are constituted by strictly nominal descriptions, thus discouraging the recourse to traditional NLP techniques. On the other hand, product descriptions appear as semi-structured texts where product features appear in a fixed (or at least regular) order. Semantic annotation of product catalogues appears therefore as a complex task requiring the combination of different types of techniques. Previous works about product information semantic annotations are quite scarce: the two main works being the European project CROSSMARC <ref type="bibr" target="#b0">[1]</ref> and the Czech national project Rainbow <ref type="bibr" target="#b1">[2]</ref>. The project CROSSMARC aims to the electronic-retail product comparison, using a combination of language engineering, machine learning and user modelling, where a domain ontology is used as "semantic glue" to link together the various analysis modules.</p><p>Within the Rainbow project, a multi-layered ontology has been defined, to integrate the more abstract aspects of the domain (domain-neutral), relative to web-sites in general, with the more specific ones (domain-dependent) relative to concepts found in sites of small organization offering products or services. Concerning the information extraction task, Rainbow makes use of lexical indicators and, depending on the document to analyze, applies HTML-centred or free-textcentred extractors, in the latter case using shallow parsing techniques.</p><p>The hybrid methodology we propose (which has been applied to Italian product catalogues belonging to the furniture domain) makes use of two different approaches: first, pattern matching techniques are resorted to for isolating individual product descriptions within the textual flow and for identifying their basic building blocks (e.g. the product name, its price as well as its natural language description). Then, for each identified product, the natural language description is processed by a battery of NLP tools <ref type="bibr">([3] [4]</ref>) in charge of identifying relevant entities (e.g. colour, material, parts of a given product) and the relations holding between them (e.g. part_of, colour_of which can be referred either to the product itself or to individual parts).</p><p>The architecture in Fig. <ref type="figure" target="#fig_1">1</ref> includes two main components, the Product catalogue Italian Semantic Annotator (PISA), and the Product catalogues Terminology Processor (PTP), both exploiting the battery of NLP modules, the former to linguistically analyze the free text part of the product descriptions, the latter to obtain the TermBank <ref type="bibr" target="#b4">[5]</ref> upon which a simple application ontology can be constructed, to be exploited for disambiguation. This "proto"-ontology forms the terminological basis for the development of the final (project) ontology, to be populated using the information derived from the multimedia semantic annotation tasks.</p><p>Once this proto-ontology has been constructed using the PTP module, it is possible to run the PISA component (Fig. <ref type="figure" target="#fig_2">2</ref>): each product description is firstly extracted by pattern matching starting from a set of regular expressions, each one matching a particular product structure. Once a description has been isolated, some of its components, interpreted on the basis of particular "groups" of the matching regular expressions ("name", "type", "product id", etc.) can be detected.</p><p>The remaining "free text" part of the description is then processed by the NLP Manager module, which is able to access the NLP tools and the ontology, the former to linguistically analyze the text, the latter to resolve possible syntactical ambiguities found during the analysis.</p><p>Consider the example in Fig. <ref type="figure" target="#fig_3">3</ref> which is relative to the annotation of the description in the box about a given product. Through pattern matching it is possible to extract its name ("Sanela"), the type (cushion), the price (€12,95), its dimensions (40 cm of width and 60 cm of length) and the product unique identifier (900.582.56), as well as the relations between this information and the product itself (i.e. name_of, price_of, etc). The natural language description identified at this stage is then passed to the NLP Manager which is in charge of acquiring, with the support of the application ontology, further information about the product: in this example, the system detects a part ("cover") and a material ("cotton"), as well as a relation holding between them ("made_of"). The box below contains a snippet of the (XMLstyle) final annotated product description, where some of the extracted features of the product are listed, including the fact, for instance, that the cover of the cushion is made of cotton.</p><p>A preliminary evaluation of the system has been done analyzing the semantic annotation of the "IKEA 2006" italian  furniture catalogue. There are two main levels of evaluation that have been taken into consideration, relative to the two main components of the system: PTP and PISA. Due to the lack of a "golden standard" furniture ontology to compare to the one obtained with the help of the PTP component, a "task based" evaluation technique has been adopted, where the coverage of the ontology has been indirectly evaluated on the basis of the quality of the obtained annotation.</p><p>To sum up the results of the preliminary evaluation we can say the ontology is able to detect approximately the 70% of the terms appearing in the free text description of the extracted product and put them in the correct relation, thus scoring a 0,7 of recall. On the other hand, considering that only terms included in the ontology can be extracted, the system scores a precision of 1. Concerning the disambiguation functionality, our analysis has proved that every time two terms are correctly detected and recognized, the disambiguation works. After all, we can assert that the quality of the linguistic analysis is strongly related to the ontology coverage.</p><p>From the pattern matching point of view, the system has scored a 0,9 of precision and 0,8 of recall, extracting 800 products out of 1000 applying 9 different regular expressions.</p><p>Concerning the task of semantic text annotation future directions of research include, on the one hand the application of the presented technique to different product catalogues (not necessarily in italian) and, on the other hand, the evolution of the methodology by the integration of a more sophisticated ontology learning from text system currently under development within the DylanLab of the ILC-CNR in Pisa.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>B. Annotation of Images</head><p>The information conveyed by a multimedia document is analyzed and extracted at two different levels: the document level, in which the document (geometrical) layout is investigated considering both text and images, and the image level, in which pictures are examined in order to describe their visual content and to recognize the depicted objects. Two interacting methodologies were adopted to analyze the document at the two different levels, both of them exploiting a domain ontology.</p><p>A catalogue document usually provides a rich source of structure. As far as the first algorithm is based on a modified version of the data mining system SUBDUE ( <ref type="bibr" target="#b5">[6]</ref>). SUBDUE is a system that discovers interesting substructures in structural data based on inexact graph matching techniques and computationally constrained beam search algorithm guided by heuristics. In particular, a substructure is evaluated based on how well it can compress the graph, according to the minimum description length principle. Highly compressing structures can be considered as building block of the entire data set.</p><p>To apply the algorithm to a document, we construct a graph representing the page layout (Fig. <ref type="figure" target="#fig_4">4</ref>). The graph representation consists of two types of vertices, image (classified in three categories: highlight, scene and miscellaneous) and text paragraph, and one type of arc representing the spatial relationships between the vertices (e.g. top-left, overlapping).</p><p>The algorithm can provide abstract structured components resulting in a hierarchical view of the document that can be analyzed at many levels of detail and focus. Usually in catalogues a common recurring structure is formed by image and caption (Fig. <ref type="figure" target="#fig_5">5</ref>). It is remarkable fact that, while text and images may be separately ambiguous, jointly they tend not to be. Establishing from the catalogue structure meaningful links between images and text paragraphs allows exploiting the semantic annotation of the textual part to semantically annotate the images or to guide image processing algorithms in order to recognize the object depicted or to infer correspondence between words and particular image structures (as in <ref type="bibr" target="#b6">[7]</ref> <ref type="bibr" target="#b7">[8]</ref>). Domain knowledge can be also added to guide the discovery process and to separate the important substructures from the irrelevant ones.</p><p>The methodology adopted to describe the image content relies on MEMORI ([9][10]), a system for the detection and recognition of objects in digital images using pre stored visual information obtained from shots or 3D models. Object recognition plays a crucial role in Computer Vision, especially in the semantic description of visual content. Although object recognition has been intensely studied, it still remains a hard and computational expensive problem. The main difficulty in  the description of image content is the lack of information about the kind and the number of objects possibly present. Moreover, objects can appear at different locations in the image and they can be deformed, rotated, rescaled, differently illuminated or even occluded with respect to a reference view. In order to simplify object detection and to reduce computational cost, many systems (e.g. <ref type="bibr" target="#b10">[11]</ref>) limit the recognition to specific classes of objects. In these cases, a priori knowledge permits to select the most descriptive features for the objects at hand and to circumscribe the search space. However, even under this restriction, high classification performance is seldom reached. Moreover, many object recognition systems rely on user interaction to label as wrong or correct the returned items or to improve system response <ref type="bibr" target="#b11">[12]</ref>.</p><p>The MEMORI system tackles the object recognition problem by segmenting the input image into regions, applying a region grouping algorithm, which interacts with an object classifier, producing a set of object candidates and filtering the candidate list (Fig. <ref type="figure" target="#fig_6">6</ref>). The object segmentation and recognition modules need domain knowledge in the form of object snapshots from multiple view points.</p><p>Supporting the extraction and recognition process with a domain ontology allows the development of context-aware strategies in order to guide and focus the multimedia semantic analysis. The knowledge of the image context permits the object recognition module to restrict the search to a limited subset of objects and to refine the heuristics weighting object hypotheses according to their own context. The other way round content based image analysis allows the acquisition and exploitation of similarity relations among multimedia entities thus allowing to refine and enrich the knowledge representation modeled in the domain ontology.</p><p>The experiments and tests conducted so far show that the proposed approach is a promising method for the detection and recognition of objects and for image annotation. An extensive assessment of the performance was conducted on a synthetic test-bed. A set of synthetic images has been created by drawing on a non uniform background rotated and rescaled version of objects taken from COIL-100 <ref type="bibr" target="#b12">[13]</ref>, a database widely adopted in the object recognition community.</p><p>The achieved recognition rates were compared with other state of the art recognition methods ( <ref type="bibr">[14][15]</ref>), MEMORI performs best in all experiments, regardless of the number of training views. The recognition rate is over 96% with as few as only 4 training views, demonstrating the robustness of the method. New techniques to make the system robust with respect to illumination changes and partial occlusions are currently under development. Future work will extend the evaluation to more complex and non synthetic images. The potential synergy between visual similarity and semantic similarity measures based on ontologies will also be investigated and exploited.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>C. Elicitation and Refinement</head><p>Once information has been extracted from text and images, and stored in the XML-based format partially described in Figure <ref type="figure" target="#fig_3">3</ref>, the next task is to make its meaning explicit (and thus machine-understandable) by transforming it into a suitable RDF/OWL representation.</p><p>This transformation has two parts<ref type="foot" target="#foot_0">1</ref> :</p><p>• The first step, called syntactic elicitation, aims at producing a collection of RDF statements, namely an explicit representation of the knowledge content of the XML elements in the schema. For example, we want to construct the fact (expressed by an RDF statement) that the string "900.582.56" is the ID of "SANELA" (actually, of some entity which happens to be a product named "SANELA"). • The second step, called semantic elicitation and refinement, aims at leveraging the RDF representation to a full semantic level, where entities are assigned a proper identifier (a URI), properties are associated to some data type or object property in some domain ontology (more precisely, are replaced by the URI of their ontological counterpart), and finally entities are assigned to the most appropriate class (e.g. "SANELA" should be assigned to the class of cushions, which belongs to a hierarchy of classes which is very likely to include among its ancestors the class of products and of physical entity).</p><p>The first step, in our implementation, is rather simple, as it is performed via a simple XSL transformation from the XML schema depicted in Figure <ref type="figure" target="#fig_3">3</ref> to a collection of RDF statements expressed in the RDF XML syntax. The only tricky part is the decision of what statements are to be produced from the XML file (in fact, we observe that the simple snippet in Figure <ref type="figure" target="#fig_3">3</ref> contains a large number of implicit statements, and therefore one needs to select the most useful ones). The outcome of this step is an RDF file containing a potentially large number of facts about the product catalog.</p><p>The second step is by far more interesting. Indeed, the goal is to qualify the RDF statements generated in the previous step by linking their constituents to some pre-existing ontology2 (this is what we call semantic elicitation). In our approach, this task can be decomposed in two different sub-tasks: (1) entitylevel elicitation, and (2) class/property-level elicitation.</p><p>The first sub-task is implemented as a problem of matching entity descriptions onto the entities stored in an repository called OKKAM. A full description of OKKAM is beyond the scope of this paper; the main idea is that it creates and stores URIs (which can then be reused through multiple applications) together with additional information, including any known description of the entity itself. Once a new entity is recognized in any digital document (plain text documents, relational databases, HTML pages, and so on), OKKAM can be queried to check whether that entity already has a known URI, which is returned for reuse; if no match is found, a new entity is created, and stored with its URI and all available descriptions.</p><p>The second task uses a tool, called CtxMatch2.0 <ref type="bibr" target="#b15">[16]</ref> for schema and ontology matching, which is a VIKEF motivated extension of a pre-existing tool. The details of semantic elicitation with CtxMatch2.0 are provided in other papers (see e.g. <ref type="bibr" target="#b16">[17]</ref>). In the scenario described in this paper, the tool is used to match categories and relations extracted from catalogs to classes and properties found in any available domain ontology. The matching method uses two sources of information: the hierarchical structure of elements (which is particularly important in product catalogs) and lexical information associated both to catalog labels and ontology elements.</p><p>The outcome of this second step is a refined RDF file where linguistic descriptions of entities are replaced by unique URIs (which can be used later to merge RDF graphs produced from different catalogs or in general from different collections of documents), category names are replaced by the URIs of ontological classes (if available), and relation names are replaced by the URIs of data type or object properties (if available). We notice that entities may be classified using complex concepts, compositionally constructed from their linguistic descriptions (for example, "SANELA" will correspond to a cushion made of cotton).</p><p>A final remark is that the outcome of this elicitation and refinement phase includes a mixture of knowledge coming from multimedia sources, namely text and pictures, in a single representation.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>D. Reuse of existing ontologies</head><p>Ontology engineering is a very time consuming and subjective task. It is time consuming because it is hard to model not only new, but also already known domains, and subjective since in most of the cases the same domains are modeled in different ways by different ontology engineers. Even if the domains are the same, the perspectives under which they are modeled are usually different. One way to detect the differences (or similarities) of the modeling perspectives is by analyzing the relations existing between the concepts. Due to the already mentioned intrinsic difficulties of the ontology engineering task, advanced research is being performed in order to find a solution to (partially) overcome these problems. The method and tools aim to support the ontology engineer in finding existing Semantic Web ontologies written in OWL (or parts of them) that model the targeted domain in a similar way; this is, with the same modeling perspective. The approach relies on the existence of searchable pool(s) of ontologies where candidate ontologies can be searched for and pre-selected based on a user-specified set of desired ontological classes (this set is called fragment). The pre-selection process searches for a specific percentile match of labels and synonyms of them appearing in the user specified fragment and ontologies in the pool of ontologies. Currently we are using the SWOOGLE (http://swoogle.umbc.edu/) ontology repository, but any other similar repository can be easily incorporated. The pre-selected ontologies are then further analyzed in detail by analyzing the labels of classes and relations in combination with a lexical resource (currently WordNet). Labels are analyzed and word meanings are related by various means in order to compute the likelihood of one or the other possible sense of a word to hold in the given context. This is then represented in two ways, in a logical formulae (following the <ref type="bibr" target="#b17">[18]</ref> approach) used for detecting similarity of concepts using a reasoner and as weights of relevance for each possible sense in order to compare concepts and decide if they are similar and to which degree this similarity holds. The relevance measures are combined with the logical results in order to give a measure of the closeness of the considered concepts. The architecture of this approach is presented in Fig. <ref type="figure" target="#fig_7">7</ref>. The ontology engineer can then decide to use any of the proposed ontologies as a basis for performing the required extension of the initially specified fragment. There exist several different techniques for matching ontologies, using very different approaches. Ontology mapping, ontology alignment and ontology matching fall all under a broader research area where correspondences between two schemata need to be discovered. In some cases rules for mapping schemata need to be created, in others it is enough to know the relations existing between two or more elements from different schemata. The area of alignment or matching of schemata has been long studied and many approaches exist in integrating database schemata, XML schemata in general, query mediation, etc. The closest approaches to ontology alignment are the approaches for schemata alignment, especially taxonomic hierarchies or classification hierarchies in general. A comprehensive study is presented in <ref type="bibr" target="#b18">[19]</ref>. In the following we will present some approaches that are more closely related to the one presented in this section.</p><p>• The iPrompt <ref type="bibr" target="#b19">[20]</ref> tool suite contains tools for ontology merging and alignment among others. Its input are a set of pairs of related/similar concept from different ontologies, based on them it proceeds to find other pairs of similar concepts relying heavily on the structure.</p><p>• The GLUE approach <ref type="bibr" target="#b20">[21]</ref> uses a machine learning approach to analyze instances. The similarity of concept meaning is defined based on the joint probability distribution of the involved concepts.</p><p>• The QOM <ref type="bibr" target="#b21">[22]</ref> approach is based on a combination of several approaches that analyze features of different ontologies. The similarity of concepts is later computed by aggregation of different similarity. The approach is iterative; every iteration uses results obtained in the previous ones in order to enhance quality of the results.</p><p>• Cupid <ref type="bibr" target="#b22">[23]</ref> presents an approach that considers schemata and not instances information for matching XML Schemata mainly. It uses linguistic information and the structure of the schema. This approach uses also keys, referential constraints and other schema information to derive more precise matching results. Cupid also handles context dependent matches of shared type definitions that are used in several larger structures in the schema.</p><p>• The MoA <ref type="bibr" target="#b23">[24]</ref> is an approach to merge and align OWL ontologies. This approach uses linguistic methods to disambiguate the meaning of concepts and proposes an algorithm to detect and represent in a Semantic Bridge semantic equivalence between concepts. The Semantic Bridge can represent equivalence of concepts and properties, subconcept and subproperties and identity of instances. A merging algorithm is also presented.</p><p>• The OMEN <ref type="bibr" target="#b24">[25]</ref> approach uses Bayesian Networks (BN) for deciding the match of concepts based on an initial hand made match. Based on known matches it analyzes the structure (e.g. domain and range of properties) to derive further matches. The algorithm operates iteratively and produces in every iteration a new match which is used in the following interactions.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>IV. CONCLUSIONS</head><p>In this paper we presented a methodology for supporting the use of background ontologies in the task of information extraction from multimedia sources, in particular from product catalogues. Our methodology tries to enable a virtuous circle by which domain ontologies are used in the extraction process, and at the same time the extraction process becomes a way for creating or extending the available ontologies. The result of the extraction process is a semantically rich representation of the content of catalogs, where knowledge extracted from texts (e.g. product descriptions) is integrated with knowledge extracted from pictures, and made available for any service one may want to build on top of it.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>A</head><label></label><figDesc>) Annotation of text -the ontology based semantic annotation of the textual part of the catalogue; B) Annotation of images -the ontology based semantic annotation of the images appearing in the catalogue; C) Elicitation and refinement -to make information extracted by the annotation component machineunderstandable and to enrich the ontology for further annotations; D) Reuse of existing ontologies -to support the ontology creation and refinement by exploiting existing ontologies.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Fig. 1 :</head><label>1</label><figDesc>Fig.1:The general architecture of the annotation system, producing a semantic annotation of product descriptions extracted from the input catalogue.</figDesc><graphic coords="2,311.76,438.48,231.12,239.40" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Fig. 2 :</head><label>2</label><figDesc>Fig. 2: The general architecture of PISA component, including the RegExp Manager committed for the pattern matching step, and the NLP Manager in charge for the analysis of the "free text" part of the product description.</figDesc><graphic coords="3,54.84,459.00,223.68,199.32" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Fig. 3 .</head><label>3</label><figDesc>Fig. 3. Example of product semantic annotation, including entities and relations among entities.</figDesc><graphic coords="3,317.64,359.64,218.76,207.00" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_4"><head>Fig. 4 .</head><label>4</label><figDesc>Fig. 4. The graph structure representing the page layout. Each graph vertex is associated to a text caption or to a product image. Images are extracted from the page background by means of a segmentation algorithm.</figDesc><graphic coords="4,41.88,564.60,241.08,153.96" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_5"><head>Fig. 5 .</head><label>5</label><figDesc>Fig. 5. The final result of the geometrical layout analysis algorithm. The figure depicts the subgraphs extracted and the thickness of the edges is related to the score obtained by the caption-figure association.</figDesc><graphic coords="4,304.80,85.68,240.96,154.08" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_6"><head>Fig. 6 .</head><label>6</label><figDesc>Fig. 6. The MEMORI system analyzes color digital images in order to detect and recognize objects. The figure shows on the left an image extracted from a catalog, and on the right the recognized object along with the resulting annotation compliant to the MPEG-7 standard.</figDesc><graphic coords="5,47.76,85.68,241.08,149.88" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_7"><head>Fig. 7 .</head><label>7</label><figDesc>Fig. 7. The architecture of the described approach for the re-use of existing ontologies.</figDesc><graphic coords="6,409.92,388.32,135.84,198.00" type="bitmap" /></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">More details are provided in<ref type="bibr" target="#b10">[11]</ref>, where the entire VIKEF knowledge pipeline is described thoroughly.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_1">See next section for what concerns the support we provide to build an ontology by reusing existing ontologies.</note>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Cross-lingual multi-agent retail comparison</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">T</forename><surname>Pazienza</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Stellato</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Vindigni</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">AWSS 2003 Workshop on Applications, Products and Services of Web-based Support Systems</title>
				<meeting><address><addrLine>Halifax Canada; WI</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2003-10-13">13 Oct. 2003. 2003</date>
		</imprint>
	</monogr>
	<note>in conjunction with Web-Intelligence international Conference</note>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Snasel: Rainbow -multiway semantic analysis of websites</title>
		<author>
			<persName><forename type="first">V</forename><surname>Svatek</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Kosek</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Labsky</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Braza</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Kavalec</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Vacura</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Vavra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 14th International Workshop on Database and Expert Systems Applications (DEXA&apos;03)</title>
				<meeting>the 14th International Workshop on Database and Expert Systems Applications (DEXA&apos;03)</meeting>
		<imprint>
			<date type="published" when="2003">2003</date>
			<biblScope unit="page">635</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">An Italian shallow parser for robust syntactic annotation</title>
		<author>
			<persName><forename type="first">A</forename><surname>Lenci</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Montemagni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Pirrelli</surname></persName>
		</author>
		<author>
			<persName><surname>Chunk-It</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Computational Linguistics in Pisa -Linguistica Computazionale a Pisa. Linguistica Computazionale, Special Issue, XVI-XVII</title>
				<editor>
			<persName><forename type="first">A</forename><surname>Zampolli</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Calzolari</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">L</forename><surname>Cignoni</surname></persName>
		</editor>
		<meeting><address><addrLine>Pisa-Roma, IEPI; Tomo I</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2003">2003. 2003</date>
			<biblScope unit="page" from="353" to="386" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Hybrid Cosntraints for Robust Parsing: First Experiments and Evaluation</title>
		<author>
			<persName><forename type="first">R</forename><surname>Bartolini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Lenci</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Montemagni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Pirrelli</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of LREC 2004: Fourth International Conference on Language Resources and Evaluation</title>
				<meeting>LREC 2004: Fourth International Conference on Language Resources and Evaluation<address><addrLine>Lisbon, Portugal; Paris</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2004-05-28">26th. 27th &amp; 28 May 2004</date>
			<biblScope unit="volume">III</biblScope>
			<biblScope unit="page" from="859" to="862" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Automatic Incremental Term Acquisition from Domain Corpora</title>
		<author>
			<persName><forename type="first">R</forename><surname>Bartolini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Giorgetti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Lenci</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Montemagni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Pirrelli</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 7th International conference on Terminology and Knowledge Engineering (TKE2005)</title>
				<meeting>the 7th International conference on Terminology and Knowledge Engineering (TKE2005)<address><addrLine>Copenhagen Business School; Copenhagen, Denmark</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2005">2005</date>
			<biblScope unit="page" from="17" to="18" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Iterative Structure Discovery in Graph-Based Data</title>
		<author>
			<persName><forename type="first">J</forename><surname>Coble</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Rathi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">J</forename><surname>Cook</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">B</forename><surname>Holder</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">International Journal on Artificial Intelligence Tools</title>
		<imprint>
			<biblScope unit="volume">14</biblScope>
			<biblScope unit="issue">1-2</biblScope>
			<biblScope unit="page" from="101" to="124" />
			<date type="published" when="2005">2005</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Object Recognition as Machine Translation: Learning a lexicon for a fixed image vocabulary</title>
		<author>
			<persName><forename type="first">P</forename><surname>Duygulu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Barnard</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Freitas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Forsyth</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">European Conference on Computer Vision (ECCV)</title>
				<meeting><address><addrLine>Copenhagen</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2002">2002</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Blobworld: Image segmentation using expectation-maximization and its application to image querying</title>
		<author>
			<persName><forename type="first">C</forename><surname>Carson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Belongie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Greenspan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Malik</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Pattern Analysis and Machine Intelligence</title>
		<imprint>
			<biblScope unit="volume">24</biblScope>
			<biblScope unit="issue">8</biblScope>
			<biblScope unit="page" from="1026" to="1038" />
			<date type="published" when="2002">2002</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Object Recognition in Color Images by the Self Configuring System MEMORI</title>
		<author>
			<persName><forename type="first">M</forename><surname>Lecca</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">International Journal of Signal Processing</title>
		<imprint>
			<biblScope unit="volume">3</biblScope>
			<biblScope unit="issue">3</biblScope>
			<biblScope unit="page" from="176" to="185" />
			<date type="published" when="2006">2006</date>
			<publisher>World Enformatika Society</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Memory-based Object Recognition in digital Images</title>
		<author>
			<persName><forename type="first">C</forename><surname>Andreatta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Lecca</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Messelodi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">10th International Fall Workshop -Vision, Modeling, and Visualization -VMV 2005</title>
				<meeting><address><addrLine>Erlangen, Germany</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2005">November 16-18, 2005</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">A common set ofperceptual observables for grouping, figure -ground discrimination, and texture classification</title>
		<author>
			<persName><forename type="first">A</forename><surname>Hoogs</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Collins</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Kaucic</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Mundy</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Transaction on Pattern Analysis and Machine Intelligence</title>
		<imprint>
			<biblScope unit="issue">4</biblScope>
			<biblScope unit="page" from="458" to="474" />
			<date type="published" when="2003">2003</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Extracting Salient Regions And Learning Importance Scores In Region-Based Image Retrieval</title>
		<author>
			<persName><forename type="first">B</forename><surname>Ko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Byun</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">International Journal of Patter Recognition and Artificial Intelligence</title>
		<imprint>
			<biblScope unit="volume">17</biblScope>
			<biblScope unit="issue">8</biblScope>
			<biblScope unit="page" from="1349" to="1367" />
			<date type="published" when="2003">2003</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<monogr>
		<title level="m" type="main">Columbia object image library (COIL-100)</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">A</forename><surname>Nene</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">K</forename><surname>Nayar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Murase</surname></persName>
		</author>
		<idno>CUCS-006-96</idno>
		<imprint>
			<date type="published" when="1996">1996</date>
		</imprint>
		<respStmt>
			<orgName>Columbia University</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">Technical Report</note>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Object recognition using local affine frames on distinguished regions</title>
		<author>
			<persName><forename type="first">S</forename><surname>Obdrzalek</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Matas</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the British Machine Vision Conference</title>
				<editor>
			<persName><forename type="first">Paul</forename><forename type="middle">L</forename><surname>Rosin</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">David</forename><surname>Marshall</surname></persName>
		</editor>
		<meeting>the British Machine Vision Conference<address><addrLine>London, UK</addrLine></address></meeting>
		<imprint>
			<publisher>BMVA</publisher>
			<date type="published" when="2002-09">September 2002</date>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="113" to="122" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Learning to Recognize 3D Objects with SNoW</title>
		<author>
			<persName><forename type="first">Ming-Hsuan</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Roth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Ahuja</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">The Proceedings of the Sixth European Conference on Computer Vision</title>
				<imprint>
			<date type="published" when="2000">2000</date>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="439" to="454" />
		</imprint>
	</monogr>
	<note>ECCV-2000</note>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Peer-to-peer semantic coordination</title>
		<author>
			<persName><forename type="first">P</forename><surname>Bouquet</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Serafini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Zanobini</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Web Semantics</title>
		<imprint>
			<biblScope unit="volume">2</biblScope>
			<date type="published" when="2005">2005</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<monogr>
		<title level="m" type="main">Bootstrapping semantics on the web. meaning elicitation from schemas</title>
		<author>
			<persName><forename type="first">P</forename><surname>Bouquet</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Serafini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Zanobini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Sceffer</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2006-05">May 2006</date>
			<publisher>WWW2006</publisher>
			<pubPlace>Edinburgh (Scotland, UK</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Enabling a knowledge supply chain: from content resources to ontologies</title>
		<author>
			<persName><forename type="first">R</forename><surname>Stecher</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Niederee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Bouquet</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Jaquin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Ait-Mokhtar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Montemagni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Brunelli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Demetriou</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the ESWC2006 workshop on &quot;Mastering the Gap: from information extraction to semantic representation</title>
		<title level="s">CEUR workshop proceedings</title>
		<meeting>the ESWC2006 workshop on &quot;Mastering the Gap: from information extraction to semantic representation<address><addrLine>Budva (Montenegro</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2006-06-11">11 June 2006</date>
			<biblScope unit="volume">187</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">A Survey of Schema-Based Matching Approaches</title>
		<author>
			<persName><forename type="first">Pavel</forename><surname>Shvaiko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jérôme</forename><surname>Euzenat</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Data Semantics IV</title>
		<imprint>
			<date type="published" when="2005">2005</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<monogr>
		<title level="m" type="main">The PROMPT suite: Interactive tools for ontology merging and mapping</title>
		<author>
			<persName><forename type="first">N</forename><surname>Noy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Musen</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2002">2002</date>
			<pubPlace>CA, USA</pubPlace>
		</imprint>
		<respStmt>
			<orgName>SMI, Stanford University</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">Technical report</note>
</biblStruct>

<biblStruct xml:id="b20">
	<monogr>
		<title level="m" type="main">Learning to match the schemas of data sources: A multistrategy approach</title>
		<author>
			<persName><forename type="first">A</forename><surname>Doan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Domingos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Halevy</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2003">2003</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">QOM -Quick Ontology Mapping</title>
		<author>
			<persName><forename type="first">Marc</forename><surname>Ehrig</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Steffen</forename><surname>Staab</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the International Semantic Web Conference</title>
				<meeting>the International Semantic Web Conference</meeting>
		<imprint>
			<date type="published" when="2004">2004</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">Generic Schema Matching with Cupid</title>
		<author>
			<persName><forename type="first">Jayant</forename><surname>Madhavan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Philip</forename><forename type="middle">A</forename><surname>Bernstein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Erhard</forename><surname>Rahm</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">The VLDB Journal</title>
		<imprint>
			<date type="published" when="2001">2001</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">MoA: OWL Ontology Merging and Alignment Tool for the Semantic Web</title>
		<author>
			<persName><forename type="first">Jaehong</forename><surname>Kim</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Minsu</forename><surname>Jang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Young-Guk</forename><surname>Ha</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Joo-Chan</forename><surname>Sohn</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Sang</forename><forename type="middle">Jo</forename><surname>Lee</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Innovations in Applied Artificial Intelligence: 18th International Conference on Industrial and Engineering Applications of Artificial Intelligence and Expert Systems</title>
				<imprint>
			<date type="published" when="2005">2005</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<analytic>
		<title level="a" type="main">OMEN: A Probabilistic Ontology Mapping Tool</title>
		<author>
			<persName><forename type="first">Prasenjit</forename><surname>Mitra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Natasha</forename><forename type="middle">F</forename><surname>Noy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Anuj</forename><surname>Jaiswal</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the International Semantic Web Conference</title>
				<meeting>the International Semantic Web Conference</meeting>
		<imprint>
			<date type="published" when="2005">2005</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
