<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Data Integration Using DataPile Structure</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">David</forename><surname>Bednárek</surname></persName>
							<email>david.bednarek@mff.cuni.cz</email>
							<affiliation key="aff0">
								<orgName type="department">Department of Software Engineering Faculty of Mathematics and Physics</orgName>
								<orgName type="institution">Charles University Prague</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">David</forename><surname>Obdržálek</surname></persName>
							<email>david.obdrzalek@mff.cuni.cz</email>
							<affiliation key="aff0">
								<orgName type="department">Department of Software Engineering Faculty of Mathematics and Physics</orgName>
								<orgName type="institution">Charles University Prague</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Jakub</forename><surname>Yaghob</surname></persName>
							<email>jakub.yaghob@mff.cuni.cz</email>
							<affiliation key="aff0">
								<orgName type="department">Department of Software Engineering Faculty of Mathematics and Physics</orgName>
								<orgName type="institution">Charles University Prague</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Filip</forename><surname>Zavoral</surname></persName>
							<email>filip.zavoral@mff.cuni.cz</email>
							<affiliation key="aff0">
								<orgName type="department">Department of Software Engineering Faculty of Mathematics and Physics</orgName>
								<orgName type="institution">Charles University Prague</orgName>
							</affiliation>
						</author>
						<title level="a" type="main">Data Integration Using DataPile Structure</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">F0EB06693D8D54E2AAA2FBBC2A4C72A5</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T11:14+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>One of the areas of data integration covers systems that maintain coherence among a heterogeneous set of databases. Such a system repeatedly collects data from the local databases, synchronizes them, and pushes the updates back. One of the key problems in this architecture is the conflict resolution. When data in a less relevant data source changes, it should not cause any data change in a store with higher relevancy. To meet such requirements, we propose a DataPile structure with following main advantages: effective storage of historical versions of data, straightforward adaptation to global schema changes, separation of data conversion and replication logic, simple implementation of data relevance. Key usage of such mechanisms is in projects with following traits or requirements: integration of heterogeneous data from sources with different reliability, data coherence of databases whose schema differs, data changes are performed on local databases and minimal load on the central database.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>The concept of data integration covers many different areas of application <ref type="bibr" target="#b2">[3,</ref><ref type="bibr" target="#b12">13]</ref>. In this paper, we focus on one kind of applications characterized by the following requirements: − Data warehousing: The data originated at the local data sources should be replicated into a central repository (data warehouse) in order to allow efficient analytical processing and querying the central system independently of local systems. − Back-propagation: Any update which occurs in a local database (performed by its local application) should be distributed to other local databases for which this kind of data is relevant. − History records: The central repository should maintain full history of all data stored therein. Each one of the requirements forms a well-known problem having well-known solutions <ref type="bibr" target="#b1">[2,</ref><ref type="bibr" target="#b7">8,</ref><ref type="bibr" target="#b8">9,</ref><ref type="bibr" target="#b9">10]</ref>; nevertheless, combining the requirements together introduces new, interesting problems, and disqualifies many of the traditional solutions. This paper presents a technique, called DataPile, which combines flexible storage technology (built upon a standard relational database system) with system architecture that separates the replication mechanisms from the schema-matching and data-conversion logic. Since the approach is inspired by XML techniques rather than relational databases, its combination with modern XML-based technologies is straightforward. Nevertheless, the system is created over relational database system and direct integration with traditional database systems is also possible.</p><p>One of the most difficult problems in the area of data integration is handling of duplicate and inconsistent information. The key issue in this problem is entity identification, i.e. determining the correspondence between different records in different data sources <ref type="bibr" target="#b10">[11,</ref><ref type="bibr" target="#b13">14]</ref>. The reality requires that the system administrators understand the principles of the entity matching algorithm; thus, various difficult formalisms presented in the theory <ref type="bibr" target="#b6">[7]</ref> are not applicable. Our approach uses a simplified entity matching system which allows the users to specify matching parameters that are easy to understand. Some researchers <ref type="bibr" target="#b5">[6]</ref> advice that successful entity identification requires additional semantics information. Since this information cannot be generally given in advance, the integrated system should be able to defer decision to the user. The system should detect inconsistencies and either resolve them, or allow users to resolve them manually. The need for user-assisted conflict resolution induces a new class of problems: The repository should be able to store data before final resolution while their relationship to the real world entities is not consistent. Consequently, the system should be able to merge entities whenever the users discover that the entities describe the same real-world entity, and, conversely, to split an entity whenever the previous merge is found invalid. Under the presence of integrity constraints and history records, this requirement needs special attention.</p><p>The relationship between the global system and local database is usually expressed using the global-as-view and local-as-view approaches <ref type="bibr" target="#b4">[5]</ref>. In our system, a mixture of these methods is used depending on the degree of integration required.</p><p>Maintenance of history records falls in the area of temporal databases and queries, where many successful solutions are known <ref type="bibr" target="#b0">[1,</ref><ref type="bibr" target="#b3">4,</ref><ref type="bibr" target="#b11">12]</ref>. The theory usually distinguishes between the valid time, for which the data element is valid in the real world, and the transaction time, recording the moments when the data entry was inserted, updated, or deleted. In our approach, the central system automatically assigns and stores the transaction time, while the local systems are responsible for maintaining the valid time where appropriate. Queries based on transaction time are processed by special algorithms implemented in the central system; queries related to valid time are processed in the same manner as queries to normal attributes.</p><p>The rest of the paper is organized as follows: The second chapter describes the principles of the DataPile technology used to flexibly store structured data in a relational database system. The next chapter focuses on entity identification using data matching and relevance weighing. The fourth chapter shows the overall architecture of the integrated system. The fifth chapter presents an evaluation based on a commercial data-integration project where the DataPile approach was used.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">The DataPile</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1">Terminology</head><p>We have used an own terminology, which is partly derived from the XML terminology. The first term is entity, which represents a type of the traditional database row. An entity consists of attributes, which are analogous to the traditional database columns. An entity instance is an instance of entity and directly equals to traditional database row contents. An attribute value is an instance of attribute and forms a value of one column in one row. A metatable is a conventional database table used by the DataPile to store schema information and other system data.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2">Data Verticalization</head><p>Usual information systems consist of some nontrivial number of conventional database tables; huge information systems have huge number of such tables. Moreover, the requirement for preserving all changes in data usually leads to the scheme, where changing one value of one column in one row causes inserting a new changed row (possibly very large) and updating the old row with some state changing column (e.g. validity termination timestamp). Another problem in conventional information systems is extensibility; adding some new columns or new tables may cause large application code rewriting. All these problems are addressed by the proposed method of storing data in different way than in traditional approaches but using standard relational databases -the DataPile. All real applications data are stored in two relational tables: one less important table DT_LOB_HASH is dedicated for storing LOBs (for performance purposes), and the second one, the most important, DT_PILE stores data of all other datatypes. This particular table is called the Pile, because all data is stored in one table without any "well-formed" internal structure or hierarchy. Each row in the pile represents one attribute, whose value is/was valid during certain interval of transaction time.</p><p>The Fig. <ref type="figure" target="#fig_0">1</ref> represents slightly simplified schema of the heart of DataPile-based information system. Tables with prefix DT_ hold real data; all other tables (with prefix MT_) are metatables. The table DT_ENTITY holds valid "global" ID for an entity instance stored in the pile together with information about the entity in form of a reference to the metatable MT_ENTITY_TYPE which stores entities. Entities consist of attributes, and this is modeled by the metatable MT_ATTR_TYPE.</p><p>Real values are stored in columns val_xxx of the main table DT_PILE , where xxx represents logical type of the attribute (number, string, datetime, ID -foreign key). Besides the actual data, other additional data is stored in the DT_PILE table: Transaction time aspect of any attribute value is represented by two columns ts_valid_xxx. The type of given attribute value can be found by reference attr_id to the MT_ATTR_TYPE. The ent_id value compounds all attribute values into one entity instance. Other columns not mentioned here serve the system for proper implementation of the functionality needed.</p><p>Such a structure easily avoids all the problems mentioned at the beginning of this paper: The number of relational tables used does not grow with an expansion of an information system; it is constant regardless on how huge the system is. Data changes are preserved with minimal overhead -one attribute value change is represented by inserting a new value into the pile -one new row is inserted into DT_PILE table not touching the rest of attribute values related to the same entity instance. Extensibility of the system is reached by the possibility to insert some new rows into metatables and therefore the possibility of defining new entities, attributes, or both.</p><p>From the above described layout we can see that this data structure fulfils two requirements put on the information system as a whole: easy extensibility of the data scheme and full information about data changes on the timeline.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Data Matching and Weighing</head><p>The requirement on data unification is solved by two algorithms: data matching and data weighing.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">Data Matching</head><p>Let us show an example, which represents usual situation we meet while processing the same data in different applications. Let application A1 have a record about a per-son with the name "Jana", surname "Teskova" with some personal identification number "806010/7000" and an address "Mother's home No. 10". The same information is stored in the application A2 as well. After Jana Teskova got married, she took her husband's surname (as it is quite usual over here). So her surname changes to "Stanclova". She also moved to live with her new husband on the address "New home 20". Our person notifies about her marriage and the accompanying changes only the office using application A1, and does not notify other office with application A2 -at first she might not even know A2 does not share the data with A1 as they both are used to keep data about people in one organization, and at second she may expect that A1 and A2 are integrated together, so changes in A1 are automatically redistributed to A2 as well (but this is a so called distribution problem, which is discussed later). So the result is A1 and A2 store different data about one entity instance. What happens when we try to merge data from A1 and A2 into a common data storage?</p><p>As our example shows, nearly all attributes have changed. But some of them are constant, especially personal identification number, which should by truly unique in our country. The association of words "should be" unfortunately means that cases exist, when different persons have the same personal identification number. On the other side, these cases are rare. Having two personal records with the same personal identification number means they belong in fact to a single person with probability of roughly 0,999999.</p><p>In this example, other attributes have changed, but a combination of some attributes can have significant meaning: e.g. name and surname together form a whole name. Even name and surname aren't commonly unique in a state, equality of such attributes means some nontrivial probability these two records describe a single person.</p><p>This example leads us to attribute classification. Every attribute is assigned one of these classes: determinant, relevant, uninteresting. − Determinant -identifies an entity instance with very high probability (e.g. personal identification number, passport number etc.). − Relevant -significant attribute, which helps identify unambiguously equality of entities (e.g. attribute types "name" and "surname" for entity type "person"). − Uninteresting -has no impact on entity matching. Following algorithm describes entity matching for two entity instances (one is already stored in the database, the second one is a newly integrated/created entity): 1. All determinant and relevant attribute values are equal -quite clear match with very high probability. 2. A nonempty subset of determinant and nonempty subset of relevant attribute values are equal, remaining determinant and relevant attribute values have no counterpart in the complimentary entity instance -very good match with quite high probability yet (example: let us extend our example with another attribute "passport number". The first entity instance has attributes "personal identification number" and "passport number" filled. The second entity instance has only "personal identification number" filled and "passport number" is missing.).</p><p>3. A nonempty subset of determinant attribute values is equal, remaining determinant attribute values has no counterpart in the complimentary entity instance, but some nonempty subset of relevant attribute values differ -this case seems to be clear as well, because the probability of match for determinant attribute values outweighs probability of different relevant attribute values, but some uncertainty remains as the probability of determinant attribute values is always &lt;1.0. We must not loose any data and their history, so the system solves such a case by considering these two entity instances as different with notification to the system administrator. The administrator can investigate this case more precisely and can merge these two entities together using an administrator application. This case directly describes the situation during entity matching from our first examplethe personal identification number as a determinant attribute value is the same, but surname as a subset of relevant attribute values differs. 4. A nonempty subset of determinant attribute values differs, remaining determinant attribute values with counterpart are equal, some nonempty subset of relevant attribute values is equal, remaining relevant attribute values have no counterpartthis case usually arises out of misspelling one determinant attribute value. This case is solved as above -entity instances are considered to be different and the system administrator is notified. 5. All other cases -input entities are different entity instances with very high probability.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">Data Weighing</head><p>Let us show another example: An employee record is usually kept in different applications in different departments, e.g. human resources department, payroll/accountants department, library, etc. Some applications and departments themselves emphasize some entities and usually some subset of attributes from entities used, e.g. staff department knows with high probability that given person has a certain name, surname, home address, etc., whereas payroll department knows with high probability his/her account number, etc. It should be beneficial for data integration to have possibility measure somehow the probability, that an application has entity instances (or more precisely on individual attribute values) filled with correct values. Therefore, every attribute value in the DataPile keeps a number which measures probability this given value is correct. This probability is stored in the DT_PILE column "relevance".</p><p>During processing of incoming data all incoming attribute values are somehow evaluated (this will be explained later) and the computed relevance is compared to current relevance of attribute value stored in the DataPile. If the new value has greater or equal computed relevance than current value has, the new value "wins", becomes the current value, and the old value is marked as archive. Otherwise (when the relevance is lower than current relevance) the new value is stored as well, but only as a remark saying the application has ineffectually tried to change this attribute value.</p><p>But there is a problem: when an unimportant application with low relevance keeps correct data (replicated from the central repository) and wants to change some attribute values (because a user has made some changes), these changes will be always ignored. This problem is solved by "approving" the data. Every application must confirm to central repository it agrees with current data replicated from central repository to this application. This confirmation is stored in the DataPile and the system knows the given application has accepted the current attribute value. When such an application (which approved a current attribute value) changes the value, the rule about weighing relevance is ignored and the attribute value is changed.</p><p>For example, a large company has usually some branch offices, where department branches can be located as well. Such branches usually show different credibility, which should be reflected by the relevance computation as well.</p><p>For computing relevancy of an attribute value following equation is therefore used:</p><formula xml:id="formula_0">at ec et iap ap a R R R R R R ⋅ ⋅ ⋅ ⋅ = (1)</formula><p>where R a means attribute value relevancy, R ap is static application relevancy (e.g. application used by staff department), R iap is static instance of application relevancy (e.g. branch of department), R et is a static entity relevancy, R ec is a computed entity instance relevancy, and R at is a static attribute relevancy. All these static values are stored in metatables as floating-point values. R ec represents computed relevancy for given entity instance and its value is computed as follows:</p><formula xml:id="formula_1">( ) i R R i ac ec ∏ = (2)</formula><p>where R ac (i) is either a value stored in metatables (when attribute i from given entity has some particular value stored in metatables as well), or it has value of 1. This computed relevancy supports changing of the relevancy based on the presence of selected attributes in the entity.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Implementation</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1">The DataPile Architecture</head><p>The whole DataPile-based system is in general shown on the following picture. data structure for collected data storage, the Temporal DB serves only as communication medium between the DataPile machinery and applications (data is revealed here only during replication and immediately deleted when replication ends). The Application server, which gives life to the whole system, is discussed in following section.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2">Application Server</head><p>The whole DataPile architecture utilizes the request/reply paradigm. The application server behaves to the rest of world passively; it waits for requests inserted into the temporal DB, fulfills them using central DB data, and writes a reply back into the temporal DB. Application server is by intention implemented so that it doesn't understand any data semantic. Everything is controlled by the content of metatables and nothing is hardcoded. The application server is primarily responsible for replication, computing algorithms like all the above mentioned data matching and weighing, and can handle other requests too (e.g. perform specialized queries upon the data stored in the DataPile).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3">Replication channel</head><p>The communication channel between an application and the DataPile machinery is not as simple as it may appear on the first look. In reality there are inserted two filters: export and import filter. They are traditional adapters, which are responsible for adapting different database schemas used by the DataPile machinery and local database. A watchful reader may note that the direction from application to the DataPile is marked as "export", opposite direction as "import"; the marking is taken from the application point of view. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.4">Cache</head><p>New applications may advantageously use collected data in central repository. Unfortunately the DataPile structure is not very well suited for direct access (e.g. searching is not very effective). Almost all applications in fact need to know only current attribute value, ignoring all history stored in the DataPile. To support such applications, the application server actively builds and maintains caches, where only current attribute values are stored and these caches are presented to applications in the form of traditional relational tables. The applications can easily search and use all other RDBMS functions on the caches.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Evaluation and Conclusions</head><p>The architecture described in this paper brings an alternative to traditional techniques. It brings several advantages but also some disadvantages.</p><p>All the concepts described in this paper were used in a real project -design and development of an information system based on data replication and synchronization of data coming from bigger number of different data sources (approx. 30 local information systems and 20 other applications producing 30 millions entries per year for 60 000 users, in this project). The project lasts from the fall 2003 and now is in the phase of finishing the pilot phase and starting roll-out.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 .</head><label>1</label><figDesc>Figure 1. A sample schema of the DataPile-based system</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 3 .</head><label>3</label><figDesc>Figure 3. Communication between the DataPile machinery and an application</figDesc></figure>
		</body>
		<back>
			<div type="annex">
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Basic and commonly usable advantage of the DataPile structure is its maintainability, easy extensibility and ability to keep the track of the whole data history. All current applications used at all branches remain preserved and functional without any change according to a strongly desired requirement. The central data repository integrates data from all data sources including all their history and sources of their changes. This enables recovering of any historical snapshot of any data. All data changes are redistributed to all other applications that contain these data, even if the source and destination schemas are different. The global schema changes affect neither data in the central repository nor local applications.</p><p>During the development of the project, we have discovered several disadvantages of our approach:</p><p>Efficiency, especially during export and matching, is low. During the initial export of one certain local system, about 500 000 entries had to be processed. This took more than 24 hours. This time complexity is caused by a relatively complex matching algorithm. Fortunately, this time complexity is not very important in everyday life because number of data changes is smaller in magnitude in comparison to the initial migration data volume.</p><p>The second disadvantage is the fact, that the structure of the central repository makes constructing direct queries difficult. Therefore the concept of caches was introduced and all the queries to non-historical data are performed on the caches instead of the DataPile itself.</p><p>The project showed that the DataPile approach is suitable for certain class of large applications, where data warehousing is coupled with maintaining consistency of local databases. In this class of applications, the drawbacks mentioned above are outweighed by integration of data warehousing features with the support for data replication, synchronization, and cleaning using back-propagation.</p></div>			</div>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Modeling Temporal Consistency in Data Warehouses</title>
		<author>
			<persName><forename type="first">R</forename><surname>Bruckner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>List</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Schiefer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">A</forename><surname>Tjoa</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">12th International Workshop on Database and Expert Systems Applications (DEXA&apos;01)</title>
				<meeting><address><addrLine>Munich, Germany</addrLine></address></meeting>
		<imprint>
			<publisher>IEEE Computer Society Press</publisher>
			<date type="published" when="2001">2001</date>
			<biblScope unit="page" from="901" to="905" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">The TSIMMIS Project: Integration of Heterogeneous Information Sources</title>
		<author>
			<persName><forename type="first">S</forename><surname>Chawathe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Garcia-Molina</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Hammer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Ireland</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Papakonstantinou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Ullman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Widom</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of IPSJ Conference</title>
				<meeting>of IPSJ Conference</meeting>
		<imprint>
			<date type="published" when="1994">1994</date>
			<biblScope unit="page" from="7" to="18" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<monogr>
		<author>
			<persName><forename type="first">I</forename><forename type="middle">K</forename><surname>Ibrahim</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Schwinger</surname></persName>
		</author>
		<title level="m">Data Integration in Digital Libraries: Approaches and Challenges</title>
				<meeting><address><addrLine>Austria</addrLine></address></meeting>
		<imprint>
			<publisher>Software Competence Center Hagenberg</publisher>
			<date type="published" when="2001">2001</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Temporal Data Management</title>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">S</forename><surname>Jensen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">T</forename><surname>Snodgrass</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE TKDE</title>
		<imprint>
			<biblScope unit="volume">11</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="36" to="45" />
			<date type="published" when="1999">1999</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Data integration: A theoretical perspective</title>
		<author>
			<persName><forename type="first">M</forename><surname>Lenzerini</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of the 21st ACM SIGACT SIGMOD SIGART Symp. on Principles of Database Systems</title>
				<meeting>of the 21st ACM SIGACT SIGMOD SIGART Symp. on Principles of Database Systems<address><addrLine>PODS</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2002">2002. 2002</date>
			<biblScope unit="page" from="233" to="246" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Entity identification in database integration</title>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">P</forename><surname>Lim</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Srivastava</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Prabhakar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Richardson</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings Ninth International Conference on Data Engineering</title>
				<meeting>Ninth International Conference on Data Engineering<address><addrLine>Vienna, Austria; Washington, DC</addrLine></address></meeting>
		<imprint>
			<publisher>IEEE Computer Society Press</publisher>
			<date type="published" when="1993">April 19--23, 1993. 1993</date>
			<biblScope unit="page" from="294" to="301" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Managing Data Quality in Cooperative Information Systems</title>
		<author>
			<persName><forename type="first">M</forename><surname>Mecella</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Scannapieco</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Virgillito</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Baldoni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Catarci</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Batini</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 10th International Conference on Cooperative Information Systems</title>
				<meeting>the 10th International Conference on Cooperative Information Systems<address><addrLine>Irvine, CA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2002">2002</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<monogr>
		<title level="m" type="main">The Lord Of The Rings: Efficient Maintenance Of Views At Dataware Houses</title>
		<author>
			<persName><forename type="first">A</forename><surname>Mostéfaoui</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Raynal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Roy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Agrawal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>El Abbadi</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2002">2002</date>
			<biblScope unit="volume">1441</biblScope>
			<pubPlace>Rennes, France</pubPlace>
		</imprint>
		<respStmt>
			<orgName>IRISA</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">Publication interne No</note>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">The CVS Algorithm for View Synchronization in Evolvable Large-Scale Information Systems</title>
		<author>
			<persName><forename type="first">A</forename><surname>Nica</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">J</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">A</forename><surname>Rundensteiner</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of International Conference on Extending Database Technology (EDBT&apos;98)</title>
				<meeting>International Conference on Extending Database Technology (EDBT&apos;98)<address><addrLine>Valencia, Spain</addrLine></address></meeting>
		<imprint>
			<date type="published" when="1998">1998</date>
			<biblScope unit="page" from="359" to="373" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Maintaining Data Warehouses Over Changing Information Sources</title>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">A</forename><surname>Rundensteiner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Koeller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Zhang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Communications of the ACM</title>
		<imprint>
			<biblScope unit="volume">43</biblScope>
			<biblScope unit="issue">6</biblScope>
			<date type="published" when="2000">2000</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Extensible and similarity-based grouping for data integration</title>
		<author>
			<persName><forename type="first">E</forename><surname>Schallehn</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Sattler</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Saake</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">8th Int. Conf. on Data Engineering (ICDE)</title>
				<meeting><address><addrLine>San Jose, CA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2002">2002</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Stratum Approaches to Temporal DBMS Implementation</title>
		<author>
			<persName><forename type="first">K</forename><surname>Torp</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">S</forename><surname>Jensen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">T</forename><surname>Snodgrass</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of IDEAS</title>
				<meeting>IDEAS<address><addrLine>Cardiff, Wales</addrLine></address></meeting>
		<imprint>
			<date type="published" when="1998">1998</date>
			<biblScope unit="page" from="4" to="13" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Research Problems in Data Warehousing</title>
		<author>
			<persName><forename type="first">J</forename><surname>Widom</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 4th Int&apos;l Conference on Information and Knowledge Management (CIKM)</title>
				<meeting>the 4th Int&apos;l Conference on Information and Knowledge Management (CIKM)</meeting>
		<imprint>
			<date type="published" when="1995">1995</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Duplicate removal in information dissemination</title>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">W</forename><surname>Yan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Garcia-Molina</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of VLDB-95</title>
				<meeting>VLDB-95<address><addrLine>Irvine, CA</addrLine></address></meeting>
		<imprint>
			<publisher>Information Systems</publisher>
			<date type="published" when="1995-09">September 1995. 2002</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
