<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Leveraging Probabilistic Existential Rules for Adversarial Deduplication</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Jose</forename><forename type="middle">N</forename><surname>Paredes</surname></persName>
							<email>jose.paredes@cs.uns.edu.ar</email>
							<affiliation key="aff0">
								<orgName type="department">Dept. of Computer Science and Engineering</orgName>
								<orgName type="institution">Universidad Nacional del Sur (UNS) Institute for Computer Science and Engineering (UNS-CONICET)</orgName>
								<address>
									<addrLine>San Andres 800</addrLine>
									<postCode>8000)</postCode>
									<settlement>Bahia Blanca</settlement>
									<country key="AR">Argentina</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Maria</forename><surname>Vanina Martinez</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Dept. of Computer Science and Engineering</orgName>
								<orgName type="institution">Universidad Nacional del Sur (UNS) Institute for Computer Science and Engineering (UNS-CONICET)</orgName>
								<address>
									<addrLine>San Andres 800</addrLine>
									<postCode>8000)</postCode>
									<settlement>Bahia Blanca</settlement>
									<country key="AR">Argentina</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Gerardo</forename><forename type="middle">I</forename><surname>Simari</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Dept. of Computer Science and Engineering</orgName>
								<orgName type="institution">Universidad Nacional del Sur (UNS) Institute for Computer Science and Engineering (UNS-CONICET)</orgName>
								<address>
									<addrLine>San Andres 800</addrLine>
									<postCode>8000)</postCode>
									<settlement>Bahia Blanca</settlement>
									<country key="AR">Argentina</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Marcelo</forename><forename type="middle">A</forename><surname>Falappa</surname></persName>
							<email>mfalappa@cs.uns.edu.ar</email>
							<affiliation key="aff0">
								<orgName type="department">Dept. of Computer Science and Engineering</orgName>
								<orgName type="institution">Universidad Nacional del Sur (UNS) Institute for Computer Science and Engineering (UNS-CONICET)</orgName>
								<address>
									<addrLine>San Andres 800</addrLine>
									<postCode>8000)</postCode>
									<settlement>Bahia Blanca</settlement>
									<country key="AR">Argentina</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Leveraging Probabilistic Existential Rules for Adversarial Deduplication</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">58BBB890A1153A49B962B900E3213EB2</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T07:32+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>The entity resolution problem in traditional databases, also known as deduplication, seeks to map multiple virtual objects to its corresponding set of real-world entities. Though the problem is challenging, it can be tackled in a variety of ways by means of leveraging several simplifying assumptions, such as the fact that the multiple virtual objects appear as the result of name or attribute ambiguity, clerical errors in data entry or formatting, missing or changing values, or abbreviations. However, in cyber security domains the entity resolution problem takes on a whole different form, since malicious actors that operate in certain environments like hacker forums and markets are highly motivated to remain semi-anonymous-this is because, though they wish to keep their true identities secret from law enforcement, they also have a reputation with their customers. The above simplifying assumptions cannot be made in this setting, and we therefore coin the term "adversarial deduplication". In this paper, we propose the use of probabilistic existential rules (also known as Datalog+/-) to model knowledge engineering solutions to this problem; we show that tuple-generating dependencies can be used to generate probabilistic deduplication hypotheses, and equality-generating dependencies can later be applied to leverage existing data towards grounding such hypotheses. The main advantage with respect to existing deduplication tools is that our model operates under the open-world assumption, and thus is capable of modeling hypotheses over unknown objects, which can later become known if new data becomes available.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>Deduplication of objects in databases, also known as entity resolution, is a classical problem in data cleaning <ref type="bibr" target="#b17">[18]</ref>; the basic idea is that databases contain objects that are potential duplicates-seemingly different records that correspond to the same entity or object in the real world-that need to be identified and merged <ref type="bibr" target="#b7">[8,</ref><ref type="bibr" target="#b13">14]</ref>. There has been a lot of interest on this topic in the last 20 years, yielding a large body of work that lies mostly in the databases literature. Traditionally, entities are resolved using pairwise similarity over the attributes of reference <ref type="bibr" target="#b17">[18]</ref>. The most recent and promising proposal are those based on so-called collective entity resolution <ref type="bibr" target="#b5">[6]</ref>, which exploits additional relational information in the data since references to different entities may co-occur. For instance, one such approach, called iterative blocking <ref type="bibr" target="#b29">[30]</ref>, iteratively groups together matching records in blocks, giving rise to the possibility of using the information processed so far to inform further decisions. Data-driven approaches <ref type="bibr" target="#b5">[6,</ref><ref type="bibr" target="#b6">7,</ref><ref type="bibr" target="#b12">13</ref>] leverage examples to determine if which pairs of records are duplicates. Recently, <ref type="bibr" target="#b2">[3]</ref> has defined a declarative framework for entity resolution based on matching dependencies (MDs). This formalism, which was first introduced in <ref type="bibr" target="#b15">[16,</ref><ref type="bibr" target="#b16">17]</ref>, consists of declarative rules that generalize entity resolution tasks, eliminating duplicates via a matching process. These rules state that certain attribute values in relational tuples, under certain similarity conditions over possibly other attribute values in those tuples, must be made the same. The original semantics for MDs, defined in <ref type="bibr" target="#b16">[17]</ref>, was redefined and extended in <ref type="bibr" target="#b4">[5]</ref>.</p><p>All of the above-mentioned methods operate under the assumption that the undesired situation that multiple records refer to the same real-world object is caused by a mix of clerical errors in data entry (simple typos), ambiguity in names or other attributes (consider, for instance, the result of transliterations such as "Kurt Gödel" being mapped to "Kurt Goedel"), inconsistent abbreviations and formatting, missing values (for instance, phone number not provided), or changing values (for example, one record has an old phone number while others have an updated one). Though these assumptions adequately reflect many real-world scenarios, we are interested in others in which they cannot be assumed to hold. One example of such a scenario is in cyber-security domains such as malicious hacker forums on the dark/deep web (the part that is not directly accessible by traditional browsers and search engines)-the entity resolution problem in this setting becomes quite different. In <ref type="bibr" target="#b25">[26]</ref>, a system for cyber threat intelligence was developed that gathers data from various social platforms on the Internet, focusing particularly on sites in the dark net and deep net. They collect and store information from hacker forum discussions and marketplaces offering products and services that focus on malicious hacking. The system is designed to extract specific information from marketplaces (regarding sales of malware/exploits) and hacker forums (discussions regarding services and threats). This extracted information is well-structured and can be stored in a relational database; they maintain two databases, one for marketplaces and the other for forums. The crawling and parsing of these sites is carried out periodically so that time-varying data can be collected. For markets, they collect product fields such as item title, item description, vendor name, CVE number (identifier given by the National Vulnerability Database <ref type="bibr" target="#b10">[11,</ref><ref type="bibr" target="#b24">25]</ref>), shipping details, item reviews, etc. For forums, they collect other fields such as topic content, post content, topic author, post author, author status, reputation, and topic interest.</p><p>These forums and marketplaces have an interesting characteristic: though malicious actors of course wish to remain anonymous from law enforcement agents that might be patrolling, they also enjoy a reputation with their customers and readers, and so they also wish to remain identifiable. In particular, the same actor typically operates under different screen names, keeping certain aspects constant so that others can recognize them. Furthermore, and perhaps most importantly, they also leave involuntary traces behind that can be leveraged in deduplication efforts. We refer to this as the adversarial deduplication/entity resolution problem. The following is a simplified example of the kind of information that we can obtain from dark web forums and marketplaces.</p><p>Example 1. As a running example, consider the simplified domain of the malicious hacker forums. Consider the following relational schema regarding information about users and their posts in forums:</p><formula xml:id="formula_0">User(Id u, Nick), Post(Id po, Id u, Id t), HasPosted(Id u, Id t), WritStSim(Id u, Id u),</formula><p>which represent users of these forums, posts by users, topics in posts, which users have posted on which topics, and the writing style similarity between users, respectively. Furthermore, id u and id po are the keys for relations User and Post, respectively. Let D 1 be an instance of this schema, where t 1 is "private sqli tool", t 2 is "Sql Injection", and t 3 is "trojan servers as a backdoor":</p><formula xml:id="formula_1">D 1 = { d 1 : Post(p 1 , a 1 , t 1 ), d 2 : Post(p 2 , a 2 , t 2 ), d 3 : Post(p 3 , a 3 , t 3 ), d 4 : User(a 1 , 'Blackstar'), d 5 : User(a 2 , 'Archer'), d 6 : User(a 3 , 'Angel Eyes'), d 7 : WritStSim(a 1 , a 2 ) }</formula><p>Finally, another basic assumption generally made by traditional approaches to entity resolution-which can be seen as a consequence of the other assumptionsis that the information contained in the databases is complete. In our setting, on the other hand, it may be the case that the relevant information is yet to be discovered; in these settings, the open-world assumption becomes crucial, allowing us to entertain deduplication hypotheses over unnamed entities. In the next section, we present a formalism that allows for this kind of reasoning, as well as deal with probabilistic uncertainty.</p><p>The rest of this paper is organized as follows: Section 2 presents the extension of Datalog+/-with probabilistic annotations and stratified negation, Section 3 presents a two-stage chase procedure that extends the classical chase with propagation of probabilistic events and EGDs. Sections 4 and 5 introduces deduplication programs and deduplication threshold queries that can be answered via the two-stage chase, respectively. We discuss conclusions and future work in Section 6.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Probabilistic Datalog+/-with Stratified Negation</head><p>In this section, we present the basics of the language of Datalog+/-from <ref type="bibr" target="#b21">[22]</ref>, its extension with probabilistic uncertainty from <ref type="bibr" target="#b18">[19]</ref>, and extend it with stratified negation; we redefine some of the basic concepts to adjust them to our presentation.</p><p>We assume the existence of the following pairwise disjoint (countably infinite) sets: a set ∆ of constants, a set V of variables, and a set N of nulls. Different constants represent different values (unique name assumption), while different nulls may represent the same value. We assume a lexicographic order on ∆ ∪ N , with every symbol in N following all symbols in ∆. We denote by X sequences of variables X 1 , . . . , X k with k &gt; 0.</p><p>Consider relational schemas R with a possibly infinite data domain ∆, a finite set of database predicates (relations) R ∈ R, and a set of built-in predicates for evaluating "=", "¬", "≤", comparisons, and arithmetic operations. Each relation R ∈ R has an associated arity n, thus R has the form R(A 1 , . . . , A n ), where each attribute A i has associated domain Dom(A i ) ⊆ ∆. For the sake of simplicity we may assume that the A i 's are different, and different predicates do not share attributes. However, different attributes may share the same domain. An atomic formula (or atom) has the form P (t 1 , . . . , t n ), where t i ∈ Dom(A i )∪V. A conjunction of atoms is often identified with the set of all its atoms. A database instance D for R is a (possibly infinite) set of ground atoms (or tuples) with predicates from R and arguments from ∆.</p><p>A homomorphism is a mapping µ :</p><formula xml:id="formula_2">∆ ∪ V ∪ N → ∆ ∪ V ∪ N such that (i) if c ∈ ∆ then µ(c) = c, (ii) if c ∈ V ∪ N then µ(c) ∈ ∆ ∪ N ,</formula><p>and (iii) µ is naturally extended to atoms, sets of atoms, and conjunctions of atoms, that is, µ(p(t 1 , . . . , t n )) = p(µ(t 1 ), . . . , µ(t n )). A grounding is a homomorphism ρ such that ρ(c) ∈ ∆ for every c ∈ ∆ ∪ V ∪ N . Also, given a conjunction of atoms Φ(X), we denote with pos(Φ(X)) the conjunction of atoms in Φ(X) that appear positive, and with neg(Φ(X)) the conjunction of atoms in Φ(X) that appear negated.</p><p>TGDs. A tuple generating dependency (TGD) is a first-order formula:</p><formula xml:id="formula_3">(σ) ∀XY Φ(X, Y) → ∃ZΨ (X, Z)<label>(1)</label></formula><p>where X, Y, Z are sequences of variables, Φ(X, Y) is a conjunction of atoms and negated atoms over R (called the body of σ, denoted body(σ)), and Ψ (X, Z) is a conjunction of atoms over R (called the head of σ, denoted head(σ)). We use body + (σ) and body − (σ) to denote pos(Φ(X, Y)) and neg(Φ(X, Y)), respectively. For ease of presentation, we sometimes omit the universal quantifiers, and assume that all variables without quantifier are universally quantified. We restrict the language so that all the variables in body − (σ) appear already in some atom in body + (σ). Furthermore, we assume Σ is stratified <ref type="bibr" target="#b19">[20]</ref>, i.e., Σ can be partitioned into mutually disjoint sets Σ 1 , . . . , Σ k such that:</p><p>-If Σ i contains a predicate A that appears positively in some rule, then there is no rule in Σ i+1 ∪ . . . ∪ Σ k with A in the head. -If Σ i contains a predicate A that appears negated in some rule, then there is no rule in Σ i ∪ . . . ∪ Σ k with A in the head.</p><p>We call Σ 1 , . . . , Σ k a stratification of Σ. As in the case of standard TGDs <ref type="bibr" target="#b21">[22]</ref>, w.l.o.g., we can assume that Ψ (X, Z) is a single atom (any set not satisfying this condition can be translated to an equivalent set that does).</p><p>Definition 1. Let σ be a TGD with body(σ) = ∀XY Φ(X, Y). We say that body(σ) is satisfied in D through homomorphism h if and only if h is such that</p><formula xml:id="formula_4">h(pos(Φ(X, Y))) ⊆ D and h(neg(Φ(X, Y))) ⊆ D.</formula><p>A TGD σ is satisfied in D if and only if whenever the body of σ is satisfied through homomorphism h, then there exists an extension h of h such that h (head(σ)) ⊆ D.</p><p>EGDs. Equality generating dependencies (EGDs) are first order formulas of the form: ∀ XΦ(X) → X i = X j , where Φ(X) is a conjunction of atoms, X is a sequence of variables, and X i , X j ∈ X. As with TGDs, we sometimes omit the universal quantifiers.</p><p>The body of an EGD , denoted body( ), corresponds to Φ(X), and the head, denoted head( ), corresponds to the equality atom on the right hand side of the rule. An instance D satisfies an EGD , denoted D |= , if whenever there exists a homomorphism h such that h(body(Φ)) ⊆ D, then it must be the case that h(X i ) = h(X j ). Otherwise, we say that is violated in D.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Probabilistic Annotations</head><p>Following <ref type="bibr" target="#b18">[19]</ref>, we further assume the existence of a finite set of constants ∆ M , a finite set of predicate names R M such that R ∩ R M = ∅, and an infinite set of variables V M , such that V ∩ V M = ∅. These sets give rise to corresponding Herbrand bases and Herbrand universes consisting of all possible ground terms and atoms, respectively; these are denoted with H M and U M , respectively.</p><p>Let M be a probabilistic model that represents a joint probability distribution over a (finite) set of random Boolean variables X = {X 1 , . . . , X n } such that there is a 1-to-1 mapping from X to the set of all ground atoms over H M . Examples of the type of probabilistic models that we assume in this work are Markov Logic <ref type="bibr" target="#b27">[28]</ref>, Bayesian networks <ref type="bibr" target="#b26">[27]</ref>, etc.</p><p>A (probabilistic) annotation λ, relative to M , is a (finite) set of expressions A = x i , where A is an atom over R M , ∆ M , and V M , x i ∈ {0, 1}, and the X i 's are pairwise distinct. A probabilistic annotation is valid if and only if for any two different expressions A = x, B = y ∈ λ, there does not exist a substitution θ such that θ(A) = θ(B). A probabilistic scenario is a valid probabilistic annotation λ for which |λ| = |X| and all A = x i are such that A is ground. We use scn(M ) to denote the set of scenarios in M . A probabilistic annotation is equivalent to a formula that consists of the disjunction of all the scenarios that belong to the set denoted by the annotation. Therefore, with a slight abuse of notation, we sometimes combine annotations with logical operators.</p><p>Intuitively, a probabilistic annotation is used to describe an event in which the random variables in a probabilistic model M are compatible with the settings of the random variables described by λ, i.e., each X i has the value x i . We will attach probabilistic annotations λ to formulas in our language to produce annotated formulas F : λ, where F is an atom, a TGD, or an EGD. Intuitively, the annotation means that F holds whenever the event associated with λ occurs. Note that whenever a random variable's value is left unspecified in a probabilistic annotation, the variable is unconstrained; in particular, a formula annotated with an empty probabilistic annotation means that the formula holds in every world. An annotated formula F : λ is referred to as a probabilistic atom whenever F is atom, a probabilistic TGD (pTGD) if F is a TGD, and a probabilistic EGD (pEGD) whenever F is an EGD. Definition 2. A probabilistic Datalog+/-knowledge base is a triple KB = (O, M, af), where O is a finite set of atoms, TGDs, and EGDs, M is probabilistic model, and af is an annotation function that maps formulas to probabilistic annotations relative to M .</p><p>For ease of presentation, we sometimes use notation "σ : e" to denote "af(σ) = e". Given a probabilistic Datalog+/-KB KB = (O, M, af) and a scenario λ ∈ scn(M ), O λ denotes the (non-probabilistic) Datalog+/-knowledge base induced by λ. Formally, O λ consists of all formulas F i (atoms, TGDs, and EGDs) such that λ i ⊆ λ for some</p><formula xml:id="formula_5">F i : λ i ∈ O.</formula><p>Example 2. Consider the DB from Example 1, and let KB 1 = (O 1 , M 1 , af 1 ) be a Datalog+/-KB. Then, O 1 could contain the following TGD ω 1 :</p><formula xml:id="formula_6">(ω 1 ) User(UID, N ) ∧ hasPosted(UID, t 1 ) → hasPosted(UID, t 2 )</formula><p>The intuition of this rule is: "Every User that has posts on topic t 1 , also has posts on topic t 2 ". The annotation function af 1 is defined such that af 1 (ω 1 ) = e 1 (t 1 , t 2 ) and e 1 (t 1 , t 2 ) is an event associated with the probability that a user posted something related to t 2 given that he has posted something related to t 1 .</p><p>O 1 could contain the following EGD 1</p><formula xml:id="formula_7">( 1 ) User(UID 1 , N ) ∧ User(UID 2 , N ) → UID 1 = UID 2 .</formula><p>The intuition of this rule is: "Two users cannot share the same user name.". This rule is annotated with the empty event, that is af 1 ( 1 ) = ∅-therefore, it always holds.</p><p>Probabilistic model M 1 in the above example can be learned from domain information, and it allows to infer probabilities about possible duplicate user and correlations regarding certain events, such as posts by users given specific topics. We will discuss this in greater detail in Section 4.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">A Two-stage Chase Procedure</head><p>We now extend the classical operational semantics for Datalog+/-, based on the chase procedure, for the case of the extensions discussed above. This procedure provides the semantics for the application of rules and constraints, extending the well-known (oblivious) chase algorithm for testing implications of data dependencies <ref type="bibr" target="#b3">[4,</ref><ref type="bibr" target="#b22">23]</ref> or completing deductive databases <ref type="bibr" target="#b0">[1]</ref>. The chase provides a mechanism for repairing a database instance relative to a set of dependencies, so that the result of the chase (a database instance) satisfies the dependencies. Note that, though <ref type="bibr" target="#b18">[19]</ref> presents a chase procedure for probabilistic Datalog+/ontologies, our two-stage chase includes stratified negation and EGDs.</p><p>The building block for the chase procedure is the TGD chase rule; in our setting, we must also deal with the propagation of probabilistic annotations.</p><p>pTGD Chase Rule. We will now define the probabilistic TGD chase rule for a knowledge base KB = (O = (D, Σ), M, af). Let σ ∈ O be a pTGD of the form ∀XY Φ(X, Y) → ∃ZΨ (X, Z) : e σ ; then, σ is applicable to D if and only if there exists a homomorphism h such that one the following cases occur: a) 1) h maps the atoms in pos(Φ(X, Y)) and neg(Φ(X, Y)) to probabilistic atoms {α 1 : e 1 , ..., α k : e k } ⊆ D and {α k+1 : e k+1 , ..., α j : e j } D, and 2) h(e σ ∧ i=1,...,k e i ) is a valid probabilistic annotation. b) 1) h maps the atoms in pos(Φ(X, Y)) and neg(Φ(X, Y)) to probabilistic atoms {α 1 : e 1 , ..., α k : e k } ⊆ D and {α k+1 : e k+1 , ..., α j : e j } ⊆ D, and 2) h(e σ ∧ i=1,...,k e i ∧ i=k+1,...,j ¬e i ) is a valid probabilistic annotations.</p><p>Let σ be applicable to D, and h be a homomorphism that extends h as follows: for each term X i ∈ X, h (X i ) = h(X i ) and h (Z) = z j , where Z is existentially quantified in head(σ) and z j is a "fresh" null, i.e., z j ∈ ∆ N , z j does not occur in D, and z j lexicographically follows all other nulls already introduced. The application of σ on D adds to D one of the following probabilistic atoms according to the case where σ is applicable: a) h (Ψ (X, Z)) : h (e σ ∧ i=1,...,k e i ). b) h (Ψ (X, Z)) : h (e σ ∧ i=1,...,k e i ∧ i=k+1,...,j ¬e i ).</p><p>As a second step, we apply the equality-generating dependencies. We do this in a separate step over the result of the pTGD chase in order to avoid undesired interactions between TGDs and EGDs that can lead to undecidability <ref type="bibr" target="#b21">[22]</ref>, since the standard way to avoid this via separability is too strong an assumption in our case (see Section 4). pEGD Chase Rule. Let KB = (O = (D, Σ), M, af) be a probabilistic Datalog+/-KB, and ∈ Σ be a pEGD of the form ∀X Φ(X) → X m = X n : e . Then, is applicable to D if and only if there exists a homomorphism h such that: a) h maps the atoms in Φ(X) to probabilistic atoms {α 1 : e 1 , ..., α k : e k } ⊆ D, and b) h(e ∧ i=1,...,k e i ) is a valid probabilistic annotation.</p><p>Let be applicable to D, and Φ Xm = {α 1 : e 1 , ..., α l : e l } ⊆ {α 1 : e 1 , ..., α k : e k } be the set of probabilistic atoms where X m occurs in α 1≤i≤l ; then, the application of on D results in:</p><p>(i) The chase fails if h(X m ) and h(X n ) are constants; this is called a hard violation. Otherwise, (ii) for all α i : e i ∈ Φ Xm replace h(X m ) with h(X n ) and e i with h(e ∧ i=1,...,k e i ), if h(X m ) is a null and h(X n ) is a constant, or h(X m ) precedes h(X n ) in the lexicographic order.</p><p>The following section shows how this chase procedure can be used to answer probabilistic deduplication queries.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Deduplication Programs</head><p>In this section we define a subclass of Probabilistic Datalog+/-programs that are specifically tailored to declaratively capture the requirements for a deduplication task in an open-world, adversarial environment.</p><p>Towards this end, we assume that there exists a set E ⊆ R that we identify as entity relations, i.e., relations that represent entities of interest in the application domain. Tuples in these relations have identifiers-keys-that allows us to compare extensions of the same predicate in different instances and to trace changes to them. Tuple identifiers can be accommodated by simply adding to each predicate E ∈ E an extra attribute (usually denoted with T ) that acts as a primary key in E. Finally, there is also a predefined predicate called dedup hypth.</p><p>Our formalism is inspired in (relational) matching dependencies (RMDs) <ref type="bibr" target="#b4">[5]</ref>, which provide a declarative way of defining conditions under which records (or attribute values) should be made equal-they also establish how this matching is supposed to be done by means of matching functions. However, as we discussed before, RMDs are developed under a closed-world assumption, i.e., everything that is not contained in the database instance is assumed to be false (or to not exist). Nevertheless, we argue that in the type of domain applications we are interested in, this assumption is not realistic and therefore we must agree in that the database instances we are dealing with are potentially incomplete. Under such scenarios, RMDs are not expressive enough to capture all potential matchings, since there might be matchings for which we cannot establish specific conditions to find them but rather just make hypotheses based on (weaker) evidence of other nature-the possibility of value invention afforded by existential rules thus becomes crucial.</p><p>To specify the kind of rules we require, we slightly modify pTGDs so they can be used to infer potential matchings (or matching hypotheses). For this, we assume that for every attribute A j appearing in a n-nary relation R ∈ R, where the schema of R is R(A 1 , . . . , A n ), there exists a binary similarity relation ≈ Aj that is reflexive and symmetric. Thus, they are capable of expressing similarity conditions over attribute values as done in RMDs, since the "≈" predicates can simply occur in the body.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>dhTGDs. A Duplication Hypothesis Tuple Generating Dependency (or dhTGD)</head><p>σ is simply a pTGD of the following form: where X, Y, Z are sequences of variables, X and Z are variables and X ∈ X, Z ∈ Z, Φ(X, Y) is a conjunction of atoms and negated atoms over R, and Γ is also a conjunction of atoms.</p><formula xml:id="formula_8">∀XY Φ(X, Y) → ∃Zdedup hypth(X, Z) ∧ Γ (X, Z) r 1  r i  r n body(r i )  …  not(head(r i ))   dedup_hyp  …</formula><p>dhTGDs are a special class of pTGDs that generate deduplication hypothesesessentially, they allow to create hypotheses about potential matchings, which are kept and later confirmed or discarded whenever new information is added to (or extracted from) the data sources. Informally, dhTGDs capture situations in which we have reasons to believe that entities may be duplicated but we do not know which other entity could correspond to them. Figure <ref type="figure" target="#fig_0">1</ref> describes one possible way in which dhTGDs could be automatically derived from data. A relational association rule mining algorithm (such as <ref type="bibr" target="#b11">[12]</ref>) is used to discover patterns in the available data sources, which are rules of the form r : body(r) → head(r)); therefore, entities that satisfy the conditions in the body but not those in the head are potential candidates for entities that have duplicate identities-see Example 3 below for a concrete scenario.</p><p>pEGDs as companions to dhTGDs. As we saw in Section 3, equalitygenerating dependencies are applied as a second step after all (dh)TGDs are applied; the goal of this step is to complete the database with the information that might be available with respect to the null values in the dedup hypth atoms. One kind of pEGDs that might be useful can be automatically derived from dhTGDs σ by taking the negation of body − (σ) as the body of the pEGD. Figure <ref type="figure">2</ref> illustrates this procedure, and the following is a simple example based on the running example.</p><p>Example 3. Consider the running example, and suppose we have obtained a relational association rule that states that "if a user has posts related to topic t 1 , then he has posts related to topic t 2 ". This can be translated into the following dhTGD:</p><formula xml:id="formula_9">(ω 2 ) User(UID, N ) ∧ hasPosted (UID, t 1 ) ∧ not(hasPosted (UID, t 2 )) → ∃Z dedup hypth(UID, Z) ∧ hasPosted(Z, t 2 ).</formula><p>(𝜔 2 ) 𝑈𝑠𝑒𝑟(𝑈𝐼𝐷, 𝑁) ∧ ℎ𝑎𝑠𝑃𝑜𝑠𝑡𝑒𝑑(𝑈𝐼𝐷, 𝑡 1 ) ∧ 𝑛𝑜𝑡(ℎ𝑎𝑠𝑃𝑜𝑠𝑡𝑒𝑑(𝑈𝐼𝐷, 𝑡 2 )) → ∃𝑍 𝑑𝑒𝑑𝑢𝑝_ℎ𝑦𝑝(𝑈𝐼𝐷, 𝑍) ∧ ℎ𝑎𝑠𝑃𝑜𝑠𝑡𝑒𝑑(𝑍, 𝑡 2 ) same atoms of body  ( 2 ) dedup_hyp atom body  ( 2 )</p><formula xml:id="formula_10">(𝜖 2 ) 𝑑𝑒𝑑𝑢𝑝_ℎ𝑦𝑝(𝑈𝐼𝐷 1 , 𝑍) ∧ 𝑈𝑠𝑒𝑟(𝑈𝐼𝐷 2 , 𝑁) ∧ ℎ𝑎𝑠𝑃𝑜𝑠𝑡𝑒𝑑(𝑈𝐼𝐷 2 , 𝑡 2 ) ∧ 𝑊𝑟𝑡𝑆𝑡𝑆𝑖𝑚(𝑈𝐼𝐷 1 , 𝑈𝐼𝐷 2 ) → 𝑍 = 𝑈𝐼𝐷 2 dedup_hyp atom same atoms of body  ( 2 )</formula><p>atoms to find similar entity Fig. <ref type="figure">2</ref>. Illustration of one possible procedure with which a pEGD can be derived from an existing dhTGD (σ)-the (unnegated) atoms from body − (σ) plus additional conditions that help find candidate entities to instantiate the deduplication hypothesis.</p><p>The intuition of this rule is "If a user that has some post on topic t 1 but he does not have any post on topic t 2 , then we generate the hypothesis that there exists another user corresponding to him (whose identity is currently unknown) who has posted on topic t 2 ".</p><p>As shown in Figure <ref type="figure">2</ref>, an equality-generating dependency can be derived from ω 2 by taking the atoms from body − (ω 2 ) and some additional conditions that, when true, can help obtain the identity of the hitherto unknown entity that participates in the deduplication hypothesis:</p><formula xml:id="formula_11">( 2 ) dedup hypth(UID 1 , Z) ∧ User(UID 2 , N ) ∧ hasPosted(UID 2 , t 2 ) ∧ WritStSim(UID 1 , UID 2 ) → Z = UID 2 .</formula><p>The intuition of this rule is "If we have a duplicate user hypothesis between user UID 1 and an unknown user Z, and user UID 2 has posts on topic t 2 with similar writing style to user UID 1 , then Z may correspond to UID 2 ".</p><p>Putting all of this together, the following example illustrates how the chase procedure unfolds over the scenario in the running example. af(ω 2 ) = e 1 (t 1 , t 2 ), where the event e 1 (t 1 , t 2 ) is associated with the probability that a user has some post on t 2 given that he has some post on t 1 .</p><p>dedup_hypth(a 1 ,a 2 )</p><formula xml:id="formula_12">dedup_hypth(a 1 ,z 1 ) hasPosted(z 1 , t 2 )</formula><p>Post(p 3 ,a  af( 2 ) = e 2 , where the event e 2 is associated with the probability that two users with similar writing style actually correspond to the same underlying user.</p><p>Figure <ref type="figure">3</ref> illustrates the chase for this example. Note that dhTGD ω 2 is applicable because body + (ω 2 ) is mapped to D 1 and body − (ω 2 ) is not mapped to D 1 ; also note how the event e 1 (t 1 , t 2 ) is propagated to the generated atoms. Then, pEGD 2 is applicable, so null value z 1 is instantiated and the events e 1 (t 1 , t 2 ) and e 2 are propagated. Finally, for simplicity we specify M directly as a joint probability distribution, considering all possible worlds λ i for the events e 1 (t 1 , t 2 ) and e 2 (cf. Figure <ref type="figure" target="#fig_1">4</ref>).</p><p>In the previous example, that the application of the dhTGD ω 2 produces a deduplication hypothesis for user a 2 , but the lack of information causes the candidate for such a duplicate to remain as a null value. However, the subsequent application of pEGD 2 lets us conclude that user a 2 -who has a similar writing cf. Figure <ref type="figure">2</ref> Fig. <ref type="figure">5</ref>. Overview of the application of probabilistic Datalog+/-to answer deduplication queries. A set of input databases is considered along with a set of dhTGDs. In the first stage of the chase procedure, a set of dedup hypth atoms is generated; in the second stage, a set of pEGDs is used to assign concrete values to nulls in such atoms-both stages involve keeping track of probabilistic events stating conditions under which inferences hold. The output is a set of dedup hypth atoms with probability values, which can be used to answer the input query.</p><p>style to a 1 -is a likely option (event e 1 (t 1 , t 2 ) ∧ e 2 has probability 0.76 according to Figure <ref type="figure" target="#fig_1">4</ref>). Note that the dhTGD could be applied multiple times, so that if there are other possible candidates, fresh dedup hypth atoms can be generated as needed.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Finding Duplicates: Threshold Deduplication Queries</head><p>The problem of identifying seemingly different entities within a knowledge base (or set of knowledge bases) that actually correspond to the same real world entity can be addressed as a query answering problem: given a knowledge base, we want to have a principled way of answering questions regarding the occurrence of duplicate entities. The chase procedure presented above can be used to answer queries of the form:</p><formula xml:id="formula_13">Q(X) = dedup(eid, X, θ),</formula><p>where eid is a specific entity of interest, and θ ∈ [0, 1] represents a threshold probability value. Answers to queries of this form simply consist of possible values for variable X such that the chase procedure generates atoms dedup hypth(eid,val ) with an associated event e with Pr M (e) ≥ θ. Figure <ref type="figure">5</ref> illustrates the overall query answering process.</p><p>The following result states that the two-stage chase procedure described in Section 3 can be stopped after a finite number of steps in order to answer threshold deduplication queries whenever the underlying classical TGDs belong to a fragment of Datalog+/-for which conjunctive queries are decidable; this can be guaranteed via syntactic conditions such such as guardedness <ref type="bibr" target="#b21">[22]</ref>, stickiness <ref type="bibr" target="#b9">[10]</ref>, acyclicity, or their weak counterparts <ref type="bibr" target="#b8">[9,</ref><ref type="bibr" target="#b9">10,</ref><ref type="bibr" target="#b14">15]</ref>, or semantic conditions such as shyness <ref type="bibr" target="#b20">[21]</ref>, finite expansion sets (fes), bounded treewidth sets (bts), and finite unification sets (fus)) <ref type="bibr" target="#b1">[2]</ref>-cf. <ref type="bibr" target="#b28">[29,</ref><ref type="bibr">Section 1.3.3]</ref> for a brief overview of many such conditions. Proposition 1. Let KB = (O = (D, Σ), M, af) be a probabilistic Datalog+/ontology with stratified negation, and Q(X) = dedup(eid, X, θ) be a threshold deduplication query. If O belongs to a fragment of Datalog+/-for which a finite portion of chase(D, Σ) can be computed in order to answer conjunctive queries, then there exists a finite portion of the two-stage probabilistic chase procedure that can be used to answer threshold deduplication query Q over KB.</p><p>Essentially, this result holds since the first stage of our chase procedure applies TGDs and dhTGDs (which are a special case of classical TGDs) in the order given by the stratification, and the second stage is simply an application of pEGDs over a fixed structure. Both stages involve carrying over probabilistic events, which are then used in conjunction with the probabilistic model to check which atoms surpass the threshold given in the query.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6">Conclusions and Future Work</head><p>In this paper, we have introduced the adversarial deduplication problem as a variant of the classical deduplication (or entity resolution) problem in which one cannot assume that the existence of multiple records corresponding to the same real-world entity is the result of innocent or casual errors such as typos, transliterations, or inconsistently-updated values, but rather the result of actors deliberately trying to avoid identification. As a motivating scenario we took malicious hacker forums and marketplaces on the dark web, where this sort of practice is commonplace and having tools at hand to address it would have a great impact on efforts such as early-warning of cyber attacks, understanding the social networks of buyers and sellers, and identification of leading vendors of specific types of products.</p><p>We proposed the use of probabilistic Datalog+/-as a way to tackle the problem of finding pairs of virtual objects for which we have reasons to believe that they correspond to the same real-world entity, where a specific kind of probabilistic TGDs (called deduplication hypothesis-generating dependencies) and EGDs can be applied in order to derive deduplication hypotheses. The main strength of this approach is that it allows to reason about unknown objects via value invention, which is essential in this setting-not having this capability renders traditional entity resolution tools ineffective. Another advantage of our approach is that dhTGDs can be automatically obtained via relational association rule mining algorithms, and corresponding EGDs can also be derived with additional domain knowledge. Finally, we developed a two-stage probabilistic chase procedure in which (dh)TGDs are first exhaustively applied, and then pEGDs are used to associate concrete values to null values generated in the first step. We showed that this modified chase procedure can be used to answer deduplication threshold queries, leveraging chase termination results from the existential rules and Datalog+/-literature.</p><p>Ongoing and future work involves investigating other kinds of deduplication queries, such as instance checking ("what is the probability that two given virtual objects correspond to the same real-world entity?"), all pairs threshold ("which pairs of virtual objects are suspected to correspond to the same real-world entity with probability at least p?"), and more general conjunctive queries. We are also interested in investigating the complexity of query answering algorithms for all such queries, and under which conditions they can be guaranteed to run in polynomial time. We plan to evaluate the performance of our approach on both synthetic and real-world data on malicious hacker forums and marketplaces on the dark web obtained via a collaboration with a cyber threat intelligence company. Finally, we would also like to explore the relationship between our formalism and other approaches to probabilistic modeling with unknown objects, such as <ref type="bibr" target="#b23">[24]</ref> </p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Fig. 1 .</head><label>1</label><figDesc>Fig.1. Overview of a possible dhTGD learning process: relational association rule mining algorithms are applied over a set of available data sources; each such rule then leads to one output dhTGD characterizing situations that violate the corresponding association rule and therefore are the basis of a duplicate entity hypothesis.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Example 4 .</head><label>4</label><figDesc>Considering the DB D 1 presented in Example 1, a probabilistic Extended Datalog+/-KB representing a deduplication program can be KB 2 = (O 2 , M, af) where O 2 = (D 1 , Σ 2 ) and M 1 is the same probabilistic model from Example 2. The set Σ 2 contains the (classical) TGD: (σ) User(UID, N ) ∧ Post(Id po, UID, T ) → hasPosted (UID, T ), and also contains the dhTGD ω 2 and pEGD 2 from the previous example. The annotation function af is defined as follows:</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Fig. 4 .</head><label>4</label><figDesc>Fig.<ref type="bibr" target="#b3">4</ref>. Joint probability distribution over the events e1(t1, t2) and e2 used in the running example. In general, such distributions can be compactly specified via a probabilistic model like a Bayesian network.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head></head><label></label><figDesc>the confidence at least  with which it is believed to hold.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head></head><label></label><figDesc>3 ,t 3 ) User(a 3 ,'Angel Eyes') Graphic representation of the chase for Example 4. Boxes with bold outline represent atoms in D1, boxes with dashed outline represent atoms that contain null values, and boxes with standard outline represent atoms without nulls obtained from some dhTGD or pEGD. The events associated with each atom are represented underneath the corresponding box; all atoms in D1 are associated with empty events, and the events propagated are represented next to the arrow of the corresponding dhTGD or pEGD.</figDesc><table><row><cell cols="2">User(a 1 ,'Blackstar')</cell><cell>Post(p 1 ,a 1 ,t 1 )</cell><cell>User(a 2 ,'Archer')</cell><cell>Post(p 2 ,a 2 ,t 2 )</cell></row><row><cell></cell><cell>TGD </cell><cell></cell><cell>TGD </cell><cell>TGD </cell></row><row><cell></cell><cell cols="2">hasPosted(a 1 ,t 1 )</cell><cell cols="2">hasPosted(a 2 ,t 2 )</cell><cell>hasPosted(a 3 ,t 3 )</cell></row><row><cell>e 1 (t 1 ,t 2 )</cell><cell>dhTGD  2</cell><cell></cell><cell>pEGD  2</cell><cell>WritStSim(a 1 ,a 2 )</cell></row><row><cell></cell><cell></cell><cell>e 1 (t 1 ,t 2 )</cell><cell></cell></row><row><cell>e 1 (t 1 ,t 2 )</cell><cell></cell><cell></cell><cell>e 2</cell></row><row><cell></cell><cell></cell><cell></cell><cell>e 1 (t 1 ,t 2 )  e 2</cell></row><row><cell cols="5">Fig. 3. λi e1(t1, t2) e2 Probability</cell></row><row><cell></cell><cell></cell><cell cols="2">λ1 true true</cell><cell>0.76</cell></row><row><cell></cell><cell></cell><cell cols="2">λ2 true false</cell><cell>0.19</cell></row><row><cell></cell><cell></cell><cell cols="2">λ3 false true</cell><cell>0.04</cell></row><row><cell></cell><cell></cell><cell cols="2">λ4 false false</cell><cell>0.01</cell></row></table></figure>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Acknowledgments. This work was supported by funds provided by CONICET, Agencia Nacional de Promoción Científica y Tecnológica, Universidad Nacional del Sur, Argentina, by the EU H2020 research and innovation programme under the Marie Sklodowska-Curie grant agreement 690974 for the project "MIREL: MIning and REasoning with Legal texts", and by the US Department of the Navy, Office of Naval Research, grant N00014-15-1-2742. Any opinions, findings, and conclusions or recommendations expressed in this material are those of the authors and do not necessarily reflect the views of the Office of Naval Research.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<title level="m" type="main">Foundations of Databases: The Logical Level</title>
		<editor>Abiteboul, S., Hull, R., Vianu, V.</editor>
		<imprint>
			<date type="published" when="1995">1995</date>
			<publisher>Addison-Wesley Longman Publishing Co., Inc</publisher>
			<pubPlace>Boston, MA, USA</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Walking the complexity lines for generalized guarded existential rules</title>
		<author>
			<persName><forename type="first">J</forename><surname>Baget</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Mugnier</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Rudolph</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Thomazo</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proce. IJCAI</title>
				<meeting>e. IJCAI</meeting>
		<imprint>
			<date type="published" when="2011">2011</date>
			<biblScope unit="page" from="712" to="717" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Erblox: Combining matching dependencies with machine learning for entity resolution</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Bahmani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">E</forename><surname>Bertossi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Vasiloglou</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Int. J. Approx. Reasoning</title>
		<imprint>
			<biblScope unit="volume">83</biblScope>
			<biblScope unit="page" from="118" to="141" />
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">A proof procedure for data dependencies</title>
		<author>
			<persName><forename type="first">C</forename><surname>Beeri</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">Y</forename><surname>Vardi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">J. ACM</title>
		<imprint>
			<biblScope unit="volume">31</biblScope>
			<biblScope unit="issue">4</biblScope>
			<biblScope unit="page" from="718" to="741" />
			<date type="published" when="1984">1984</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Data cleaning and query answering with matching dependencies and matching functions</title>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">E</forename><surname>Bertossi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Kolahi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">V S</forename><surname>Lakshmanan</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Theory Comput. Syst</title>
		<imprint>
			<biblScope unit="volume">52</biblScope>
			<biblScope unit="issue">3</biblScope>
			<biblScope unit="page" from="441" to="482" />
			<date type="published" when="2013">2013</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Collective entity resolution in relational data</title>
		<author>
			<persName><forename type="first">I</forename><surname>Bhattacharya</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Getoor</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">ACM Trans. Knowl. Discov. Data</title>
		<imprint>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="issue">1</biblScope>
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Query-time entity resolution</title>
		<author>
			<persName><forename type="first">I</forename><surname>Bhattacharya</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Getoor</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">J. Artif. Intell. Res</title>
		<imprint>
			<biblScope unit="volume">30</biblScope>
			<biblScope unit="page" from="621" to="657" />
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Data fusion</title>
		<author>
			<persName><forename type="first">J</forename><surname>Bleiholder</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Naumann</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">ACM Comput. Surv</title>
		<imprint>
			<biblScope unit="volume">41</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="1" to="41" />
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Taming the infinite chase: Query answering under expressive relational constraints</title>
		<author>
			<persName><forename type="first">A</forename><surname>Calì</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Gottlob</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Kifer</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">J. Artif. Intell. Res</title>
		<imprint>
			<biblScope unit="volume">48</biblScope>
			<biblScope unit="page" from="115" to="174" />
			<date type="published" when="2013">2013</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Towards more expressive ontology languages: The query answering problem</title>
		<author>
			<persName><forename type="first">A</forename><surname>Calì</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Gottlob</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Pieris</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Artif. Intell</title>
		<imprint>
			<biblScope unit="volume">193</biblScope>
			<biblScope unit="page" from="87" to="128" />
			<date type="published" when="2012">2012</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<monogr>
		<ptr target="http://cve.mitre.org/" />
		<title level="m">CVE: Common vulnerabilities and exposures: The standard for information security vulnerability names</title>
				<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Discovery of relational association rules</title>
		<author>
			<persName><forename type="first">L</forename><surname>Dehaspe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Toivonen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Relational data mining</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2001">2001</date>
			<biblScope unit="page" from="189" to="212" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">An automatic system for identifying authorities in digital libraries</title>
		<author>
			<persName><forename type="first">I</forename><surname>Diaz-Valenzuela</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">J</forename><surname>Martin-Bautista</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">A</forename><surname>Vila</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">R</forename><surname>Campaña</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Expert Syst. Appl</title>
		<imprint>
			<biblScope unit="volume">40</biblScope>
			<biblScope unit="issue">10</biblScope>
			<biblScope unit="page" from="3994" to="4002" />
			<date type="published" when="2013">2013</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Duplicate record detection: A survey</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">K</forename><surname>Elmagarmid</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">G</forename><surname>Ipeirotis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><forename type="middle">S</forename><surname>Verykios</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE T. Knowl. Data En</title>
		<imprint>
			<biblScope unit="volume">19</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="1" to="16" />
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Data exchange: semantics and query answering</title>
		<author>
			<persName><forename type="first">R</forename><surname>Fagin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">G</forename><surname>Kolaitis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">J</forename><surname>Miller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Popa</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Theor. Comput. Sci</title>
		<imprint>
			<biblScope unit="volume">336</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="89" to="124" />
			<date type="published" when="2005">2005</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Dependencies revisited for improving data quality</title>
		<author>
			<persName><forename type="first">W</forename><surname>Fan</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. ACM PODS</title>
				<meeting>ACM PODS</meeting>
		<imprint>
			<date type="published" when="2008">2008</date>
			<biblScope unit="page" from="159" to="170" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Reasoning about record matching rules</title>
		<author>
			<persName><forename type="first">W</forename><surname>Fan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Jia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Ma</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">PVLDB</title>
		<imprint>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="407" to="418" />
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Entity resolution: Theory, practice and open challenges</title>
		<author>
			<persName><forename type="first">L</forename><surname>Getoor</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Machanavajjhala</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">PVLDB</title>
		<imprint>
			<biblScope unit="volume">5</biblScope>
			<biblScope unit="issue">12</biblScope>
			<biblScope unit="page" from="2018" to="2019" />
			<date type="published" when="2012">2012</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Query answering under probabilistic uncertainty in Datalog+/-ontologies</title>
		<author>
			<persName><forename type="first">G</forename><surname>Gottlob</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Lukasiewicz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">V</forename><surname>Martinez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">I</forename><surname>Simari</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Ann. Math. Artif. Intell</title>
		<imprint>
			<biblScope unit="volume">69</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="37" to="72" />
			<date type="published" when="2013">2013</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Expressiveness of guarded existential rule languages</title>
		<author>
			<persName><forename type="first">G</forename><surname>Gottlob</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Rudolph</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Simkus</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. ACM PODS</title>
				<meeting>ACM PODS</meeting>
		<imprint>
			<date type="published" when="2014">2014</date>
			<biblScope unit="page" from="27" to="38" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">Efficiently computable datalog∃ programs</title>
		<author>
			<persName><forename type="first">N</forename><surname>Leone</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Manna</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Terracina</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Veltri</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. KR</title>
				<meeting>KR</meeting>
		<imprint>
			<date type="published" when="2012">2012</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">A general Datalog-based framework for tractable query answering over ontologies</title>
		<author>
			<persName><forename type="first">T</forename><surname>Lukasiewicz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Cali</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Gottlob</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">J. Web Semant</title>
		<imprint>
			<biblScope unit="volume">14</biblScope>
			<biblScope unit="issue">0</biblScope>
			<date type="published" when="2012">2012</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">Testing implications of data dependencies</title>
		<author>
			<persName><forename type="first">D</forename><surname>Maier</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">O</forename><surname>Mendelzon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Sagiv</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">ACM T. Database Syst</title>
		<imprint>
			<biblScope unit="volume">4</biblScope>
			<biblScope unit="issue">4</biblScope>
			<biblScope unit="page" from="455" to="469" />
			<date type="published" when="1979">1979</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">BLOG: probabilistic models with unknown objects</title>
		<author>
			<persName><forename type="first">B</forename><surname>Milch</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Marthi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">J</forename><surname>Russell</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Sontag</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">L</forename><surname>Ong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Kolobov</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. IJCAI</title>
				<meeting>IJCAI</meeting>
		<imprint>
			<date type="published" when="2005">2005</date>
			<biblScope unit="page" from="1352" to="1359" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<monogr>
		<ptr target="https://nvd.nist.gov/(2018" />
		<title level="m">National Vulnerability Database</title>
				<imprint/>
		<respStmt>
			<orgName>NIST</orgName>
		</respStmt>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<analytic>
		<title level="a" type="main">Darknet and deepnet mining for proactive cybersecurity threat intelligence</title>
		<author>
			<persName><forename type="first">E</forename><surname>Nunes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Diab</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">T</forename><surname>Gunn</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Marin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Mishra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Paliath</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Robertson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Shakarian</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Thart</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Shakarian</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. ISI</title>
				<meeting>ISI</meeting>
		<imprint>
			<date type="published" when="2016">2016</date>
			<biblScope unit="page" from="7" to="12" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<monogr>
		<title level="m" type="main">Probabilistic Reasoning in Intelligent Systems: Networks of Plausible Inference</title>
		<author>
			<persName><forename type="first">J</forename><surname>Pearl</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1988">1988</date>
			<publisher>Morgan Kaufmann Publishers Inc</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b27">
	<analytic>
		<title level="a" type="main">Markov logic networks</title>
		<author>
			<persName><forename type="first">M</forename><surname>Richardson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Domingos</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Mach. Learn</title>
		<imprint>
			<biblScope unit="volume">62</biblScope>
			<biblScope unit="issue">1-2</biblScope>
			<biblScope unit="page" from="107" to="136" />
			<date type="published" when="2006">2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b28">
	<analytic>
		<title level="a" type="main">Ontology-Based Data Access Leveraging Subjective Reports</title>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">I</forename><surname>Simari</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Molinaro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">V</forename><surname>Martinez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Lukasiewicz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Predoiu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="s">Springer Briefs in Computer Science</title>
		<imprint>
			<date type="published" when="2017">2017</date>
			<publisher>Springer</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b29">
	<analytic>
		<title level="a" type="main">Entity resolution with iterative blocking</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">E</forename><surname>Whang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Menestrina</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Koutrika</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Theobald</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Garcia-Molina</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. SIGMOD. Stanford</title>
				<meeting>SIGMOD. Stanford</meeting>
		<imprint>
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
