<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Statistical Modeling vs. Machine Learning for Deduplication of Customer Records (industrial paper)</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Witold</forename><surname>Andrzejewski</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">Poznan University of Technology</orgName>
								<address>
									<settlement>Poznań</settlement>
									<country key="PL">Poland</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Bartosz</forename><surname>Bębel</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">Poznan University of Technology</orgName>
								<address>
									<settlement>Poznań</settlement>
									<country key="PL">Poland</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Paweł</forename><surname>Boiński</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">Poznan University of Technology</orgName>
								<address>
									<settlement>Poznań</settlement>
									<country key="PL">Poland</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Justyna</forename><surname>Kowalewska</surname></persName>
							<affiliation key="aff2">
								<orgName type="institution">PKO BP</orgName>
								<address>
									<settlement>Warsaw</settlement>
									<country key="PL">Poland</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Agnieszka</forename><surname>Marszałek</surname></persName>
							<affiliation key="aff2">
								<orgName type="institution">PKO BP</orgName>
								<address>
									<settlement>Warsaw</settlement>
									<country key="PL">Poland</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Robert</forename><surname>Wrembel</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">Poznan University of Technology</orgName>
								<address>
									<settlement>Poznań</settlement>
									<country key="PL">Poland</country>
								</address>
							</affiliation>
							<affiliation key="aff1">
								<orgName type="laboratory">Interdisciplinary Centre for Artificial Intelligence and Cybersecurity</orgName>
								<address>
									<settlement>Poznań</settlement>
									<country key="PL">Poland</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Statistical Modeling vs. Machine Learning for Deduplication of Customer Records (industrial paper)</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">1559C08BEFED2AE2B1AC71CE1A0FC1BF</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T17:17+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>data quality</term>
					<term>data deduplication</term>
					<term>statistical modeling</term>
					<term>machine learning</term>
					<term>classification</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Large companies typically face a problem of multiple database records describing the same physical object (a.k.a. duplicates). There are multiple sources of duplicates, e.g., using multiple but not synchronized data repositories, applications that do not check for duplicates before inserting data into a repository, and data errors. This problem is of particular importance while dealing with personal data, e.g., in healthcare, banking, insurance. To handle the problem of duplicates, the state-of-the-art data deduplication pipeline was developed. The pipeline is equipped with multiple complex algorithms. A promising direction in data deduplication is machine learning. In this paper, we report our experience in researching and developing two approaches to deduplication of customer records in a financial institution. These approaches are based on: (1) statistical modeling and (2) machine learning. In particular, this paper summarizes our findings from comparing these two approaches. The reported research was done within a R&amp;D project for the biggest Polish bank -PKO BP.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Large organizations/companies face the problem of duplicated data. This problem is frequent for companies that store customer data. Duplicated and outdated data cause economic losess, increase customer dissatisfaction, and deteriorate the reputation of a company. For these reasons, data integration, cleaning, and deduplication of customer records are one of the core processes in data governance.</p><p>Data deduplication has been extensively researched in multiple research centers worldwide. The research resulted in a base-line data deduplication pipeline, e.g., <ref type="bibr" target="#b0">[1,</ref><ref type="bibr" target="#b1">2,</ref><ref type="bibr" target="#b2">3]</ref>. It has become a standard pipeline for multiple data deduplication projects in various application domains. The pipeline includes four basic tasks, namely:</p><p>• blocking (a.k.a. indexing), which arranges records into groups, such that each group is likely to include duplicates, • block processing (a.k.a. filtering), which eliminates records that do not have to be compared, • entity matching (a.k.a. similarity computation), which computes similarity values between record Each of these tasks is supported by multiple algorithms. Some of them apply very complex statistical models (SM), whereas others are based on standard machine learning (ML), including deep learning (DL). The ML techniques typically apply various classification models to divide records into classes of matches, probably matches, and non-matches, e.g., <ref type="bibr" target="#b3">[4,</ref><ref type="bibr" target="#b4">5,</ref><ref type="bibr" target="#b5">6]</ref>. The DL techniques apply language models for both record blocking and record matching, e.g., <ref type="bibr" target="#b6">[7,</ref><ref type="bibr" target="#b7">8,</ref><ref type="bibr" target="#b8">9,</ref><ref type="bibr" target="#b9">10,</ref><ref type="bibr" target="#b10">11]</ref>.</p><p>Since there are two families of approaches to data deduplication, a question is which family offers more accurate deduplication models. In this paper, we outline our experience and findings on designing a deduplication pipeline for customer data. We designed two versions of the pipeline. The first one uses statistical modeling for discovering duplicates. The second one uses ML techniques. Both pipelines were developed within a R&amp;D project for the biggest Polish Bank PKO BP (https://pkobp.pl). The pipelines were tested on a real data set of over 5 million of customer records. To the best of our knowledge, this is the biggest deduplication experiment comparing SM and ML approaches, reported in the research literature.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Tasks in the SM pipeline</head><p>We built the SM pipeline as the extension of the base-line data deduplication pipeline (BLDDP), to serve the par- </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Figure 1:</head><p>The tasks in the deduplication pipeline based on statistical modeling ticular goals of the project. First, our pipeline explicitly includes all steps that we found to be crucial for the deduplication process (whereas in the BLDDP, some steps are implicit). Second, the last two tasks in our pipeline extend the BLDDP and they allow to further merge groups of similar records into subgraphs. Third, our pipeline accepts partially dirty data, as in practice it is impossible to perfectly clean all data. The SM pipeline applied in the project is shown in Figure <ref type="figure">1</ref>.</p><p>SMT1: In this task, grouping attributes are selected, which allow to co-locate similar records in the same group. In our project, initially 20 candidate grouping attributes were selected by domain experts. The attributes allowed either to identify customers, or they were errorfree and not null, or their values haven't changed over time. The candidate attributes were further verified by means of a method that we developed. For each attribute, the method computed a value representing its fit as a grouping attribute <ref type="bibr" target="#b11">[12]</ref>. Finally, the obtained ranking of attributes was verified by domain experts again. Based on their input, the final set of grouping attributes was selected.</p><p>SMT2: Using grouping attributes obtained from SMT1, records are arranged into groups via sorting. The method we used is very similar to the one described in <ref type="bibr" target="#b12">[13]</ref>, however multiple sortings were used.</p><p>SMT3: The goal of this task is to select attributes that will contribute to assessing a similarity value between records in each pair. Attributes that: (1) represent record identifiers, (2) do not include nulls, (3) include cleaned values, (4) include unified (homogenized) values, (5) do not change values over time are good candidates. Unfortunately, in real scenarios, attributes exposing all these characteristics are often unavailable. In our project, the set of attributes selected for comparing record pairs is based on the aforementioned preferable attribute characteristics and on expert knowledge. The set includes 18 attributes describing individual customers (including personal data and address).</p><p>SMT4: Similarities between corresponding attributes (selected in SMT3) in pairs of records are compared by means of similarity measures. The literature lists well over 30 different similarity measures for text data (e.g., <ref type="bibr" target="#b13">[14,</ref><ref type="bibr" target="#b14">15]</ref>). Unfortunately, there are no rules that would guide a data scientist for selecting the right similarity measure for an attribute having a given characteristic of its values. For this reason, in our project, the selection of the most suitable similarity measures for our problem was based on excessive experimental evaluation <ref type="bibr" target="#b15">[16,</ref><ref type="bibr" target="#b16">17]</ref>.</p><p>SMT5: Measures selected in SMT4 are applied to computing similarities of corresponding attribute values, constituting customer pairs of records. An adequate similarity measure is applied to every pair of attributes. Based on attribute similarities, an overall record similarity value is computed. In a simple scenario, all compared attributes are equally important. In practice, some attributes may contribute more to records similarity than others. For example, a last name is more important (usually more clean and not null) than an email address (frequently null and changing in time). For this reason, the compared attributes must be weighted, resulting in an overall weighted similarity of records. The first challenge in task SMT5 is to define adequate weights for attributes. In practice, these weights are defined by a try and error procedure and by applying expert knowledge. Based on the similarity of a pair of records, the pair is classified either as duplicates (class T), or probably duplicates (class P), or non-duplicates (class N). For this kind of classification, the so-called similarity thresholds have to be defined for each class. These thresholds impact the number of true positives and false negatives. Setting adequate values of the thresholds is another challenge in SMT5. In practice, the thresholds are defined experimentally with the support of domain experts <ref type="bibr" target="#b13">[14]</ref>. In our project, to find attribute weights and similarity thresholds we apply mathematical programming <ref type="bibr" target="#b16">[17]</ref>.</p><p>SMT6: To compare records in a group, a popular technique, called sorted neighborhood is used. It accepts one parameter that is the size of a sliding window in which records are compared. This parameter impacts the computational performance and the number of discovered (potential) duplicates. Defining the size of the sliding window is challenging. A window that is too narrow may prevent from discovering all potential duplicates, whereas a window that is too wide will allow to discover more potential duplicates at a cost of unnecessary time overhead caused by comparing more records -some of them will be false positives. In our project, the window size was selected by means of excessive experiments <ref type="bibr" target="#b12">[13]</ref>.</p><p>SMT7: Since similar records may form groups larger than pairs, in order to find such groups, all similar pairs of records have to be combined. To this end, we cre- SMT8: Similar records in the record similarity graph can be extracted by cutting the graph into sub-graphs. In our project, we applied clustering of records by using similarity values of pairs of records computed in SMT5. For this purpose, we modified the Highly Connected Subgraphs (HCS) clustering algorithm <ref type="bibr" target="#b17">[18]</ref>. HCS recursively applies the minimum cut to the record similarity graph (separately for each connected component) until the obtained sub-graphs are highly connected, i.e. the size of the minimum cut is larger than the number of nodes divided by two. The resulting sub-graphs form the groups of similar records. The modifications include: using a minimum weighted cut instead of a minimum cut, searching for additional edges in sufficiently small subgraphs and assigning resulting singletons to the most similar groups after the clustering ends. The performance of the modified HCS is close to linear w.r.t. the number of pairs, since larger connected components in the similarity graph are rare.</p><p>Notice that the pipeline we proposed, is mapped to the BLDDP as follows: SMT1 realizes block building in the BLDDP; SMT2 realizes block processing; SMT3, SMT4, and SMT5 realize entity matching; SMT6, SMT7, and SMT8 realize entity clustering. A more detailed description of our pipeline is available in <ref type="bibr" target="#b16">[17]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Tasks in the ML pipeline</head><p>The machine learning pipeline applied in the project is shown in Figure <ref type="figure" target="#fig_1">2</ref>.</p><p>MLT1: in this task, 1000 pairs of records were manually assigned labels by domain experts. A label indicated one of the three classes, namely T, P, or N. All the labels were roughly equally represented. Since the Bank database included over 20 millions of customer records, it was not possible to label manually a sufficiently large sample of record pairs. For this reason, in MLT2 the set of rules was built manually with the support of domain experts. It is important to stress that high quality of these rules is crucial for subsequent tasks. The rules were next applied in task MLT3 to automatically label a set of 50 000 pairs of customer records.</p><p>It was assumed that the effect of applying the rules was to determine decisions consistent with those of the domain experts. Even for a small set of attributes being compared in pairs (18 attributes in our case), attempts to create simple rules of type if X then Y else Z turned out to be challenging. It was due to the very different scenarios present in the data, in particular, the lack of 100% similarity between the compared attribute values.</p><p>It turned out that rules that only confirmed whether a given pair could be a representative of a given class T, P, or N were much easier to interpret. This led to a set of 𝑚 rules under the following assumptions:</p><p>• each rule can return only one answer: T, P or N;</p><p>• a rule for a given pair of customer records may return no answer; • for a given pair, successive rules are applied; an answer can be returned by a set of 0 to 89 rules; • rules have a certain priority -the strongest are rules that assign class N, followed by rules that assign class T, and the weakest rules are those assigning P.</p><p>The above assumptions cause that the final decision made by the rules is determined as described in Tab. 1. It shows the number of N, P, and T responses generated by a set of rules. Symbol * denotes any number of rules returning a given answer. In general, the following four cases are possible. Case 1: when at least one rule returns N, then the final decision is N. This is due to the highest priority of rules returning N. Case 2: decision P is taken when no rule returns N and no rule returns T and at least one rule returns P. Case 3: decision T is taken when no rule returns N and at least one rule returns T. Case 4: no rule returns an answer. In such a case, given the bank's conservative policy, the final decision is N, since a possible false negative is much less costly than a false positive. Despite tremendous effort, designing rules that matched exactly the experts decisions proved to be im-possible in a reasonable amount of time. The confusion matrix for the set of 89 rules created is shown in Tab. 2. In addition to the T, P, and N classes, it is easy to see that for a total of 70 pairs, the rules made no decision (column None). This means that for 1 000 pairs, the coverage of pairs by rules is equal to 93%. It is noteworthy that the assessments made by the rules were more conservative than those made by the experts, i.e., in the case of confusion, they assigned class N to pairs considered to be T and class P to pairs considered to be T. It is notable that never a pair of class N was considered either P or T. In MLT3, from the test set of over 5 million of customer records, 2.5 million of pairs were created and then they were labeled by means of the rules. The set of pairs to be labeled was not random, and a good portion of it contained common as well as problematic examples. Therefore, it can be considered that it was somehow biased and unrepresentative for the coverage assessment measure. It turned out to be much more interesting to assess coverage by experimenting with the set of pairs obtained by applying step SMT2 (from the SM approach). For the set of 2.5 million of pairs analyzed, in only 23 352 cases the rules were not able to make a final decision. This means the rule coverage of 99.987%. It should be noted here that of the 89 rules, the most popular rules (i.e., those that produced a result for the largest number of pairs) assigned class N. The most popular rule produced N for more than 83% of pairs. When considering the entire population of pairs (i.e., any pairs from the set), the result would be even higher, as the chance to find duplicates decreases significantly.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 1 Determination of the final rule decision</head><p>In, MLT4, from the set of pairs built in MLT3, a training and a testing data sets were created by stratified sampling. Having created these two subsets, we have added new features inspired by the conditions of the expert rules (task MLT5).</p><p>In MLT6, we run several preparatory experiments to get the "feel" of the data. In one of them, we used PCA method to obtain two principal components and rendered each example in the dataset of 1000 pairs manually labeled by domain experts on a two-dimensional scatter plot, with classes T, P, and N marked in colors. The result was that classes T and N were separated quite well, while P was scattered all over the plot. This observation meant that class P would be problematic.</p><p>Having run the preparatory experiments, four different classification models were created based on the training set. The models included: decision tree, random forest, SVM, and feed forward neural network. These models were built using Python package sklearn. Hyperparameters of the models were adjusted manually or (if needed) with the help of the optuna package, based on the dataset labeled by domain experts. We also attempted to use an auto-ML approach by means of the tpot library, but it was not able to produce any model that would be more accurate than the models mentioned above, within the given time frame. Finally, in MLT7, the quality of all the produced models was tested based on the testing set obtained from the automatically labeled data as well as based on the set of 1000 pairs manually labeled by domain experts. The model quality was measured by means of precision, recall, measure F1, and accuracy.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Development</head><p>Pipeline implementation environment: Both the SM and ML pipelines were implemented in a typical data science environment, which included: (1) a relational database management system to store customer data, (2) csv files to store temporary data produced by the tasks in the pipelines, (3) spreadsheet files to store groups of duplicated records, and (4) Python programs and standard packages to implement tasks in both pipelines.</p><p>Data size: In the reported project, in the development and testing phases we used 2.5 million of pairs of customer records. In the production system, the pipelines run on the Bank customer database, which includes over 20 million of records.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Results</head><p>Tab. 3 presents precision, recall, measure F1, and accuracy achieved by the ML models on the dataset of 1000 pairs manually labeled by the domain experts. We also add results achieved by the Statistical Modeling approach, for comparison. The numbers in brackets represent the ranking of the values (best-1, worst-5). As it can be observed, the best performance on this dataset was obtained by the Statistical Model. The worst result was obtained by the decision tree. In the rest of the cases the ranking is not so obvious, as not all measures point to the same rank. However, since F1 is an aggregation of precision and recall, and the ranking of F1 is consistent with the accuracy, we can assume that the second best is random forest, followed by SVM and feed-forward neural network (FFNN).</p><p>The best results obtained via the Statistical Model stem most probably from the fact that all parameters of the There are a few interesting observations to be made here. First, notice that the orders of precision, F1, and accuracy are consistent, while recall is different. However, similarly as before we can ignore precision and recall in favor of their aggregation, i.e., F1, to obtain the ranking of the models. In this approach, the best model is random forest, followed by decision tree, FFNN, SVM, and the Statistical Model. Thus, the obtained order is completely different. We explain the results as follows.</p><p>Since training and testing datasets had labels generated by the same set of rules, the classifiers that were trained on the training dataset tried to generalize the set of these rules. Thus, their performance on the testing dataset is quite good. The Statistical Model was not built based on the set of rules at all, and thus, its performance on the testing dataset is worse. It is also interesting that treebased classifiers, i.e., decision tree and random forest perform better than mathematical models such as FFNN and SVM, which come down to division of attribute space by hyperplanes. We suspect that this stems from the fact, that trees are just representations of sets of rules. Thus, the tree model structure fits better the underlying source of the labels than the mathematical models. One can also observe that a single test in a node of a tree is also a hyperplane dividing the attribute space, but such planes are orthogonal to one of the axis, while in mathematical models, the hyperplanes are aligned arbitrarily. Larger degree of freedom makes it harder for the mathematical models to find the original model that produced the labels, which leads to worse results.</p><p>Regardless of the testing dataset, the best results were achieved by the random forest model. Its confusion matrix for the dataset labeled by the domain experts is given in Tab. 5. One of the most important observations here is the fact that the number of erroneous classifications involving class P is an order of magnitude greater than the number of erroneous classifications involving classes T and N. This observation confirms the result mentioned in Section 3. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Summary</head><p>In this paper, we reported our experience from a R&amp;D project on deduplicating customers data. The project was realized for the biggest Polish Bank PKO BP. While presented solutions were designed explicitly for the Bank dataset, the findings are general and the approach can be fitted to other, similar problems.</p><p>In the project, we designed two deduplication pipelines: (1) based on statistical modeling and (2) based on machine learning. The SM pipeline extends the standard deduplication pipeline from the literature to the par-ticular characteristics of data being deduplicated and to the project requirements. The ML pipeline was designed from scratch within the project. Both pipelines were implemented, tested on a data set of 2.5 million of pairs of customer records, and verified by domain experts. The obtained results were accepted by the Bank. The project ended with the decision to run the SM pipeline in the production system on the Bank database, which includes over 20 million of customer records. This pipeline has already been deployed in the Bank.</p><p>It must be stressed that the real settings of the project differ from the ones assumed in the research literature (the ideal ones) among others w.r.t.: (1) the quality of data being deduplicated, (2) the sizes of deduplicated data, (3) the availability of tagged data for ML algorithms, (4) available development environments, (5) the performance of a deduplication pipeline (time and quality). The real settings, which are far from the ideal ones, made the project very challenging.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head></head><label></label><figDesc>Published in the Proceedings of the Workshops of the EDBT/ICDT 2024 Joint Conference (March 25-28, 2024), Paestum, Italy robert.wrembel@cs.put.poznan.pl (R. Wrembel) 0000-0001-9486-929X (W. Andrzejewski); 0000-0002-6426-3809 (B. Bębel); 0000-0003-4914-9394 (P. Boiński); 0000-0001-6037-5718 (R. Wrembel) pairs, and • entity clustering, which creates larger clusters of similar records.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: The tasks in the deduplication pipeline based on machine learning</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2</head><label>2</label><figDesc>Confusion matrix for the final set of rules</figDesc><table><row><cell></cell><cell></cell><cell></cell><cell cols="2">Rules decision</cell></row><row><cell></cell><cell></cell><cell>N</cell><cell>P</cell><cell>T</cell><cell>None</cell></row><row><cell>Expert decision</cell><cell>N P T</cell><cell cols="3">310 0 8 373 0 0 0 7 232 2 38 30</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 3</head><label>3</label><figDesc>Comparison of models on the expert labeled dataset ML models on the testing dataset, obtained by means of 89 rules. It is crucial to underline that the results presented in the table compare how</figDesc><table><row><cell></cell><cell cols="2">Decision Tree Random forest</cell><cell>FFNN</cell><cell>SVM</cell><cell>Statistical model</cell></row><row><cell>Precision</cell><cell>[5] 0.7620</cell><cell>[3] 0.8369</cell><cell cols="2">[4] 0.8176 [2] 0.8422</cell><cell>[1] 0.8607</cell></row><row><cell>Recall</cell><cell>[5] 0.7801</cell><cell>[2] 0.8498</cell><cell cols="2">[3] 0.8391 [4] 0.8208</cell><cell>[1] 0.8799</cell></row><row><cell>F1</cell><cell>[5] 0.7631</cell><cell>[2] 0.8407</cell><cell cols="2">[4] 0.8241 [3] 0.8296</cell><cell>[1] 0.8673</cell></row><row><cell>Accuracy</cell><cell>[5] 0.7612</cell><cell>[2] 0.8383</cell><cell cols="2">[4] 0.8201 [3] 0.8276</cell><cell>[1] 0.8640</cell></row><row><cell>Table 4</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell cols="2">Comparison of models on the testing dataset</cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell></cell><cell cols="2">Decision Tree Random Forest</cell><cell>FFNN</cell><cell>SVM</cell><cell>Statistical model</cell></row><row><cell>Precision</cell><cell>[2] 0.9577</cell><cell>[1] 0.9629</cell><cell>[3] 0.9444</cell><cell>[4] 0.8898</cell><cell>[5] 0.8051</cell></row><row><cell>Recall</cell><cell>[3] 0.9399</cell><cell>[2] 0.9551</cell><cell cols="2">[4] 0.9368 [1] 0.9619</cell><cell>[5] 0.8182</cell></row><row><cell>F1</cell><cell>[2] 0.9484</cell><cell>[1] 0.9590</cell><cell>[3] 0.9405</cell><cell>[4] 0.9201</cell><cell>[5] 0.8065</cell></row><row><cell>Accuracy</cell><cell>[2] 0.9958</cell><cell>[1] 0.9967</cell><cell>[3] 0.9953</cell><cell>[4] 0.9926</cell><cell>[5] 0.9859</cell></row><row><cell cols="3">model (mainly attribute weights and similarity thresh-</cell><cell></cell><cell></cell><cell></cell></row><row><cell cols="3">olds) were optimized directly based on the expert labeled</cell><cell></cell><cell></cell><cell></cell></row><row><cell cols="3">dataset also used for testing. The same dataset was only</cell><cell></cell><cell></cell><cell></cell></row><row><cell cols="3">used for optimizing hyperparameters for other models</cell><cell></cell><cell></cell><cell></cell></row><row><cell cols="3">(random forest, SVM, FFNN). Decision trees were ad-</cell><cell></cell><cell></cell><cell></cell></row><row><cell>justed manually.</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell cols="3">Tab. 4 presents precision, recall, F1, and accuracy</cell><cell></cell><cell></cell><cell></cell></row><row><cell>achieved by the</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>close are the results of the ML model to the results obtained by applying the rules, and not to the ground truth (which is not available).</head><label></label><figDesc>The structure of the table is the same one as Tab. 3.</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Table 5</head><label>5</label><figDesc>Confusion matrix for random forest</figDesc><table><row><cell></cell><cell></cell><cell cols="2">Random forest decision</cell></row><row><cell></cell><cell></cell><cell>N</cell><cell>P</cell><cell>T</cell></row><row><cell>Expert decision</cell><cell>N P T</cell><cell cols="2">318 24 24 319 68 6 1 28 212</cell></row></table></figure>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgements.</head><p>The project is supported by the grant from the National Center for Research and Development no. POIR.01.01.01-00-0287/19.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Duplicate record detection: A survey</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">K</forename><surname>Elmagarmid</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">G</forename><surname>Ipeirotis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><forename type="middle">S</forename><surname>Verykios</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Knowledge and Data Engineering</title>
		<imprint>
			<biblScope unit="volume">19</biblScope>
			<biblScope unit="page" from="1" to="16" />
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Frameworks for entity matching: A comparison</title>
		<author>
			<persName><forename type="first">H</forename><surname>Köpcke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Rahm</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Data &amp; Knowledge Engineering</title>
		<imprint>
			<biblScope unit="volume">69</biblScope>
			<biblScope unit="page" from="197" to="210" />
			<date type="published" when="2010">2010</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Domainand structure-agnostic end-to-end entity resolution with jedai</title>
		<author>
			<persName><forename type="first">G</forename><surname>Papadakis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Tsekouras</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Thanos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Giannakopoulos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Palpanas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Koubarakis</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">SIGMOD Record</title>
		<imprint>
			<biblScope unit="volume">48</biblScope>
			<biblScope unit="page" from="30" to="36" />
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Heterogeneous committee-based active learning for entity resolution (healer)</title>
		<author>
			<persName><forename type="first">X</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Broneske</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">C</forename><surname>Durand</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Zoun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Saake</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">European Conf. on Advances in Databases and Information Systems ADBIS</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2019">2019</date>
			<biblScope unit="volume">11695</biblScope>
			<biblScope unit="page" from="69" to="85" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Interactive deduplication using active learning</title>
		<author>
			<persName><forename type="first">S</forename><surname>Sarawagi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Bhamidipaty</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">ACM SIGKDD Int. Conf. on Knowledge Discovery and Data Mining</title>
				<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2002">2002</date>
			<biblScope unit="page" from="269" to="278" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">A multiclass classification approach for incremental entity resolution on short textual data</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">A</forename><surname>Silva</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">A</forename><surname>Pereira</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Int. Journal of Business Intelligence and Data Mining</title>
		<imprint>
			<biblScope unit="volume">18</biblScope>
			<biblScope unit="page" from="218" to="245" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Entity matching with transformer architectures -A step forward in data integration</title>
		<author>
			<persName><forename type="first">U</forename><surname>Brunner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Stockinger</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Int. Conf. on Extending Database Technology (EDBT), OpenProceedings.org</title>
				<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="463" to="473" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Deep indexed active learning for matching heterogeneous entity representations</title>
		<author>
			<persName><forename type="first">A</forename><surname>Jain</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Sarawagi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Sen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">VLDB Endowment</title>
		<imprint>
			<biblScope unit="volume">15</biblScope>
			<biblScope unit="page" from="31" to="45" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Deep learning for entity matching: A design space exploration</title>
		<author>
			<persName><forename type="first">S</forename><surname>Mudgal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Rekatsinas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Doan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Park</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Krishnan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Deep</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Arcaute</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Raghavendra</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">SIGMOD Int. Conf. on Management of Data, ACM</title>
				<imprint>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="19" to="34" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Deep learning for blocking in entity matching: A design space exploration</title>
		<author>
			<persName><forename type="first">S</forename><surname>Thirumuruganathan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Tang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Ouzzani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Govind</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Paulsen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Fung</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Doan</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Proc. VLDB Endowment</title>
		<imprint>
			<biblScope unit="volume">14</biblScope>
			<biblScope unit="page" from="2459" to="2472" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Pre-trained embeddings for entity resolution: An experimental analysis</title>
		<author>
			<persName><forename type="first">A</forename><surname>Zeakis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Papadakis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Skoutas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Koubarakis</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Proc. VLDB Endowment</title>
		<imprint>
			<biblScope unit="volume">16</biblScope>
			<biblScope unit="page" from="2225" to="2238" />
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">On customer data deduplication: Lessons learned from a r&amp;d project in the financial sector</title>
		<author>
			<persName><forename type="first">P</forename><surname>Boiński</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Sienkiewicz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Bębel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Wrembel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Gałęzowski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Graniszewski</surname></persName>
		</author>
		<ptr target=".org" />
	</analytic>
	<monogr>
		<title level="m">Workshops of the EDBT/ICDT 2022 Joint Conference</title>
		<title level="s">CEUR Workshop Proceedings</title>
		<imprint>
			<publisher>CEUR-WS</publisher>
			<date type="published" when="2022">2022</date>
			<biblScope unit="volume">3135</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">On tuning the sorted neighborhood method for record comparisons in a data deduplication pipeline -industrial experience report</title>
		<author>
			<persName><forename type="first">P</forename><surname>Boinski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Andrzejewski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Bebel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Wrembel</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Int. Conf. Database and Expert Systems Applications (DEXA)</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2023">2023</date>
			<biblScope unit="volume">14146</biblScope>
			<biblScope unit="page" from="164" to="178" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Data Matching -Concepts and Techniques for Record Linkage, Entity Resolution, and Duplicate Detection</title>
		<author>
			<persName><forename type="first">P</forename><surname>Christen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Data-Centric Systems and Applications</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2012">2012</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<monogr>
		<title level="m" type="main">Similarity measures</title>
		<author>
			<persName><forename type="first">F</forename><surname>Naumann</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2013">2013</date>
			<publisher>Hasso Plattner Institut</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Text similarity measures in a data deduplication pipeline for customers records</title>
		<author>
			<persName><forename type="first">W</forename><surname>Andrzejewski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Bebel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Boinski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Sienkiewicz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Wrembel</surname></persName>
		</author>
		<ptr target=".org" />
	</analytic>
	<monogr>
		<title level="m">Int. Workshop on Design, Optimization, Languages and Analytical Processing of Big Data (DOLAP)</title>
		<title level="s">CEUR Workshop Proceedings</title>
		<imprint>
			<publisher>CEUR-WS</publisher>
			<date type="published" when="2023">2023</date>
			<biblScope unit="volume">3369</biblScope>
			<biblScope unit="page" from="33" to="42" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">On tuning parameters guiding similarity computations in a data deduplication pipeline for customers records: Experience from a r&amp;d project</title>
		<author>
			<persName><forename type="first">W</forename><surname>Andrzejewski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Bębel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Boiński</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Wrembel</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Information Systems</title>
		<imprint>
			<biblScope unit="volume">121</biblScope>
			<biblScope unit="page">102323</biblScope>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Partitioning biological networks into highly connected clusters with maximum edge coverage</title>
		<author>
			<persName><forename type="first">F</forename><surname>Hüffner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Komusiewicz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Liebtrau</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Niedermeier</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE/ACM Transactions on Computational Biology and Bioinformatics</title>
		<imprint>
			<biblScope unit="volume">11</biblScope>
			<biblScope unit="page" from="455" to="467" />
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
