<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>SEBD</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Attribute Ambiguity Discovery: A Deep Learning Approach via Unsupervised Learning</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>(Discussion Paper)</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Enzo Veltri</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Gilbert Badaro</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Mohammed Saeed</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Paolo Papotti</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>EURECOM</institution>
          ,
          <addr-line>Biot</addr-line>
          ,
          <country country="FR">France</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Università degli Studi della Basilicata (UNIBAS)</institution>
          ,
          <addr-line>Potenza</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2023</year>
      </pub-date>
      <volume>31</volume>
      <fpage>02</fpage>
      <lpage>05</lpage>
      <abstract>
        <p>Several applications, such as text-to-SQL and computational fact checking, exploit the relationship between relational data and natural language text. However, state of the art solutions simply fail in managing “data-ambiguity", i.e., the case when there are multiple interpretations of the relationship between text and data. Given the ambiguity in language, text can be mapped to diferent subsets of data, but existing training corpora only have examples in which every sentence/question is annotated precisely w.r.t. the relation. This unrealistic assumption leaves the target applications unable to handle ambiguous cases. To tackle this problem, we present a deep learning method that identifies every pair of data ambiguous attributes and a label that describes both columns. Such metadata can then be used to generate examples with data ambiguities for any input table.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Unsupervised Text Generation</kwd>
        <kwd>Data to Text Generation</kwd>
        <kwd>Data Ambiguity</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Ambiguity in natural language comes in many forms [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. Factual sentences or questions can
be “data-ambiguous” w.r.t the data available in a table. A simple example is the question “Is
Curry the best shooter in NBA history?”. Based on the NBA oficial data, the answer changes
depending on the interpretation of shooting in terms of table attributes [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. Indeed, in the context
of querying relational databases using natural language (text-to-SQL), “ambiguity of natural
language queries is one of the most dificult challenges” [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. The problem of data ambiguities in
text is also relevant for many natural language processing (NLP) applications that use relational
data. These span from computational fact checking, i.e., verify if a given claim holds w.r.t. a
table [
        <xref ref-type="bibr" rid="ref4 ref5">4, 5</xref>
        ], to question answering in general [
        <xref ref-type="bibr" rid="ref6 ref7">6, 7</xref>
        ], and document classification [
        <xref ref-type="bibr" rid="ref8 ref9">8, 9</xref>
        ].
      </p>
      <p>Consider a fact checking application that verifies a textual claim, such as “Carter LA has
higher shooting than Smith SF”, against a relational  as in Table 1. Even as humans, it is hard
to state if the sentence is true or false w.r.t. the data in . The challenge is due to the two
1
2</p>
      <p>Team</p>
      <p>LA
SF
3FG%
47
50
fouls
4
4
apps
5
7
diferent meanings that can be matched to shooting: the claim can refer to attribute Field Goal
(FG%) or to 3-point Field Goal (3FG%). The same challenge applies with a SQL query expressed
in natural language such as “Did Carter LA has higher shooting than Smith SF?”. We refer to
this issue as data ambiguity, i.e., the existence of more than one interpretation of a text w.r.t. the
data for a human reader.</p>
      <p>While existing corpora of examples come from extensive and expensive manual eforts, they
do not contain examples with ambiguous text. Existing applications fail in these scenarios: the
two examples above would lead to a single interpretation, which is incorrect in 50% of the cases.</p>
      <p>
        In this work, we focus on the discovery of ambiguities over a relational schema, i.e., on attribute
ambiguities. Ambiguities over data, and the combination of attributes and data ambiguities are
explored in the full paper [
        <xref ref-type="bibr" rid="ref10 ref11">10, 11</xref>
        ].
      </p>
      <p>
        Given a relation, the key idea is to identify the set of attributes involved in ambiguity over
the data. To handle these challenges we introduce deep learning models that predict if two
attributes are ambiguous and if so, it also predicts a text or “label” that makes them ambiguous.
Such information can be used to generate a large corpus of ambiguous claims [
        <xref ref-type="bibr" rid="ref10 ref12">10, 12</xref>
        ].
      </p>
      <p>Ambiguity Discovery is presented in Section 2 and the Experimental Evaluation in Section 3.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Ambiguity Discovery</title>
      <p>From our analysis of an online fact-checking application’s log, we estimate that 40% of the
user-submitted sentences present attribute ambiguity.</p>
      <p>We describe the overview of the Attribute Discovery (Figure 1) to obtain the ambiguity
metadata for a given relational table.</p>
      <p>
        1) In the training phase, a transformer pre-trained model is fine-tuned with examples of pairs
of ambiguous attributes1. As training data does not exist for this task, we use six unsupervised,
noisy annotator functions. Given a pair of attributes for a table, the annotators produce a
label for them if they are ambiguous. For example, given the attributes ‘length’ and ‘width’, an
annotator function should produce labels such as ‘dimension’ or ‘magnitude’ that yield some
ambiguity w.r.t. the input attributes. We run these noisy annotators on a large collection of
tables, such as the WebTables corpus [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ]. The unsupervised functions are detailed in Section 2.1.
      </p>
      <p>
        2) The noisy output of the annotator functions is used as training examples for a fine-tuning
task on an encoder-decoder language model [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ]. This text-to-text architecture achieves
stateof-the-art results on several NLP tasks. We introduce the novel task of predicting (1) if two
attributes are ambiguous and (2) the label for such ambiguity. We design two variants that take
into consideration diferent levels of access to the tables. One considers as input the schema
only, while the second assumes that a sample of the data can be fed to the model. We discuss
this component in Section 2.2.
      </p>
      <p>3) Once the model has been trained, we use it at test time with any unseen input table . We
test all the possible pairs of non-key attributes in  and, if a label is predicted, we consider the
input attribute pair in question as ambiguous.</p>
      <p>
        4) Such ambiguous metadata can be used in a generic Text-Generator for the target
application [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ].
      </p>
      <sec id="sec-2-1">
        <title>2.1. Annotator Functions</title>
        <p>Our goal is to create training data for a model that, given two attributes alongside table schema
(or table schema with a data sample), either produces a label if both attributes are ambiguous or
abstains otherwise. Unfortunately, we cannot count on a large amount of manually annotated
examples. To obtain such data in an unsupervised fashion, we start the process with a set of
basic annotator functions that exploit external resources and heuristics to find candidates for
ambiguity.</p>
        <p>
          The first set of our weak supervision annotators [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ] are designed in a two-step approach.
First, an alias function automatically finds possible aliases for a given word. For example, a
set of possible aliases for the word length are {measurement, measure, range}. Then the alias
for two candidate attributes are compared and eventually selected. We design alias functions
that use diferent external datasets. Four alias functions use the ConceptNet graph [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ]. Given
the input word (attribute label), we look for its relationships in the graph that are alternative
1We focus on pair of attributes as we found it efective for example generation in our target application. The
method can be extended with a post-processing step that analyzes pairs to compute larger sets, e.g., 1pt%, 2pt%,
3pt%.
representations of the input: )  synonyms; )   related words; )  words it derived
from; and )  subtypes. For a second external resource, we use the Wikipedia API to search
top page titles as other possible aliases with the wiki function. Table 2 shows examples of the
generated aliases.
        </p>
        <p>Once the aliases have been collected, given a pair of attributes represented by their names,
we say that two attributes are ambiguous if the intersection of their aliases values is non-empty.
The ambiguous labels are the intersection. If the intersection is empty, then the two pairs of
attributes are not ambiguous. An annotation function makes use of the intersection for every
external resource. If the attribute names are not meaningful, the annotator functions output
empty results, e.g., attribute with name “A12” has empty results for all annotator functions.</p>
        <p>As a sixth function, we find the Longest Common Substring ( LCS) between two attribute
names. Since the result may contain sequences of characters without meaning, we filter the
generated word with a dictionary.</p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Fine-Tuning the Language Model</title>
        <p>
          Pre-trained transformer based language models (LMs) such as BERT [17] or T5 [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ] have
shown to perform well in diferent NLP tasks such as question answering and text classification.
Typically, these models are pre-trained on large text corpora such as articles from Wikipedia,
news articles, or Common Crawl. The model is pre-trained in an unsupervised manner, for
example by predicting a missing token or the next-sentence in a paragraph. Unlike conventional
word embedding techniques, pre-trained transformer-based LMs learn better the semantics
of words and provide diferent representations for the same word when utilized in diferent
contexts. LMs can then be further fine-tuned for supervised, specific tasks. Tuning with task
specific data is one of the advantages of pre-trained LMs, thanks to the advances in transfer
learning. Fine-tuning is performed in a supervised manner by providing the LM the labeled
input. Fine-tuning allows to generate tasks with new prompts while efectively exploiting
transfer-learning for natural language understanding.
        </p>
        <p>We define two variants of the same fine-tuning task with the objective of identifying a pair of
ambiguous attributes and their common label, in case such label exists. We consider a sequence
generation task where the goal is to learn a function  :  → , where  is a sequence of text
containing the pair of attributes and  is a sequence of text corresponding to the label (where
none means no ambiguity). The training data comes from the annotator functions discussed
above. The diference between the two tasks lies in the prompt provided to the LM.</p>
        <p>In the schema-task, function  takes a table schema  and a pair of its attribute with labels 
as input, the goal is to predict a text label  in case a similarity exists between the two attributes
or None. Hence  = (, );  ∈ (,  ).</p>
        <p>In the data-task, the table header and randomly selected rows are concatenated with the
attributes along with the corresponding label. In this case,  ′ assumes as input a sample of the
table data cells , in addition to  and . We denote the schema and the data together as
. Hence, ′ = (, );  ∈ (,  ). For this task, we discuss alternative representations
of the table cells, namely a row serialization and a column serialization. Our decision of the two
alternatives is based on the typical serialization in the literature to create neural representations
for database tables [18]. While other contextual features are sometimes encoded, we keep our</p>
        <sec id="sec-2-2-1">
          <title>Schema-task prompt: concatenate , ,</title>
          <p>Player | Team | FG% | 3FG% | fouls | apps, attr1: FG% attr2: 3FG%, [y: shooting]</p>
        </sec>
        <sec id="sec-2-2-2">
          <title>Data-task prompt (Row): concatenate , ,</title>
          <p>Player | Team | FG% | 3FG% | fouls | apps || Carter | LA | 56 |47 | 4 | 5 || Smith | SF | ... | 7 || Carter|
..., attr1: FG% attr2: 3FG%, [y: shooting]</p>
        </sec>
        <sec id="sec-2-2-3">
          <title>Data-task prompt (Column): concatenate , ,</title>
          <p>Player | Carter | Smith | Carter || Team | LA | SF | SF || FG%| 56 | 55 | 60 || 3FG% | ... || fouls | 4| ...,
attr1: FG% attr2: 3FG%, [y: field goal]
model simple and generic.</p>
          <p>Figure 2 shows a sample of the prompt for the fine tuning of the model based on the basket
data in Table 1. The special tokens are used to help the model distinguish the start and the end
of a cell, of a row, and of a column depending on the configuration.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Experiments</title>
      <p>We now analyze in detail the solution proposed to identify attributes with ambiguities and their
corresponding labels. In the following, we refer to our methods described in Section 2.2 as
Schema and Data.</p>
      <p>
        Datasets. For the training, we take a sample of 500k tables from the Web Tables corpus [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ].
We use relational tables with a header as first row and horizontal orientation. For the test, we
use a crowd-annotated corpus of 13 tables from the popular UCI Dataset. We asked the crowd
to annotate ambiguities for two tasks where (i) the schema or (ii) the schema and the data were
provided. We presented pair of attributes with possible ambiguous labels generated from us.
We asked to mark the ambiguous pairs, moreover, for all the ambiguous pairs, we asked also to
report more ambiguous words if possible.
      </p>
      <p>Metrics. We do not report accuracy, defined as the fraction of correct predictions. As the
number of pairs of attributes defined as ambiguous is much smaller than the number of attribute
combinations, the label distribution is skewed (output is dominated by true negative). We
therefore report precision (P), recall (R), and their combination in the f-measure (F1) for the
prediction of the models, where a prediction for a label is a true positive if such label is in the
ground truth.</p>
      <p>Baselines. The annotator functions introduced in Section 2.1 perform very poorly when
naively applied to the test dataset. We therefore introduce two baseline methods. The first, is an
unsupervised labeling heuristic function, uLabel, that uses both ConceptNet and Wikipedia to
ifnd possible common words for the attributes with intersection. If the result is still empty, then
it returns the output of the LCS function. The second is a supervised labeling solution (sLabel),
that also makes use of the pre-trained LM. This fine tuning task takes a single attribute as input
and produces a list of possible labels. It starts with the examples from the same annotators,
i.e., synonyms of the attribute, ConceptNet labels, Wikipedia page titles, and the least common
sub-string between the attribute and every other attribute in the table. The resulting list of
possible labels is then used as training data. For testing, each attribute is submitted to the model
and the pairs of attribute with non-empty intersection of their outputs are added to the results.</p>
      <p>For the evaluation, we start by comparing the diferent methods. We then show how diferent
parameters have an impact on our solution. All results are reported for the binary task of
stating if two attributes are ambiguous (Ambiguity) and the task of predicting the correct label
(Labeling).</p>
      <p>Results w.r.t. the baselines. Table 3 shows the results for the four methods on the test corpus.
We observe that the unsupervised baselines obtains good precision in both tasks, but very low
recall. The other methods perform all well in terms of detecting ambiguous pairs (Ambiguity),
with sLabel close to our methods in terms of f-measure thanks to very high precision and
recall. However, in the task of predicting the label, both our models clearly outperform both
baselines with an f-measure of 82%, while sLabel achieves only 53%. Interestingly, both models
not using data (sLabel and Schema) achieve high precision, while the model that uses schema
and data (Data) achieves much higher recall. This is because the comparison of data values
may lead to ambiguities that are not captured by looking at attribute label only. However, this
may be misleading in some cases, such as those with numerical values with similar domains
but diferent value distributions. For example, for attributes FG_PCT and FG3M, the human
annotators agree on ‘FG’ as ambiguous label, but Data returns none as they have diferent value
distributions. For Data, we found experimentally that the best quality results are achieved
with the maximum number of rows to consider in the  equals to five and that the row
representation outperforms the column representation.</p>
      <p>Results w.r.t. the number of training steps. Figure 3 (on the left) shows that both methods
improve the quality of the results with an increasing number of training steps. Results nearly
converge after 3k steps. Training Schema and Data models requires 1.5 hours.</p>
      <p>Results w.r.t. the number of T5 parameters (model size). Figure 3 (on the right) reports
the impact of the number of training parameter on the quality of the results. Increasing the
size of the model parameters (sizes: Small &lt; Base &lt; Large &lt; 3B) increases the quality of the
predictions in the final model. LMs with a small number of parameters after 2000 training steps
start to converge to low-quality results. The number of parameters also influences the inference
time. The biggest model (3B) takes on average two seconds per prediction on our machine.</p>
      <p>Schema Amb.</p>
      <p>Schema Label</p>
      <p>Data Amb.</p>
      <p>Data Label
0.8</p>
      <sec id="sec-3-1">
        <title>3.1. End-To-End User Evaluation</title>
        <p>
          We conduct a second user study to evaluate the text generated by an end-to-end system [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ]
over 11 datasets. The system is configured to use our ambiguity discovery module. We generate
at least four ambiguous sentences and two not ambiguous ones for each table. We ask eleven
participants to manually annotate the generated text in two ways. First, ambiguity detection:
state if the text is ambiguous w.r.t. the associated schema and a sample of data. Second, attribute
ambiguity detection: if the text is ambiguous, mark the ambiguous attributes. We instruct the
participants with data ambiguity definitions and examples. We give each participant three
datasets to annotate, so that every dataset has three annotations. Then, for every annotator
and dataset, we measure P, R and F1 for both ambiguity and attribute ambiguity detection. For
ambiguity, we count a match if the annotation agrees with the binary label for the text in the
ground truth. For the attribute ambiguity, we count a match if at least one of the annotated
attributes is in the ground truth of the text. Results per dataset (averaged over three participants)
Abalone 1.000
Adults 0.923
Basket Acronyms 0.750
Basket 1.000
Heart Diseases 0.875
Iris 1.000
Superstore 0.950
Wine Quality 0.950
Laptop 0.923
Mushroom 1.000
Soccer 0.762
are shown in Table 4. We observe an average F1 (over all datasets) of 84% and 82.5% for ambiguity
and attribute ambiguity detection, respectively. Such results show that (1) humans recognize
text with and without data ambiguity, (2) they also recognize the right attributes when there is
ambiguity. As expected, the second action has lower F1 as it is an harder task.
[17] J. Devlin, M. Chang, K. Lee, K. Toutanova, BERT: pre-training of deep bidirectional
transformers for language understanding, in: NAACL-HLT, ACL, 2019, pp. 4171–4186.
        </p>
        <p>URL: https://doi.org/10.18653/v1/n19-1423. doi:10.18653/v1/n19-1423.
[18] G. Badaro, M. Saeed, P. Papotti, Transformers for Tabular Data Representation: A Survey
of Models and Applications, TACL (2023). URL: https://hal.science/hal-03877085.</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1] [n.d.], Notes on ambiguity, https://cs.nyu.edu/~davise/ai/ambiguity.html,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <article-title>[2] BleacherReport, Stephen Curry is 'the best shooter who ever lived' per Warriors HC Steve Kerr</article-title>
          , https://bleacherreport.com/articles/10009994,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>G.</given-names>
            <surname>Katsogiannis-Meimarakis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Koutrika</surname>
          </string-name>
          ,
          <article-title>A deep dive into deep learning approaches for text-to-sql systems</article-title>
          , in: SIGMOD, ACM,
          <year>2021</year>
          , pp.
          <fpage>2846</fpage>
          -
          <lpage>2851</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>R.</given-names>
            <surname>Aly</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Guo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. S.</given-names>
            <surname>Schlichtkrull</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Thorne</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Vlachos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Christodoulopoulos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Cocarascu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mittal</surname>
          </string-name>
          , FEVEROUS:
          <article-title>Fact extraction and VERification over unstructured and structured information</article-title>
          , in: NIPS,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>G.</given-names>
            <surname>Karagiannis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Saeed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Papotti</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Trummer</surname>
          </string-name>
          ,
          <article-title>Scrutinizer: A mixed-initiative approach to large-scale, data-driven claim verification</article-title>
          ,
          <source>Proc. VLDB Endow</source>
          .
          <volume>13</volume>
          (
          <year>2020</year>
          )
          <fpage>2508</fpage>
          -
          <lpage>2521</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>W.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Schlinger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W. Y.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W. W.</given-names>
            <surname>Cohen</surname>
          </string-name>
          ,
          <article-title>Open question answering over tables and text</article-title>
          , in: ICLR, OpenReview.net,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>J.</given-names>
            <surname>Thorne</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Yazdani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Saeidi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Silvestri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Riedel</surname>
          </string-name>
          ,
          <string-name>
            <surname>A. Y. Levy</surname>
          </string-name>
          ,
          <article-title>From natural language processing to neural databases</article-title>
          ,
          <source>Proc. VLDB Endow</source>
          .
          <volume>14</volume>
          (
          <year>2021</year>
          )
          <fpage>1033</fpage>
          -
          <lpage>1039</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>P.</given-names>
            <surname>Lapadula</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Mecca</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Santoro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Solimando</surname>
          </string-name>
          , E. Veltri,
          <article-title>Humanity is overrated. or not. automatic diagnostic suggestions by greg</article-title>
          ,
          <source>ml (extended abstract)</source>
          ,
          <source>Communications in Computer and Information Science</source>
          <volume>909</volume>
          (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>P.</given-names>
            <surname>Lapadula</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Mecca</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Santoro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Solimando</surname>
          </string-name>
          , E. Veltri, Greg, ml
          <article-title>- machine learning for healthcare at a scale</article-title>
          ,
          <source>Health and Technology</source>
          <volume>10</volume>
          (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>E.</given-names>
            <surname>Veltri</surname>
          </string-name>
          , G. Badaro,
          <string-name>
            <given-names>M.</given-names>
            <surname>Saeed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Papotti</surname>
          </string-name>
          ,
          <article-title>Data ambiguity profiling for the generation of training examples</article-title>
          , in: ICDE, IEEE,
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>E.</given-names>
            <surname>Veltri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Santoro</surname>
          </string-name>
          , G. Badaro,
          <string-name>
            <given-names>M.</given-names>
            <surname>Saeed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Papotti</surname>
          </string-name>
          , Pythia:
          <article-title>Unsupervised generation of ambiguous textual claims from relational data</article-title>
          ,
          <source>in: SIGMOD</source>
          <year>2022</year>
          , ACM,
          <year>2022</year>
          , p.
          <fpage>2409</fpage>
          -
          <lpage>2412</lpage>
          . doi:
          <volume>10</volume>
          .1145/3514221.3520164.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>E.</given-names>
            <surname>Veltri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Santoro</surname>
          </string-name>
          , G. Badaro,
          <string-name>
            <given-names>M.</given-names>
            <surname>Saeed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Papotti</surname>
          </string-name>
          ,
          <article-title>Ambiguity detection and textual claims generation from relational data</article-title>
          ,
          <source>in: SEBD</source>
          , volume
          <volume>3194</volume>
          , CEUR-WS.org,
          <year>2022</year>
          , pp.
          <fpage>333</fpage>
          -
          <lpage>340</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>O.</given-names>
            <surname>Lehmberg</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Ritze</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Meusel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Bizer</surname>
          </string-name>
          ,
          <article-title>A large public corpus of web tables containing time and context metadata</article-title>
          , in: WWW, ACM,
          <year>2016</year>
          , pp.
          <fpage>75</fpage>
          -
          <lpage>76</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>C.</given-names>
            <surname>Rafel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Shazeer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Roberts</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Narang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Matena</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. J.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <article-title>Exploring the limits of transfer learning with a unified text-to-text transformer</article-title>
          ,
          <source>J. Mach. Learn. Res</source>
          .
          <volume>21</volume>
          (
          <year>2020</year>
          )
          <volume>140</volume>
          :
          <fpage>1</fpage>
          -
          <lpage>140</lpage>
          :
          <fpage>67</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>A.</given-names>
            <surname>Ratner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. H.</given-names>
            <surname>Bach</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Ehrenberg</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Fries</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Ré</surname>
          </string-name>
          , Snorkel:
          <article-title>Rapid training data creation with weak supervision</article-title>
          ,
          <source>in: VLDB</source>
          , volume
          <volume>11</volume>
          ,
          <year>2017</year>
          , p.
          <fpage>269</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>R.</given-names>
            <surname>Speer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Chin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Havasi</surname>
          </string-name>
          ,
          <article-title>Conceptnet 5.5: An open multilingual graph of general knowledge</article-title>
          ,
          <source>in: Thirty-first AAAI conference on artificial intelligence</source>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>