<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>A Sequence to Sequence approach for Knowledge Base Relation Linking</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Vito Barbara</string-name>
          <email>vitobarbara3@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Manuel Borroto</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Francesco Ricca</string-name>
          <email>francesco.riccag@unical.it</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>University of Calabria</institution>
          ,
          <addr-line>Rende CS 87036</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Currently, Question Answering (QA) has taken a central role in the area of the Semantic Web, allowing access to a large amount of information in knowledge bases in a simple way. In this eld, this is known as Knowledge Base Question Answering (KBQA). The most recent KBQA systems allow to abstract all those lay users from the complexities of the query languages like SPARQL and the ontology de nition by posing questions in natural language to retrieve information. Relation Linking (RL) is an essential component in end-to-end KBQA systems, allowing to map relations in natural language questions to corresponding relations in knowledge bases. In this paper, we propose an approach for RL that leverages the strength of the Deep Learning Sequence to Sequence models. We demonstrate the potential of our system by presenting its results on DBNQA and QALD-9, which are well-known datasets for Question Answering over DBpedia.</p>
      </abstract>
      <kwd-group>
        <kwd>Relation Linking Processing</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        The development of knowledge bases has gathered nowadays large volumes of
information concerning multiple domains. Among the best known is DBpedia [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ],
which covers various domains of interest. Unfortunately, access to this
information is complicated for those users unfamiliar with the SPARQL [
        <xref ref-type="bibr" rid="ref26">26</xref>
        ] query
language and the knowledge base de nition, and so the Knowledge Base
Question Answering systems have been introduced, as [
        <xref ref-type="bibr" rid="ref12 ref3 ref4">4,3,12</xref>
        ]. KBQA is a discipline
that permits to answer automatically to natural language questions, retrieving
information from knowledge bases. RL is a fundamental component in a KBQA
system. Its goal is to associate each relation in a natural language question with
the corresponding predicate in the knowledge base. For instance, given the
natural language question "How tall is the Ei el Tower?" an ideal RL system should
return the predicate "dbo:height". This task is made particularly di cult due
to a large number of predicates in knowledge bases and the syntactic di erence
between the relations in natural language questions and the relations in
knowledge bases. The possibility of multiple or implicit relations for a single question
makes the work more complicated. Our idea is to treat the relations as a sequence
of predicates, such that the RL task can be seen as a sequence to sequence [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ]
problem, where the input is a question, and the output is a sequence of relations.
To this aim, we propose a system called SeqRL. The considered knowledge base
is DBpedia, so any reference to knowledge bases is related to it. However, our
approach is quite exible: if it is necessary to use another knowledge base, it is
enough only to change the output vocabulary and re-train the model on a new
appropriate dataset.
      </p>
      <p>
        We empirically test the system on two datasets for question answering on
the DBpedia ontology, namely the DBpedia Neural Question Answering dataset
(DBNQA) [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ], and QALD-9 [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ].
      </p>
      <p>The remainder of the paper is structured as follows. In section 2, we talk
about some related works. Section 3 provides the preliminary elements necessary
to understand how our approach works. In section 4, we go into the particular
details of our approach. Section 5 focuses on the discussion of experiments and
results. And nally, in section 6, we provide some conclusions and aspects for
future work.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Related work</title>
      <p>
        Many question answering systems over knowledge graphs rely on relation
linking components in order to do a proper mapping between the natural language
question and the underlying knowledge graph. In this sense, Dubey et al.[
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]
proposed the EARL system, introducing the novelty that Entity Linking (EL) and
Relation Linking tasks are done jointly, improving not only e ectiveness but also
execution times. EARL proposes two ways to face the problem. The rst treats
the task as a Generalised Travelling Salesman Problem (GTSP) instance, while
the second uses an Machine Learning (ML) approach based on Text Similarity.
The authors proved the performance of the system using LC-QuAD v1.0 [
        <xref ref-type="bibr" rid="ref22">22</xref>
        ]
and QALD-7 [
        <xref ref-type="bibr" rid="ref24">24</xref>
        ] datasets, obtaining good results. They used accuracy to assess
the system. The main limitation of EARL is that it does not address questions
with hidden relations, that is, relations that cannot be directly inferred from the
text.
      </p>
      <p>
        Pan et al. [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ] proposed a new framework for Relation Linking, called EERL,
which exploits entities contained in questions to support the task. They expand
the set of relation candidates by using properties that are logically connected to
the target entities. The framework takes as input a natural language question
and a knowledge graph, and it is composed of ve modules: Relation Keyword
Extractor, Keyword-based Relation Expansion, Entity Linking, Entity-based
Relation Expansion, Relation Ranking. The rst component, Relation Keyword
Extractor, permits extraction of relation phrases contained in the question. In
the second module, Keyword-based Relation Expansion, the authors use
background knowledge to get a list of associated relation phrases. The Entity Linking
module extrapolates the entities included in the question. Entity-based Relation
Expansion is the main module, which allows to obtain all relation candidates,
by expanding explicit and implicit relations. Relation expansion is based on the
following hypothesis \The relations in questions are properties of the entities
occurring in the question or properties of the types of these entities". In the
end, the last module, Relation Candidate Ranking, is useful to select the best
relations from the candidates. The authors test the system on three datasets:
QALD-5 [
        <xref ref-type="bibr" rid="ref23">23</xref>
        ], QALD-7, LC-QuAD v1.0. The main weakness of the system is the
real e ectiveness of the proposed hypothesis, in fact, they noticed that some of
the questions' relations don't appear in the relation candidates.
      </p>
      <p>
        Ahmad Sakor et al. [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ] introduced a rule-based approach to address the
problem of joint entity and relation linking within the short text, called Falcon.
The system mainly appeals to the fundamental principles of English morphology,
such as compounding, right-hand rules for headword identi cation, and also uses
an extended knowledge base. The authors use the well-known spaCy [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] library
to annotate the text with POS tag information and then create a list of candidate
entities and relations. For relations, candidates are proposed based on the verbs
found in the text. Finally, a ranking mechanism organizes the candidates. Falcon
was tested using the QALD-7 and LC-QuAD v1.0 datasets, outperforming the
compared systems like EARL. Falcon, like EARL, has problems with questions
that contain implicit relations due to the nature of the approach.
      </p>
      <p>
        An interesting approach is the Semantic LINkinG system (SLING) proposed
by Mihindukulasooriya et al. [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] SLING is a distant supervision-based
system that leverages semantic parsing such as Abstract Meaning Representation
(AMR). The authors use distance supervised techniques to address the
challenge of lack of training data. The system is divided into two main components:
Question Metadata Generation and Relation Linking. The rst component takes
as input the question text and its AMR graph. The graph is converted into a
set of intermediate AMR triples, and the query text is used to obtain
additional knowledge base information by using Entity Linking tools, like BLINK.
The second component works using four di erent modules for Relation Linking,
two supervised and two unsupervised approaches. Finally, the system aggregates
the results of the modules to obtain a nal ranked list of relations. SLING was
executed on the well-known QALD-7, QALD-9, Lc-QUAD v1.0 datasets,
demonstrating better performance than the related systems. It is not clear from the
paper the limitations of this approach, so we think it may su er from the same
problem as the previously discussed RL approaches since it depends on the
semantic mapping of the text that does not contain a reference to the implicit
relationships.
      </p>
      <p>
        The Generative relation linking for question answering over knowledge bases
(GenRL) approach by Rossiello et al. [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ] is considered the state-of-the-art in
this type of task. GenRL works by using a Seq2Seq model to generate a
sequence of relations from the question text. This approach extends the question
text by introducing information about the entities present and their linked
relations, providing additional context for training the baseline model. Another
interesting point is that GenRL starts from a pre-trained transformer model,
called BART [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] that is ne-tuned depending on the di erent datasets present
in the literature. The system also includes a validation mechanism that helps
to improve the results obtained by the neural network. During the evaluation,
GenRL demonstrated better performance than related systems on the DBpedia
and Wikidata [
        <xref ref-type="bibr" rid="ref25">25</xref>
        ] knowledge bases.
      </p>
      <p>
        Our work is based on the results described in [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], which were extended with a
deeper experimental analysis. SeqRL has some similarities with GenRL, like the
usage of Seq2Seq models, but it was developed independently by using di erent
tools. The main di erence between the two approaches is that they use a
pretrained seq2seq model, BART, which is based on a transformer architecture,
while we use an LSTM-based model trained from scratch.
      </p>
      <p>Also, they extend their model with knowledge integration and validation
mechanisms, and this has positively impacted their performance.
3
3.1</p>
    </sec>
    <sec id="sec-3">
      <title>Preliminaries</title>
      <sec id="sec-3-1">
        <title>Knowledge Bases and SPARQL</title>
        <p>
          A knowledge base can be de ned as a formal description of a domain of interest
that is suitable to be managed by an engine reasoning about the facts modeled
in the knowledge base itself, e.g., query existing knowledge or obtain new
knowledge. A formal description of knowledge as a set of concepts within a domain
and the relationships that hold between them is called ontology [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ].
        </p>
        <p>
          Ontologies allow the organization of information in such a way that is
interpretable by both humans and machines. To enable such a description, it is
necessary to formally specify some components such as individuals, classes,
attributes, and relations as well as restrictions, rules, and axioms. As a result,
ontologies do not only introduce a shared and reusable knowledge
representation but can also be used to infer new knowledge about the domain. An ontology,
together with a set of individual instances of the ontology classes, constitutes a
knowledge base (KB) [
          <xref ref-type="bibr" rid="ref17">17</xref>
          ].
        </p>
        <p>
          To create an Ontology, it is necessary to resort to a set of languages and
technologies de ned to this aim. The most common ontologies are de ned by
using the RDF, RDFS, and OWL languages, which are maintained by W3C [
          <xref ref-type="bibr" rid="ref28">28</xref>
          ]
and constitute the de-facto standards for this type of task. In this paper, we adopt
RDF as the ontology speci cation language of reference. With RDF, information
is represented by a collection of hsubject predicate objecti triples, where the
predicate establishes a binary relationship between subject and object. So, a
knowledge base is composed of a set of triples known as RDF-graph [
          <xref ref-type="bibr" rid="ref27">27</xref>
          ].
        </p>
        <p>The resources in a KB can be de ned using URIs, allowing to reference
non-local resources. This property permits the interaction among multiple KBs,
making the accessible information grow considerably, not only in volume but
also in the diversity of domains. To retrieve the information, given a knowledge
base, one has to resort to a proper query language like SPARQL, the de-facto
standard for this type of task.</p>
        <p>
          SPARQL is an SQL-like language to query RDF-graphs. The syntax and
semantics of the language allow the user to query a KB by de ning triples looking
for a match with subject-predicate-object patterns within the graph [
          <xref ref-type="bibr" rid="ref28">28</xref>
          ]. The
following is an example of a SPARQL query:
PREFIX dbo: &lt;http://dbpedia.org/ontology/&gt;
PREFIX dbr: &lt;http://dbpedia.org/resource/&gt;
SELECT ?place WHERE { dbr:Barack_Obama dbo:birthPlace ?place }
        </p>
        <p>The PREFIX construct de nes a namespace useful to disambiguate concepts
with the same name. In the third line, SELECT returns a list containing the
values of the variable ?place, while the WHERE clause contains the triples to
be matched against the RDF-graph. In the above example, dbr:Barack Obama
identi es a KB resource, and dbo:birthPlace is a predicate used to link the place
of birth. Note that the identi ers that begin with \?" are considered variables.
The query models the answer to the question \Where was Barack Obama born?".
In addition to the SELECT clause, there are other ways to obtain the results:
{ CONSTRUCT constructs and returns an RDF-graph by substituting
variables in the query pattern.
{ DESCRIBE provides an RDF-graph describing the resources that were found.
{ ASK returns a boolean value indicating whether the query pattern matches
or not.</p>
        <p>Like SQL, the language provides di erent constructs to modify the results of a
query, as ORDER BY, DISTINCT, OFFSET, and LIMIT. Another signi cant
construct is the FILTER clause, which permits to apply restrictions to results,
by using, for instance, a simple boolean expression or a regular expression.
3.2</p>
      </sec>
      <sec id="sec-3-2">
        <title>Recurrent Neural Networks</title>
        <p>
          The approach proposed in this paper takes advantage of the development achieved
in the Deep Learning eld and addresses the current problem using Arti cial
Neural Networks (ANN) [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ], speci cally Recurrent Neural Networks (RNN) [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ].
Due to a particular way of processing data, RNNs have contributed to obtaining
good results in tasks involving the processing of data sequences, and therefore
they are also good at processing natural language as well, if we consider the
natural language as a semantically ordered sequence of words.
        </p>
        <p>
          An RNN works by iterating over the elements of the sequence S and keeping a
state h that contains information relative to what was already processed so that
the result of processing the element at time t is also conditioned by the previous
information t 1 [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ]. At each time step t the state h(t) is updated by mean of
h(t) = f (h(t 1); xt), where f is a non-linear function. This way of operation is
useful to capture semantic relationships between the words in a sentence.
        </p>
        <p>
          Long short-term memory networks. When processing natural language, it is
common to deal with sentences long enough to cause the memory mechanism
employed by RNNs to collapse, making it hard to learn long-term dependencies.
The phenomenon explained before is known as Vanishing Gradient Descent, and
it was addressed by Hochreiter et. al [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ] in 1997, introducing a particular type of
RNN called Long short-term memory (LSTM). LSTM adds a way of transporting
information through many time steps. Imagine a conveyor belt running parallel
to the sequence you are processing. Information from the sequence can jump
onto the conveyor belt at any point, be transported to a later timestep, and
jump o , intact, when you need it. Essentially, an LSTM saves information for
later, thus preventing older signals from gradually vanishing during processing
[
          <xref ref-type="bibr" rid="ref6">6</xref>
          ]. The success of the LSTMs lies in a mechanism called Gates used to compute
the hidden states. The gating mechanism can regulate the ow of information
and decide what information is important to keep or throw away. This is done
by mean of:
i = (xtU i + st 1W i)
o = (xtU o + st 1W o)
ct = ct 1 f + g i
f = (xtU f + st 1W f )
g = tanh(xtU g + st 1W g)
st = tanh(ct) o
where the input i, forget f , and output o represent gates that are squashed
by the sigmoid into vectors of values between 0 and 1. Multiplying the vectors
determines how much of the other vectors to let into the current input state. g
is a candidate hidden state that is computed based on the current input and the
previous hidden state. ct is used as the internal memory, which is a combination
of the previous memory ct 1 multiplied by the input gate, and the hidden state
st is a combination of the internal memory and the output gate.
4
        </p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Neural network for Relation Linking</title>
      <p>A naive neural approach for RL is to train a multi-class multi-label classi er
that predicts predicates contained in the questions. As previously mentioned,
the number of predicates is huge, just think that DBpedia contains more than
50:000 predicates, an unmanageable quantity in terms of resources for a classical
neural network. In this paper, we tackle the task of RL as a sequence to sequence
problem by using well-known Seq2Seq models. Therefore, the task of our
system is to take a natural language question as input and returns a sequence of
predicates contained in the underlying KB. For these models, the de nition of
input and output vocabularies is necessary. In this case, input vocabulary is the
English language, while output vocabulary is the set of predicates included in
DBpedia. All predicates have been obtained through a SPARQL query.</p>
      <p>To provide the model with the proper input to improve the learning phase,
we decided to rewrite the questions and the predicates to lowercase. We also
normalized the predicates by deleting the pre x, to simplify the linking process.
The normalization process has been done because, in the vocabulary, there exist
relations semantically equivalent that the model is not able to distinguish, like
"dbo:name" and "dbp:name".</p>
      <p>
        The output vocabulary has a reasonable number of predicates, especially
when compared to the input one; for instance, consider that the Oxford English
Dictionary contains more than 700:000 words. Neural networks, generally, can
manage words, which they have already seen during the training phase and so
contained in the training set, but currently, for this task, there are no datasets
complete enough to include all words. To address this problem, we used
pretrained Word Embedding. In particular, we used models that are trained on
Wikipedia through FastText [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. They allow to have access to thousands of word
vectors, learned over millions of words, and so, they can provide a vectorial
representation even for terms not included in the training set. Note that output
vocabulary is even primarily English, therefore, we used FastText models to
extract the embeddings from predicates.
      </p>
      <p>In this case, the Seq2Seq model has a classical structure Encoder-Decoder, in
which the encoder has the goal to extract a vector V from the question, namely
context vector, that encapsulates the entire meaning of the phrase; while the
decoder tries to translate this vector in a sequence of predicates, belonging to
output vocabulary. Both are composed of a recurrent layer of type LSTM with
128 units. During the training phase, we used the Teacher forcing, a famous
training strategy for the recurrent network, which permits a rapid convergence.
It consists in using as input of the decoder the ground truths of the previous
step, instead of the last output predicted by the model. Figure 1 shows the
model architecture used during the training phase.</p>
      <p>During the inference phase, the ground truths are not available, so the input
of the decoder is simply what it predicts at the previous step. Also, in this phase,
the iterative process of the recurrent network needs to be implemented at hand,
splitting the original model into two sub-models: encoder and decoder, as shown
in Figure 2.</p>
      <p>(a) Encoder
(b) Decoder</p>
      <p>
        Another characteristic of SeqRL, suggested by the authors of article [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ], is
to reverse the words of the input question.
      </p>
      <p>Figure 3 shows an example of an instance during the training phase. More
deeply, the word embeddings are the actual input of the encoder and the
decoder, generated by FastText ; while the output of the decoder is a probability
distribution over the output vocabulary, obtained through the Softmax
activation function.
5
5.1</p>
    </sec>
    <sec id="sec-5">
      <title>Experiments</title>
      <sec id="sec-5-1">
        <title>Experiment Setup</title>
        <p>We have implemented our models by using Keras, a well-known framework for
deep learning, on top of TensorFlow. The experiments were executed on Google
Collaboratory, a virtual environment based on Jupyter Notebook, and on a Linux
server with 500 GB of ram and two GPU Nvidia Tesla V100 with 16 GB of
memory. We considered two well-known datasets for KBQA over the DBpedia
ontology: DBNQA and QALD-9. To assess the system, we adopted precision,
recall and F1-score metrics.
5.2</p>
      </sec>
      <sec id="sec-5-2">
        <title>Evaluation on DBNQA</title>
        <p>DBpedia Neural Question Answering (DBNQA) is a huge dataset for KBQA,
composed by 894:499 question-query pairs. From each query, we extrapolated all
predicates, so they can be used as ground truths. Also, we removed all statistical
predicates, e.g. wikiPageLength. To evaluate the performance of the system, we
decided to use k-fold cross-validation, with k = 5, since there are no other
systems, to the best of our knowledge, that use DBNQA dataset for evaluation
for Relation Linking, and so we cannot make a proper comparison. Notice that
the aforementioned dataset is used mainly for Question Answering. For each step,
the model was trained for 4 epochs, and Table 1 shows the obtained results. As
shown, our system performs very well, reaching a score greater than 0:98 in all
metrics. More in detail, we report that the dataset is quite repetitive, and this
facilitates the behavior of the model.</p>
        <p>In Table 2, we show a set of examples to analyze the behavior of the model.
The examples in rows 1 and 7 show the ability of SeqRL to infer relations that
are not directly expressed in the text, in fact, it is able to predict the unseen
relations architect and locatedinarea.</p>
        <p>Who built the Ei el Tower?
Which instruments does Cat
Stevens play?
Which book has the most pages?
When did princess Diana die?
Where did princess Diana die?</p>
        <p>dbo:locatedInArea, locatedinarea,
What is the highest mountain in Italy? rdf:type, dbo:Mountain, type, mountain,
dbo:elevation, elevation</p>
        <p>Many approaches in the literature are not able to tackle this situation. For
instance, Falcon1, given the natural language question "Who built the Ei el
Tower?", returns the wrong relation numberBuilt.
5.3</p>
      </sec>
      <sec id="sec-5-3">
        <title>Evaluation on QALD-9</title>
        <p>Our approach is designed mainly for huge datasets, like DBNQA. But, none
of the related works use DBNQA to evaluate their system, so we decided to
consider a dataset largely used in the literature to make comparison:
QALD9. The Question Answering over Linked Data (QALD) is a series of challenges
that aim to provide benchmarks for assessing and comparing KBQA systems on
DBpedia. We considered the benchmark proposed as part of the ninth edition
of QALD, known as QALD-9. The dataset contains 558 question-query pairs
in 11 di erent languages. The data is split into 408 training and 150 testing
questions, and we focus on the ones expressed in English. It is important to note
that this dataset is very challenging to be approached using learning techniques,
given the very small training set not covering all questions types of the test set.
For this reason, we trained our model on both DBNQA and QALD-9 training
set. In particular we rst, obtained a pre-trained model on DBNQA and then
ne-tuned it performing training on QALD-9. Notice that DBNQA is built by
using Lc-QUAD v1.0 and QALD-7-Train.</p>
        <p>
          Table 3 shows comparison of SeqRL with some related approaches. The
results for the related approaches have been taken from [
          <xref ref-type="bibr" rid="ref19">19</xref>
          ] (page 7 table 1). It is
important to observe that the performance of FALCON 1.0, SLING and GenRL
can be in uenced by the problem of the equivalent predicates, due to the pre x
that we ignore. As we can see, our approach performs good, outperforming
systems like FALCON 1.0 and SLING, while the most recent work, GenRL, has a
comparable performance (if not better performance).
1 https://labs.tib.eu/falcon/
The paper presented an approach based on deep neural networks for Relation
Linking. We took advantage of the Sequence to Sequence models based on LSTM
networks to address the problem by considering the predicates as a sequence of
words. Our system demonstrated its great potential, obtaining optimal results
on DBNQA and good ones on the challenging dataset QALD-9.
        </p>
        <p>As future work, we want to extend our approach with other Natural Language
Processing tools, like Entity Linking (EL). EL permits to identify entities into
the question, which can be helpful to extract a set of candidate predicates to be
used to simplify the process of linking. We also plan to extend our experiments
considering other recently-developed KBQA benchmarks.</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Barbara</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Borroto</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ricca</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          :
          <article-title>A Sequence to Sequence approach for Knowledge Base Relation Linking</article-title>
          .
          <source>Master's thesis</source>
          , University of Calabria (Sep
          <year>2021</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Bojanowski</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Grave</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Joulin</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mikolov</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>Enriching word vectors with subword information</article-title>
          .
          <source>arXiv preprint arXiv:1607.04606</source>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Borroto</surname>
            ,
            <given-names>M.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ricca</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cuteri</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          :
          <article-title>Reducing the impact of out of vocabulary words in the translation of natural language questions into sparql queries</article-title>
          .
          <source>ArXiv abs/2111</source>
          .03000 (
          <year>2021</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Cabrio</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cojan</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Aprosio</surname>
            ,
            <given-names>A.P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Magnini</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lavelli</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gandon</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          :
          <article-title>Qakis: an open domain qa system based on relational patterns</article-title>
          . In: International Semantic Web Conference,
          <string-name>
            <surname>ISWC</surname>
          </string-name>
          <year>2012</year>
          (
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Dubey</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Banerjee</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chaudhuri</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lehmann</surname>
          </string-name>
          , J.:
          <article-title>Earl: joint entity and relation linking for question answering over knowledge graphs</article-title>
          .
          <source>In: International Semantic Web Conference</source>
          . pp.
          <volume>108</volume>
          {
          <fpage>126</fpage>
          . Springer (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Francois</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>Deep learning with Python</article-title>
          . Manning Publications Company (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Giles</surname>
            ,
            <given-names>C.L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kuhn</surname>
            ,
            <given-names>G.M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Williams</surname>
            ,
            <given-names>R.J.:</given-names>
          </string-name>
          <article-title>Dynamic recurrent neural networks: Theory and applications</article-title>
          .
          <source>IEEE Trans. on Neur.Net</source>
          .
          <volume>5</volume>
          (
          <issue>2</issue>
          ),
          <volume>153</volume>
          {
          <fpage>156</fpage>
          (
          <year>1994</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Gruber</surname>
            ,
            <given-names>T.R.</given-names>
          </string-name>
          :
          <article-title>Toward principles for the design of ontologies used for knowledge sharing? Int</article-title>
          .
          <string-name>
            <given-names>J.</given-names>
            <surname>Hum</surname>
          </string-name>
          .-Comput. Stud.
          <volume>43</volume>
          (
          <issue>5-6</issue>
          ),
          <volume>907</volume>
          {
          <fpage>928</fpage>
          (
          <year>1995</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Hartmann</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Marx</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Soru</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>Generating a large dataset for neural question answering over the DBpedia knowledge base (</article-title>
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Hochreiter</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Schmidhuber</surname>
            ,
            <given-names>J.:</given-names>
          </string-name>
          <article-title>Long short-term memory</article-title>
          .
          <source>Neural Computation</source>
          <volume>9</volume>
          (
          <issue>8</issue>
          ),
          <volume>1735</volume>
          {
          <fpage>1780</fpage>
          (
          <year>1997</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Honnibal</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Montani</surname>
            , I., Van Landeghem,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Boyd</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>spaCy: Industrial-strength Natural Language Processing in Python (</article-title>
          <year>2020</year>
          ). https://doi.org/10.5281/zenodo.1212303, https://doi.org/10.5281/zenodo. 1212303
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Kapanipathi</surname>
          </string-name>
          , et al.:
          <article-title>Question answering over knowledge bases by leveraging semantic parsing and neuro-symbolic reasoning</article-title>
          . arXiv preprint arXiv:
          <year>2012</year>
          .
          <volume>01707</volume>
          (
          <year>2020</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Lehmann</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Isele</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jakob</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jentzsch</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kontokostas</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mendes</surname>
            ,
            <given-names>P.N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hellmann</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Morsey</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Van Kleef</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          , et al.:
          <article-title>Dbpedia{a large-scale, multilingual knowledge base extracted from wikipedia</article-title>
          .
          <source>Semantic Web</source>
          <volume>6</volume>
          (
          <issue>2</issue>
          ),
          <volume>167</volume>
          {
          <fpage>195</fpage>
          (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Lewis</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Liu</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Goyal</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ghazvininejad</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mohamed</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Levy</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stoyanov</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zettlemoyer</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          :
          <article-title>Bart: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension</article-title>
          . arXiv preprint arXiv:
          <year>1910</year>
          .
          <volume>13461</volume>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Mihindukulasooriya</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rossiello</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kapanipathi</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Abdelaziz</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ravishankar</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yu</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gliozzo</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Roukos</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gray</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Leveraging semantic parsing for relation linking over knowledge bases</article-title>
          .
          <source>In: International Semantic Web Conference</source>
          . pp.
          <volume>402</volume>
          {
          <fpage>419</fpage>
          . Springer (
          <year>2020</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>Ngomo</surname>
          </string-name>
          , N.:
          <article-title>9th challenge on question answering over linked data (qald-9). language 7(1) (</article-title>
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <surname>Noy</surname>
            ,
            <given-names>N.F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>McGuinness</surname>
            ,
            <given-names>D.L.</given-names>
          </string-name>
          , et al.:
          <article-title>Ontology development 101: A guide to creating your rst ontology (</article-title>
          <year>2001</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          18.
          <string-name>
            <surname>Pan</surname>
            ,
            <given-names>J.Z.</given-names>
          </string-name>
          , Zhang,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Singh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            ,
            <surname>Harmelen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.v.</given-names>
            ,
            <surname>Gu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            ,
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <surname>Z.</surname>
          </string-name>
          :
          <article-title>Entity enabled relation linking</article-title>
          . In: Ghidini,
          <string-name>
            <given-names>C.</given-names>
            ,
            <surname>Hartig</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            ,
            <surname>Maleshkova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Svatek</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            ,
            <surname>Cruz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I.</given-names>
            ,
            <surname>Hogan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            ,
            <surname>Song</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            ,
            <surname>Lefrancois</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Gandon</surname>
          </string-name>
          ,
          <string-name>
            <surname>F</surname>
          </string-name>
          . (eds.)
          <source>The Semantic Web { ISWC 2019</source>
          . pp.
          <volume>523</volume>
          {
          <fpage>538</fpage>
          . Springer International Publishing,
          <string-name>
            <surname>Cham</surname>
          </string-name>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          19.
          <string-name>
            <surname>Rossiello</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mihindukulasooriya</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Abdelaziz</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bornea</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gliozzo</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Naseem</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kapanipathi</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Generative relation linking for question answering over knowledge bases</article-title>
          .
          <source>In: International Semantic Web Conference</source>
          . pp.
          <volume>321</volume>
          {
          <fpage>337</fpage>
          . Springer (
          <year>2021</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          20.
          <string-name>
            <surname>Sakor</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          , Onando Mulang',
          <string-name>
            <given-names>I.</given-names>
            ,
            <surname>Singh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            ,
            <surname>Shekarpour</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            ,
            <surname>Esther</surname>
          </string-name>
          <string-name>
            <surname>Vidal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Lehmann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            ,
            <surname>Auer</surname>
          </string-name>
          ,
          <string-name>
            <surname>S.</surname>
          </string-name>
          : Old is gold:
          <article-title>Linguistic driven approach for entity and relation linking of short text</article-title>
          .
          <source>In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</source>
          , Volume
          <volume>1</volume>
          (Long and Short Papers). pp.
          <volume>2336</volume>
          {
          <fpage>2346</fpage>
          . Association for Computational Linguistics, Minneapolis,
          <source>Minnesota (Jun</source>
          <year>2019</year>
          ). https://doi.org/10.18653/v1/
          <fpage>N19</fpage>
          -1243, https://aclanthology.org/N19-1243
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          21.
          <string-name>
            <surname>Sutskever</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vinyals</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Le</surname>
            ,
            <given-names>Q.V.</given-names>
          </string-name>
          :
          <article-title>Sequence to sequence learning with neural networks</article-title>
          .
          <source>In: NIPS</source>
          . pp.
          <volume>3104</volume>
          {
          <issue>3112</issue>
          (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          22.
          <string-name>
            <surname>Trivedi</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Maheshwari</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dubey</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lehmann</surname>
          </string-name>
          , J.:
          <article-title>Lc-quad: A corpus for complex question answering over knowledge graphs</article-title>
          .
          <source>In: International Semantic Web Conference</source>
          . pp.
          <volume>210</volume>
          {
          <fpage>218</fpage>
          . Springer (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          23.
          <string-name>
            <surname>Unger</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Forascu</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lopez</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ngomo</surname>
            ,
            <given-names>A.C.N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cabrio</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cimiano</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Walter</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Question answering over linked data (qald-5)</article-title>
          . In: CLEF (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          24.
          <string-name>
            <surname>Usbeck</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ngomo</surname>
            ,
            <given-names>A.C.N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Haarmann</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Krithara</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          , Roder,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Napolitano</surname>
          </string-name>
          , G.:
          <article-title>7th open challenge on question answering over linked data (qald-7)</article-title>
          . In: Dragoni,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Solanki</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Blomqvist</surname>
          </string-name>
          , E. (eds.) Semantic Web Challenges. pp.
          <volume>59</volume>
          {
          <fpage>69</fpage>
          . Springer International Publishing,
          <string-name>
            <surname>Cham</surname>
          </string-name>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          25.
          <string-name>
            <surname>Vrandecic</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          , Krotzsch, M.:
          <article-title>Wikidata: A free collaborative knowledgebase</article-title>
          .
          <source>Commun. ACM</source>
          <volume>57</volume>
          (
          <issue>10</issue>
          ),
          <volume>7885</volume>
          (Sep
          <year>2014</year>
          ). https://doi.org/10.1145/2629489, https: //doi.org/10.1145/2629489
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          26.
          <article-title>W3C: SPARQL 1.1 query language (</article-title>
          <year>2013</year>
          ), https://www.w3.org/TR/ sparql11-query/
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          27. W3C:
          <article-title>Resource description framework (RDF) (</article-title>
          <year>2014</year>
          ), https://www.w3.org/RDF/
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          28. W3C:
          <article-title>Semantic web standards (</article-title>
          <year>2014</year>
          ), https://www.w3.org
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>