<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Comparing Research Contributions in a Scholarly Knowledge Graph</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Allard Oelen</string-name>
          <email>oelen@l3s.de</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Mohamad Yaser Jaradeh</string-name>
          <email>jaradeh@l3s.de</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Kheir Eddine Farfar</string-name>
          <email>kheir.farfar@tib.eu</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Markus Stocker</string-name>
          <email>markus.stocker@tib.eu</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Sören Auer</string-name>
          <email>auer@tib.eu</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>L3S Research Center</institution>
          ,
          <addr-line>Leibniz</addr-line>
          ,
          <institution>University of Hannover</institution>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>TIB Leibniz Information Centre for, Science and Technology</institution>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2019</year>
      </pub-date>
      <abstract>
        <p>Conducting a scientific literature review is a time consuming activity. This holds for both finding and comparing the related literature. In this paper, we present a workflow and system designed to, among other things, compare research contributions in a scientific knowledge graph. In order to compare contributions, multiple tasks are performed, including finding similar contributions, mapping properties and visualizing the comparison. The presented workflow is implemented in the Open Research Knowledge Graph (ORKG) which enables researchers to find and compare related literature. A preliminary evaluation has been conducted with researchers. Results show that researchers are satisfied with the usability of the user interface, but more importantly, they acknowledge the need and usefulness of contribution comparisons.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>INTRODUCTION</title>
      <p>
        When conducting scientific research, finding and comparing
stateof-the-art literature is an important activity. Mainly due to the
unstructured way of publishing scholarly knowledge it is currently
time consuming to find and compare related literature. The Open
Research Knowledge Graph1 (ORKG) [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] is a system designed to
acquire, publish and process structured scholarly knowledge
published in the scholarly literature. One of the main features of the
ORKG is the ability to automatically compare related literature.
      </p>
      <p>
        The benefits of having scholarly knowledge structured include,
among others, the ability to easily find, retrieve but also compare
such knowledge. Comparing resources (scholarly knowledge or
other) can be useful in many contexts [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ], for instance resources
describing cities for their population, area and other attributes.
Comparing structured data is useful particularly when compared
resources are described with similar or even same properties.
Although knowledge graphs are of course structured, resources—even
1http://orkg.org
Copyright ©2019 for this paper by its authors. Use permitted under Creative Commons
License Attribution 4.0 International (CC BY 4.0).
of the same type—are often described using diferent and diferently
named attributes. Moreover, diferent hierarchical structures can
complicate a comparison of two resources.
      </p>
      <p>In this paper, we present a workflow that describes how to select
and compare resources describing scholarly knowledge in graph
databases. We implement this workflow in the ORKG, which
enables the comparison of related literature, including state-of-the-art
overviews. In the ORKG, these resources are specifically called
research contributions. A research contribution relates the research
problem addressed by the contribution, the research method and (at
least one) research result. Currently, we do not further constrain the
description of these resources. Users can adopt arbitrary third-party
vocabularies to describe problems, methods, and results. We thus
tackle the following research questions:
• RQ1: How to compare research contributions in a graph based
system?
• RQ2: How to efectively specify and visualize research
contribution comparisons in a user interface?
2</p>
    </sec>
    <sec id="sec-2">
      <title>RELATED WORK</title>
      <p>
        Resource (or entity) comparison is a well-known task in a variety of
information systems, for instance in e-commerce or hotel booking
systems. In e-commerce, products can be compared in order to help
customers during the decision process [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ]. These comparisons are
often based on a predefined set of properties to compare (e.g., price
and color). This does not apply when comparing resources in a
community-created knowledge graph where there is no predefined
set of properties. Petrova et al. created a framework to compare
heterogeneous entities in RDF graphs using SPARQL queries [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ].
In this framework, both the similarities and diferences between
entities are determined.
      </p>
      <p>
        The task of comparing research contributions in a graph system
can be decomposed into multiple sub tasks. The first sub task is
ifnding suitable contributions to compare. The most suitable
comparison candidates are similar resources. The actual retrieval of
similar resources can be seen as an information retrieval problem,
with techniques such as TF-IDF [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. Measures to calculate the
structural similarity between RDF graphs have been proposed in
the literature (e.g., Maillot et al. [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]). A second sub task is
matching semantically similar predicates. Determining the similarity of
resources is a recurring problem in dataset interlinking [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] or the
more general task of ontology alignment/matching [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ]. For
property mapping, techniques of interest include edit distance (e.g.,
Jaro-Winkler [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ] or Levenshtein [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]) and vector distance.
Gromann and Declerck evaluated the performance of word vectors for
ontology alignment and found that FastText [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] performed best [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ].
      </p>
      <p>
        As suggested above, an efective automated related literature
comparison relies on scholarly knowledge being structured. There
is substantial related work on representing scholarly knowledge in
structured form. Building on the work of numerous philosophers of
science, Hars [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] proposed a comprehensive scientific knowledge
model that includes concepts such as theory, methodology and
statement. More recently, ontologies were engineered to describe
diferent aspects of the scholarly communication process. Among
them are CiTO2 and C4O3 for recording citation related concepts,
FaBiO4 and BiRO5 for capturing bibliographic data and PRO6, PSO7
and PWO8 for the publication process. Additionally DoCO9 can be
used to describe the structure of a document which can be
complemented by DEO10 to also include rhetorical elements to describe
the scientific discourse [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. Among others, these ontologies are part
of the Semantic Publishing and Referencing Ontologies (SPAR), a
collection of ontologies that can be used to describe scholarly
publishing and referencing of documents [
        <xref ref-type="bibr" rid="ref13 ref14 ref6">6, 13, 14</xref>
        ]. Ruiz Iniesta and
Corcho [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ] reviewed the state-of-the-art ontologies to describe
scholarly articles.
      </p>
      <p>
        These ontologies are designed to capture primarily metadata
about and structure of scholarly articles, not the actual research
contributions (scholarly knowledge) communicated in articles. In
order to conduct a useful comparison, only comparing article
metadata and structure is oftentimes not suficient. Rather, a
comparison should include (structured descriptions of) problem, materials,
methods, results and perhaps other aspects of scholarly work. The
Open Research Knowledge Graph (ORKG) [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] supports creating
such structured descriptions of scholarly knowledge. We thus
embed the research contribution comparison presented in this paper
in the larger efort of the ORKG project, which aims to advance
scholarly communication infrastructure.
3
      </p>
    </sec>
    <sec id="sec-3">
      <title>WORKFLOW</title>
      <p>We present a workflow that describes how to perform a
comparison of research contributions, thereafter more generally referred
to as resources. This workflow consists of four diferent steps: 1)
select comparison candidates, 2) select related statements, 3) map
properties and 4) visualize comparison. The workflow is depicted in
Figure 1. Section 4 presents the implementation for the individual
steps. We now discuss each step of the workflow in more detail.
3.1</p>
    </sec>
    <sec id="sec-4">
      <title>Select comparison candidates</title>
      <p>To perform a comparison, a starting resource is needed. This
resource is called main resource and is always manually selected
2http://purl.org/spar/cito
3http://purl.org/spar/c4o
4http://purl.org/spar/fabio
5http://purl.org/spar/biro
6http://purl.org/spar/pro
7http://purl.org/spar/pso
8http://purl.org/spar/pwo
9http://purl.org/spar/doco
10http://purl.org/spar/deo
by a user. The main resource is compared against other
comparison resources. There are two diferent approaches for selecting
the comparison resources. The first approach automatically selects
comparison resources based on similarity. The second approach
lets users manually select resources.</p>
      <p>3.1.1 Find similar resources. Comparing resources makes only
sense when resources can sensibly be compared. For example, it
does not make (much) sense to compare a city (e.g., dbpedia:berlin)
to a car brand (e.g., dbpedia:volkswagen). This of course does not
only apply to comparison in knowledge graphs but also applies to
comparison in other kinds of databases. We thus argue that it makes
only sense to compare resources that are similar. More specifically,
resources that share the same (or a similar set of) properties are
good comparison candidates. To illustrate this, consider following
resources: dbpedia:berlin and dbpedia:new_york_city. Both resources
share the property dbo:populationTotal which makes them suitable
for comparison. Finding similar resources is therefore based on
ifnding resources that share the same or similar properties.</p>
      <p>
        To do so, each comparison resource is converted into a string.
This string is generated by concatenating all properties of the
resource (Algorithm 1). The resulting string is stored. TF-IDF [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] is
used to query the store and the string for the main resource is used
as query. The search returns the most similar resources. The top-k
resources are selected and form a set of resources that is used in
the next step.
      </p>
      <p>Algorithm 1 Paper indexing
1: procedure IndexPaper(paper)
2: for each property in paper do
3: propertyStrinд ← propertyStrinд + property
4: save propertyStrinд
3.1.2 Manual selection. There are scenarios where comparison
based on similarity is not suitable. For example, a user may want
to compare Germany and France to see which country has the
highest GDP. In this case, there is no need to automatically select
resources to be compared because they are determined by the user.
Therefore, manual selection of resources should also be supported.
How to manually select resources is an implementation detail and
a proposal for how to do this is presented in Section 4. The result
of manual selection is a set of resources used for the comparison.
3.2
This step selects the statements related to the resources to be
compared returned in the previous step. Statements are selected
transitively to match resources in subject or object position. This search
is performed until a predefined maximum transitive depth δ has
been reached. The intuition is that the deeper a property is nested
the less likely is its relevance for the comparison.
As described in the first step, comparisons are built using shared or
similar properties of resources. In case the same property has been
used between resources, these properties are grouped and form
one comparison row. However, often diferent properties are used
to describe the same concept. This occurs for various reasons. The
most obvious reason is when two diferent ontologies are used to
describe the same property. For example, for describing the
population of a city, DBpedia uses dbo:populationTotal while WikiData
uses WikiData:population (actually the property identifier is P1082;
for the purpose here we use the label). When comparing resources,
these properties should be considered as equivalent. Especially
for community-created knowledge graphs, diferently identified
properties likely exist that are, in fact, equivalent.</p>
      <p>
        To overcome this problem, we use FastText [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] word embeddings
to determine the similarity of properties. If the similarity is higher
than a predetermined threshold τ , the properties are considered
same and are grouped. In the end, each group of predicates will be
visualized as one row in the comparison table. The result of this
step is a list of statements for each comparison resource, where
similar predicates are grouped.
      </p>
      <p>The similarity matrix γ is generated
γ = cos(→p−i , →p−j )i</p>
      <p>h
with cos(.) as the cosine similarity of vector embeddings for
predicate pairs (pi , pj ) ∈ P, whereby P is the set of all resources.</p>
      <p>Furthermore, we create a mask matrix Φ that selects predicates of
resources ci ∈ C, whereby C is the set of resources to be compared.
Formally,
Φi, j =</p>
      <p>Next, for each selected predicate p we create the matrix φ that
slices Φ to include only similar predicates. Formally,
φi, j = (Φi, j )</p>
      <p>
        ci ∈ C
pj ∈sim(p)
where sim(p) is the set of predicates with similarity values γ [p] ≥
τ with predicate p. Finally, φ is used to eficiently compute the
common set of predicates [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. This process is displayed in Algorithm
2.
3.4
      </p>
    </sec>
    <sec id="sec-5">
      <title>Visualize comparison</title>
      <p>The final step of the workflow is to visualize the comparison and
present the data in a human understandable format. Tabular format
is often appropriate for visualizing comparisons. Another aspect
of the visualization is determining which properties should be
(1)
(2)
(3)</p>
      <p>Algorithm 2 Property mapping
1: procedure MapProperties(properties, threshold)
2: for each property p1 ∈ properties do
3: for each property p2 ∈ properties do
4: similarity ← cos(FastText(p1), FastText(p2))
5: if similarity &gt; threshold then
6: similar Props ← similar Props ∪ {p1, p2}
return similar Props
displayed and which ones should be hidden. A property is displayed
when it is shared among a predetermined amount τ of papers, where
τ mainly depends on comparison use and can be determined based
on the total amount of resources in the comparison.</p>
      <p>Another aspect of comparison visualization is the possibility
to customize the resulting table. This is needed because of the
similarity-based matching of properties and the use of
predetermined thresholds. For example, users should be able to enable or
disable properties. They should also get feedback on property
provenance (i.e., the property path). Ultimately, this contributes to a better
user experience, with the possibility to manually correct mistakes
made by the system.
4</p>
    </sec>
    <sec id="sec-6">
      <title>IMPLEMENTATION</title>
      <p>The presented four-step workflow for comparing resources is
implemented in the Open Research Knowledge Graph (ORKG),
specifically to compare research contributions as a special type of resource.
As a companion to the description here, an online video summarizes
and demonstrates the ORKG comparison feature11.</p>
      <p>The user interface of the comparison feature is seamlessly
integrated with the ORKG front end, which is written in JavaScript
using the React framework12 and is publicly available online13. The
back end of the comparison feature is written as a service separate
from the ORKG back end, is written in Python and is available
online14.</p>
      <p>In the ORKG, each paper consists of at least one research
contribution which addresses at least one research problem and is further
described with contribution data including for instance materials,
methods, implementation, results or other aspects. In the ORKG, it
is research contributions that are compared rather than papers.</p>
      <p>We will now discuss each step of the presented workflow to
illustrate how it is implemented in the ORKG.
4.1</p>
    </sec>
    <sec id="sec-7">
      <title>Select comparison candidates</title>
      <p>Both approaches presented, namely find similar resources and
manual selection, are implemented in the ORKG. The reason for
implementing both is that they complement each other. Conducting
a comparison based on similarity is useful when a user wants to
compare a certain contribution with other (automatically
determined similar) contributions (for example, addressing the same
problem), while manual contribution selection can be helpful to
compare a user defined set of contributions. Figure 2 shows both
approaches. As depicted, three similar contributions are suggested
11https://youtu.be/mbe-cVyW_us
12https://reactjs.org/
13https://gitlab.com/TIBHannover/orkg/orkg-frontend
14https://gitlab.com/TIBHannover/orkg/orkg-similarity
to the user. (The corresponding similarity percentage is displayed
next to paper title.) These suggested contributions can be directly
compared. In contrast, the manual approach works similarly to an
online shopping cart. When the “Add to comparison” checkbox is
checked, a box is displayed at the bottom of the page. This box
shows the manually selected contributions that will be used for the
comparison (Figure 3).</p>
      <p>To retrieve contributions that are similar to a given contribution,
we developed an API endpoint. This endpoint takes the given
contribution as input and returns five similar contributions (of which
three are displayed). For performance reasons, each contribution is
indexed by concatenating the properties to a string (Section 3.1).
This string is stored inside a document-oriented database. The
indexing happens as soon as a contribution is added. The result of this
step is a set of contribution IDs used to perform the comparison.
4.2</p>
    </sec>
    <sec id="sec-8">
      <title>Select related statements</title>
      <p>An additional API endpoint was developed for the comparison. This
endpoint takes the set of contribution IDs as input and returns the
data used to display the comparison. The comparison endpoint
is responsible for steps two and three of the workflow: selecting
the related statements and mapping the properties. For each listed
contribution, an ORKG back end query selects all related statements.
This is done as described in Section 3.2. The process of selecting
statements is repeated until depth δ = 5. This number is chosen to
include statements that are not directly related to the resource, but
to exclude statements that are less relevant because they are nested
too deep.
4.3</p>
    </sec>
    <sec id="sec-9">
      <title>Map properties</title>
      <p>Using the API of the previous step, the properties of the selected
statements are mapped. As described in the workflow, for each
property pair the similarity is calculated using word embeddings. In
case the similarity threshold τ ≥ 0.9, the properties are considered
to be equivalent and are grouped. The threshold is determined by a
trial and error method. Then the results from the API are returned
to the UI where they are displayed.
Because the comparisons are made for humans, visualizing them
efectively is essential and therefore we invested considerable efort
on this aspect. Figure 4 displays a comparison for research
contributions related to visualization tools published in the literature. In
this example, four properties are displayed. Literals are displayed as
plain text while resources are displayed as links. When a resource
link is selected, a popup is displayed showing the statements related
to this resource. By default, only properties that are common to at
least two contributions (τ ≥ 2) are displayed. The UI implements
some additional features that are particularly useful to compare
research contributions. We will now discuss these features is more
detail.</p>
      <p>4.4.1 Customization. Users can customize comparisons
according to their needs. The customization includes transposing the
table and customizing the properties. The properties can be
enabled/disabled and they can be sorted. Especially the option to
disable properties is helpful when resources with many statements
are compared. Only properties considered relevant to the user can
be selected to display. Customizing the comparison table can be
useful before exporting or sharing the comparison.</p>
      <p>4.4.2 Sharing and persistence. The comparison can be shared
using a link. For sharing the comparison, a persistence mechanism
has been built in. Especially when sharing the comparison for
research purposes, it is important to share the original comparison.
Since resource descriptions may change over time comparisons
may also change. To support persistency, the whole state of the
comparison is stored in a document-based database.</p>
      <p>4.4.3 Export. It is possible to export comparisons in formats
PDF, CSV and LATEX. Especially the LATEX export is useful for the
ORKG, since the export be directly used in research papers. In
addition to the generated LATEX table, a BibTeX file is generated
containing the bibliographic information of the papers used in the
comparison. Also, a link referring back to the comparison by the
ORKG is showed as footnote. Just like the shareable link, this link
is persistent and is therefore suitable for use in articles.</p>
    </sec>
    <sec id="sec-10">
      <title>PRELIMINARY EVALUATION</title>
      <p>In this section we present a preliminary evaluation of the
implemented comparison functionality.
5.1</p>
    </sec>
    <sec id="sec-11">
      <title>User evaluation</title>
      <p>
        A qualitative evaluation is conducted to determine the usability of
our implementation. Additionally, the evaluation is used to
determine the usefulness of the comparison functionality in general. The
usability is determined using the System Usability Scale (SUS) [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. In
total, five participants were part of the evaluation. All participants
are researchers. At the start of the evaluation, each participant was
asked to watch a video that explained the basic concepts of the
comparison functionality. Afterwards an instructor asked the
participant to perform certain tasks in the system, specifically creating
a comparison (based on similarity and manually), customizing this
comparison and exporting the comparison. The tasks were chosen
to include all main functionalities of the comparison
functionality. In case a participant was not able to complete the task he was
allowed to ask an instructor for help. After interacting with the
system, users were asked to fill out an online questionnaire 15. The
questionnaire contained ten questions from the SUS, each questions
could be answered on a scale from 1 (strong disagree) to 5 (strongly
agree). Afterwards, a short interview was conducted to get the
opinions of the participants on the usefulness of the comparison
feature.
      </p>
      <p>
        The SUS score ranges from 1 to 100. In our evaluation, the SUS
score is 81, which is considered excellent [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. Figure 5 depicts the
score per question. This indicates that participants did not have
problems with using the user interface to create, customize and
export their related work comparisons. This is in line with the
positive verbal feedback that was provided to the instructor during the
evaluation. In addition to the usability questions, three questions
were asked related to the usefulness of the related literature
comparison functionality. All participants agreed that such a functionality
is useful and can potentially save them time while conducting
research. Finally, participants were asked to give additional feedback.
Among others, participant #1 remarked "It would be nice if it is
explained how similarity of papers is determined"; participant #3
suggested "Show text labels next to properties, explaining what this
15https://forms.gle/x2t7SYkAzCkCekUp8
In order to evaluate the performance of the overall comparison, we
compared the implemented ORKG approach to a baseline approach
for comparing multiple resources. In Table 1 the time needed for
a comparison is displayed for both the baseline and the ORKG
approach. In total eight papers are compared with on average ten
properties per paper. In the baseline approach, the “Map properties”
step is not scaling well. This is because each property is compared
against all other properties. If multiple contributions are selected,
the amount of property similarity checks grows exponentially. As
displayed in the table, the ORKG approach outperforms the baseline
approach. The total amount of papers used for the evaluation is
limited to eight because the baseline approach does not scale to
larger sets.
6
      </p>
    </sec>
    <sec id="sec-12">
      <title>DISCUSSION &amp; FUTURE WORK</title>
      <p>The aim of the contribution comparison functionality is to support
literature reviews and make it more eficient. To live up to this aim,
the knowledge graph should contain more data. As described in
Section 2, structured data is needed to perform an efective and accurate
comparison. Currently, such a graph containing research
contributions does not exist since most existing initiatives focus solely
on document metadata. This is why the ORKG focuses on making
the actual research contributions machine-readable. Although the
amount of papers in the ORKG is growing, it is currently not
suficient for the comparison functionality to be used efectively. The
evaluation results suggest that the comparison feature performs
well and that users are satisfied with the usability. Additionally,
they see the potential of the functionality. Thus, the technical
infrastructure is in place for the related literature comparison but
more data is needed for an extensive evaluation and real-world use.</p>
      <p>In order to evaluate the usability of the interface, a user
evaluation is arguably the most suitable method. In total there were
only five participants for the user evaluation presented here. While
this is not suficient to make any definitive conclusions, it helps to
understand what users expect from such a system. The individually
provided feedback is also helpful to guide further developments.
One of the important outcomes of the evaluation is that all
participants agreed on the usefulness of the feature. They saw the
potential of conducting literature reviews with the ORKG instead
of doing it entirely manually.</p>
      <p>Future work will focus on a more extensive evaluation of the
individual components of the system. This includes the merging
of properties and the similarity functionality. In order to perform
such an evaluation, more data should be added to the ORKG. Given
that automated related literature comparisons is one of the many
advantages of structured scholarly knowledge, more functionalities
leveraging this structured data will be developed. An example is
faceted search, which provides an alternative to the full-text search
commonly used to find related literature.
7</p>
    </sec>
    <sec id="sec-13">
      <title>CONCLUSION</title>
      <p>The presented workflow shows how research contributions in a
graph database can be compared, which answers our first research
question. The workflow consists of four steps in which comparison
candidates are selected, related statements are fetched, properties
are mapped and finally the comparison is visualized. We presented,
evaluated and discussed an implementation of the workflow in the
ORKG. The implementation answers our second research question
by showing how the comparisons can be efectively visualized in a
user interface. The performance evaluation results show that the
system scales well. The user evaluation indicates that users see the
potential of a related literature comparison functionality, and that
the current implementation is user-friendly.</p>
    </sec>
    <sec id="sec-14">
      <title>ACKNOWLEDGMENTS</title>
      <p>This work was co-funded by the European Research Council for
the project ScienceGRAPH (Grant agreement ID: 819536) and the
TIB Leibniz Information Centre for Science and Technology. The
authors would like to thank the participants of the user evaluation.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>Samur</given-names>
            <surname>Araujo</surname>
          </string-name>
          , Jan Hidders, Daniel Schwabe, and
          <string-name>
            <surname>Arjen P. De Vries</surname>
          </string-name>
          .
          <year>2011</year>
          .
          <article-title>SERIMI - Resource description similarity, RDF instance matching and interlinking</article-title>
          .
          <source>CEUR Workshop Proceedings</source>
          <volume>814</volume>
          (
          <year>2011</year>
          ),
          <fpage>246</fpage>
          -
          <lpage>247</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>Piotr</given-names>
            <surname>Bojanowski</surname>
          </string-name>
          , Edouard Grave, Armand Joulin, and
          <string-name>
            <given-names>Tomas</given-names>
            <surname>Mikolov</surname>
          </string-name>
          .
          <year>2017</year>
          .
          <article-title>Enriching Word Vectors with Subword Information. Transactions of the Association for Computational Linguistics 5 (</article-title>
          <year>2017</year>
          ),
          <fpage>135</fpage>
          -
          <lpage>146</lpage>
          . https://doi.org/10.1162/tacl_a_
          <fpage>00051</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>John</given-names>
            <surname>Brooke</surname>
          </string-name>
          et al.
          <year>1996</year>
          .
          <article-title>SUS-A quick and dirty usability scale</article-title>
          .
          <source>Usability evaluation in industry 189</source>
          ,
          <issue>194</issue>
          (
          <year>1996</year>
          ),
          <fpage>4</fpage>
          -
          <lpage>7</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>Alexandru</given-names>
            <surname>Constantin</surname>
          </string-name>
          , Silvio Peroni, Steve Pettifer, David Shotton,
          <string-name>
            <given-names>and Fabio</given-names>
            <surname>Vitali</surname>
          </string-name>
          .
          <year>2016</year>
          .
          <article-title>The Document Components Ontology (DoCO)</article-title>
          .
          <source>Semantic Web</source>
          <volume>7</volume>
          ,
          <issue>2</issue>
          (
          <year>2016</year>
          ),
          <fpage>167</fpage>
          -
          <lpage>181</lpage>
          . https://doi.org/10.3233/SW-150177
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>Tim</given-names>
            <surname>Donovan</surname>
          </string-name>
          ,
          <string-name>
            <surname>Lambert M. Felix</surname>
            ,
            <given-names>James D.</given-names>
          </string-name>
          <string-name>
            <surname>Chalmers</surname>
            ,
            <given-names>Stephen J.</given-names>
          </string-name>
          <string-name>
            <surname>Milan</surname>
            , Alexander G. Mathioudakis, and
            <given-names>Sally</given-names>
          </string-name>
          <string-name>
            <surname>Spencer</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>Continuous versus intermittent antibiotics for bronchiectasis</article-title>
          .
          <source>Cochrane Database of Systematic Reviews</source>
          <year>2018</year>
          ,
          <volume>6</volume>
          (
          <year>2018</year>
          ),
          <fpage>114</fpage>
          -
          <lpage>123</lpage>
          . https://doi.org/10.1002/14651858.CD012733.pub2
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>Aldo</given-names>
            <surname>Gangemi</surname>
          </string-name>
          , Silvio Peroni, David Shotton,
          <string-name>
            <given-names>and Fabio</given-names>
            <surname>Vitali</surname>
          </string-name>
          .
          <year>2017</year>
          .
          <article-title>The Publishing Workflow Ontology (PWO)</article-title>
          .
          <source>Semantic Web</source>
          <volume>8</volume>
          ,
          <issue>5</issue>
          (
          <year>2017</year>
          ),
          <fpage>703</fpage>
          -
          <lpage>718</lpage>
          . https://doi.org/10.3233/SW-160230
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>Dagmar</given-names>
            <surname>Gromann</surname>
          </string-name>
          and
          <string-name>
            <given-names>Thierry</given-names>
            <surname>Declerck</surname>
          </string-name>
          .
          <year>2019</year>
          .
          <article-title>Comparing pretrained multilingual word embeddings on an ontology alignment task</article-title>
          .
          <source>LREC 2018 - 11th International Conference on Language Resources and Evaluation</source>
          (
          <year>2019</year>
          ),
          <fpage>230</fpage>
          -
          <lpage>236</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>Alexander</given-names>
            <surname>Hars</surname>
          </string-name>
          .
          <year>2001</year>
          .
          <article-title>Designing Scientific Knowledge Infrastructures: The Contribution of Epistemology</article-title>
          .
          <source>Information Systems Frontiers</source>
          <volume>3</volume>
          ,
          <issue>1</issue>
          (
          <year>2001</year>
          ),
          <fpage>63</fpage>
          -
          <lpage>73</lpage>
          . https://doi.org/10.1023/A:1011401704862
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>Mohamad</given-names>
            <surname>Yaser</surname>
          </string-name>
          <string-name>
            <given-names>Jaradeh</given-names>
            , Allard Oelen, Manuel Prinz,
            <surname>Jennifer D'Souza</surname>
          </string-name>
          , Gábor Kismihók, Markus Stocker, and
          <string-name>
            <given-names>Sören</given-names>
            <surname>Auer</surname>
          </string-name>
          .
          <year>2019</year>
          .
          <article-title>Open Research Knowledge Graph: Next Generation Infrastructure for Semantic Scholarly Knowledge (in press)</article-title>
          .
          <source>In In Proceedings of the 10th International Conference on Knowledge Capture (K-CAP '19)</source>
          . ACM. https://doi.org/10.1145/3360901.3364435
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>Vladimir</surname>
            <given-names>I</given-names>
          </string-name>
          <string-name>
            <surname>Levenshtein</surname>
          </string-name>
          .
          <year>1966</year>
          .
          <article-title>Binary codes capable of correcting deletions, insertions, and reversals</article-title>
          .
          <source>In Soviet physics doklady</source>
          , Vol.
          <volume>10</volume>
          .
          <fpage>707</fpage>
          -
          <lpage>710</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>Pierre</surname>
            <given-names>Maillot</given-names>
          </string-name>
          , Carlos Bobed, Pierre Maillot, Carlos Bobed, Pierre Maillot, and
          <string-name>
            <given-names>Carlos</given-names>
            <surname>Bobed</surname>
          </string-name>
          .
          <year>2019</year>
          .
          <article-title>Measuring structural similarity between RDF graphs To cite this version : HAL Id : hal-01940449 Measuring Structural Similarity Between RDF Graphs</article-title>
          .
          <article-title>(</article-title>
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          <source>[12] Carme Pinya Medina and Maria Rosa Rosselló Ramon</source>
          .
          <year>2015</year>
          .
          <article-title>Using TF-IDF to Determine Word Relevance in Document Queries Juan</article-title>
          .
          <source>New Educational Review</source>
          <volume>42</volume>
          ,
          <issue>4</issue>
          (
          <year>2015</year>
          ),
          <fpage>40</fpage>
          -
          <lpage>51</lpage>
          . https://doi.org/10.15804/tner.
          <year>2015</year>
          .
          <volume>42</volume>
          .4.
          <fpage>03</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>Silvio</given-names>
            <surname>Peroni</surname>
          </string-name>
          and
          <string-name>
            <given-names>David</given-names>
            <surname>Shotton</surname>
          </string-name>
          .
          <year>2012</year>
          .
          <article-title>FaBiO and CiTO: Ontologies for describing bibliographic resources and citations</article-title>
          .
          <source>Journal of Web Semantics</source>
          <volume>17</volume>
          (
          <year>2012</year>
          ),
          <fpage>33</fpage>
          -
          <lpage>43</lpage>
          . https://doi.org/10.1016/j.websem.
          <year>2012</year>
          .
          <volume>08</volume>
          .001
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>Silvio</given-names>
            <surname>Peroni</surname>
          </string-name>
          and
          <string-name>
            <given-names>David</given-names>
            <surname>Shotton</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>The SPAR ontologies</article-title>
          .
          <source>In International Semantic Web Conference</source>
          . Springer,
          <fpage>119</fpage>
          -
          <lpage>136</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <surname>Alina</surname>
            <given-names>Petrova</given-names>
          </string-name>
          , Evgeny Sherkhonov, Bernardo Cuenca Grau, and
          <string-name>
            <given-names>Ian</given-names>
            <surname>Horrocks</surname>
          </string-name>
          .
          <year>2017</year>
          .
          <article-title>Entity comparison in RDF graphs</article-title>
          .
          <source>Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) 10587 LNCS</source>
          (
          <year>2017</year>
          ),
          <fpage>526</fpage>
          -
          <lpage>541</lpage>
          . https://doi.org/10.1007/978-3-
          <fpage>319</fpage>
          -68288-4_
          <fpage>31</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>Almudena</given-names>
            <surname>Ruiz</surname>
          </string-name>
          Iniesta and
          <string-name>
            <given-names>Oscar</given-names>
            <surname>Corcho</surname>
          </string-name>
          .
          <year>2014</year>
          .
          <article-title>A review of ontologies for describing scholarly and scientific documents</article-title>
          .
          <source>In 4th Workshop on Semantic Publishing (SePublica) (CEUR Workshop Proceedings)</source>
          . http://ceur-ws.
          <source>org/</source>
          Vol-
          <volume>1155</volume>
          #
          <fpage>paper</fpage>
          -
          <lpage>07</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>Pavel</given-names>
            <surname>Shvaiko</surname>
          </string-name>
          and
          <string-name>
            <given-names>Jérôme</given-names>
            <surname>Euzenat</surname>
          </string-name>
          .
          <year>2013</year>
          .
          <article-title>Ontology matching: State of the art and future challenges</article-title>
          .
          <source>IEEE Transactions on Knowledge and Data Engineering</source>
          <volume>25</volume>
          ,
          <issue>1</issue>
          (
          <year>2013</year>
          ),
          <fpage>158</fpage>
          -
          <lpage>176</lpage>
          . https://doi.org/10.1109/TKDE.
          <year>2011</year>
          .253
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <surname>William</surname>
            <given-names>E.</given-names>
          </string-name>
          <string-name>
            <surname>Winkler</surname>
          </string-name>
          .
          <year>1990</year>
          .
          <article-title>String Comparator Metrics and Enhanced Decision Rules in the Fellegi-Sunter Model of Record Linkage</article-title>
          .
          <source>Proceedings of the Section on Survey Research</source>
          , American Statistical Association (
          <year>1990</year>
          ),
          <fpage>354</fpage>
          -
          <lpage>359</lpage>
          . https://doi. org/10.1007/978-1-
          <fpage>4612</fpage>
          -2856-1_
          <fpage>101</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <surname>Paweł</surname>
            <given-names>Ziemba</given-names>
          </string-name>
          , Jarosław Jankowski, and
          <string-name>
            <given-names>Jarosław</given-names>
            <surname>Wątróbski</surname>
          </string-name>
          .
          <year>2017</year>
          .
          <article-title>Online comparison system with certain and uncertain criteria based on multi-criteria decision analysis method</article-title>
          .
          <source>In International Conference on Computational Collective Intelligence</source>
          . Springer,
          <fpage>579</fpage>
          -
          <lpage>589</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>