=Paper= {{Paper |id=Vol-2699/paper06 |storemode=property |title=Symbolic Vs Sub-symbolic AI Methods: Friends or Enemies? |pdfUrl=https://ceur-ws.org/Vol-2699/paper06.pdf |volume=Vol-2699 |authors=Eleni Ilkou,Maria Koutraki |dblpUrl=https://dblp.org/rec/conf/cikm/IlkouK20 }} ==Symbolic Vs Sub-symbolic AI Methods: Friends or Enemies?== https://ceur-ws.org/Vol-2699/paper06.pdf
Symbolic Vs Sub-symbolic AI Methods: Friends or
Enemies?
Eleni Ilkoua,b , Maria Koutrakia,b
a
    L3S Research Center, Appelstrasse 9a, 30167 Hannover, Germany
b
    Leibniz University of Hannover, Welfengarten 1, 30167 Hannover, Germany


                                             Abstract
                                             There is a long and unresolved debate between the symbolic and sub-symbolic methods. However, in recent years, there
                                             is a push towards in-between methods. In this work, we provide a comprehensive overview of the symbolic, sub-symbolic
                                             and in-between approaches focused in the domain of knowledge graphs, namely, schema representation, schema matching,
                                             knowledge graph completion, link prediction, entity resolution, entity classification and triple classification. We critically present
                                             key characteristics, advantages and disadvantages of the main algorithms in each domain, and review the use of these
                                             methods in knowledge graph related applications.

                                             Keywords
                                             Symbolic methods, sub-symbolic methods, in-between methods, knowledge graph tasks, knowledge graph completion,
                                             schema


1. Introduction                                                                                                       combination of symbolic and sub-symbolic AI approaches,
                                                                                                                      which we refer to as in-between methods.
Symbolic and sub-symbolic represent the two main bran-                                                                   Table 1 shows an overview of some of the basic differ-
ches of Artificial Intelligence (AI). The AI field saw huge                                                           ent characteristics of the symbolic and sub-symbolic AI
progress and established itself in the 1950s, after some of                                                           methods. It presents an easy visual comparison between
the most notable and inaugural works of McCulloch and                                                                 the two AI fields; as it was discussed in [1, 2] and accord-
Pittes, who in 1943 established the foundations of neural                                                             ing to our thorough analysis of the fields. Apart from the
networks (NN), and Turing’s work, who introduced in                                                                   core symbolic or sub-symbolic methods, nowadays there
1950s the test of intelligence for machines, known as the                                                             are symbolic applications with sub-symbolic characteris-
Turing test.                                                                                                          tics and vice versa [3]. We choose to adopt an annotation
   Since its invention, the field has seen ups and downs                                                              where a method belongs to symbolic or sub-symbolic if
in its development, which are colloquially known as the                                                               it uses only symbolic or sub-symbolic parts respectively;
AI seasons, and are characterised as “summers” and “win-                                                              otherwise we categorise it in the in-between methods.
ters”. The exact periods of these ups and downs are                                                                   The main differences between these two AI fields are
unclear, however, we adopt an intermediate convention                                                                 the following: (1) symbolic approaches produce logical
based on Wikipedia and Henry Kautz’s talk1 “The Third                                                                 conclusions, whereas sub-symbolic approaches provide
AI Summer” in AAAI 2020. We display a timeline of these                                                               associative results. (2) The human intervention is com-
developments in Figure 1.                                                                                             mon in the symbolic methods, while the sub-symbolic
   The first AI summer, also called the golden years, be-                                                             learn and adapt to the given data. (3) The symbolic meth-
gins a few years after the birth of AI, and it was based                                                              ods perform best when dealing with relatively small and
on the optimism in problem solving and reasoning. The                                                                 precise data, while the sub-symbolic ones are able to
dominant paradigm was symbolic AI until the 1980s. This                                                               handle large and noisy datasets.
is when the sub-symbolic AI starts taking the lead and                                                                   In this paper, we discuss in detail some of the well-
gains attention until the recent years. There is a long and                                                           known approaches in each AI domain, and their appli-
unresolved debate between the two different approaches.                                                               cation use-cases in some of the most prominent down-
However, this grapple between the different AI domains                                                                stream tasks in the domain of knowledge graphs. We
is approaching to its end, as we are currently experienc-                                                             focus on their applicability in the schema representa-
ing the third AI summer, where the presiding wave is the                                                              tion, schema matching, knowledge graph completion
                                                                                                                      and more specifically in entity resolution, link prediction,
Proceedings of the CIKM 2020 Workshops, October 19-20, Galway,                                                        entity and triple classification. In this work, we make the
Ireland.                                                                                                              following contributions:
" ilkou@l3s.de (E. Ilkou); koutraki@l3s.de (M. Koutraki)
 0000-0002-4847-6177 (E. Ilkou)                                                                                           • A overview of the characteristics, advantages and
                                       © 2020 Copyright for this paper by its authors. Use permitted under Creative
                                       Commons License Attribution 4.0 International (CC BY 4.0).                            disadvantages of the symbolic and sub-symbolic
    CEUR

                          CEUR Workshop Proceedings (CEUR-WS.org)
                                                                                                                             AI methods (Sections 2 and 3).
                  http://ceur-ws.org
    Workshop      ISSN 1613-0073
    Proceedings


                  1
                      https://roc-hci.com/announcements/the-third-ai-summer/
                                                Golden years:
                                                 1st summer             The boom:                                The 3rd
                            Turing                                     2nd summer                                summer
                             Test




             1940           1950              1960         1970      1980             1990    2000        2010       2020
              Foundations
                 of NN             Birth of                             Start of
                                      AI                          in-between models


Figure 1: The timeline of Artificial Intelligence methods



     • An analysis of the in-between methods and their                      rule based systems have the advantage of rule modular-
       different categories as they are presented in the                    ity, as the rules are discrete and autonomous knowledge
       literature, and their general characteristics (Sec-                  units that can easily be inserted or removed from a knowl-
       tion 4).                                                             edge base [6]. Moreover, they provide knowledge inter-
     • An overview of the most common applications                          operability; meaning that in closely related applications,
       of the symbolic, sub-symbolic and in-between                         knowledge transfer is possible. Also, they are better for
       methods in knowledge graphs (Section 5).                             abstract problems as they are not highly dependent on
                                                                            the input data.
   The rest of this paper is structured as following: Sec-                     On the other hand, the symbolic methods are typi-
tion 2 presents an overview of the main characteristics of                  cally not well-suitable for cases where datasets have data-
the symbolic AI methods. Similarly, in Section 3 we dis-                    quality issues and might be prone to noise. Under such
cuss the main characteristics of the sub-symbolic meth-                     circumstances, they are often yielding to sub-optimal
ods. In Section 4, we present an overview of the ap-                        results [5], and they are not possible to conclude (“brittle-
proaches that combine both symbolic and sub-symbolic                        ness”) [7]. Further, the rules and the knowledge usually
methods, namely, the in-between methods. Then, in                           are hard and hand-coded, creating the Knowledge Ac-
Section 5 we present some of the most important down-                       quisition Bottleneck [8], which refers to the high cost of
stream tasks in the field of knowledge graphs and we                        human involvement in converting real-world problems
analyse the different approaches (symbolic, sub-symbolic                    into inputs for symbolic AI systems. Finally, the main-
and in-between methods) that have been followed in the                      tenance of rule bases is difficult as it requires complex
literature to tackle these tasks.                                           verification and validation.
                                                                               In terms of applications, the symbolic methods work
2. Symbolic Methods                                                         best on well-defined and static problems, and on manipu-
                                                                            lating and modelling abstractions. However, traditionally,
Symbolic methods, also known as Good Old Fashioned                          they do not have good performance in real-time dynamic
Artificial Intelligence (GOFAI), refer to human-readable                    assessments and massive empirical data streams.
and explainable processes. The symbolic techniques are
defined by explicit symbolic methods, such as formal
methods and programming languages, and are usually
                                                                            3. Sub-symbolic Methods
used for deductive knowledge [4]. They consist of first-                    Contrary to symbolic methods, where the learning hap-
order logic rules, while other methods include rules, on-                   pens through the human supervision and intervention,
tologies, decision trees, planning and reasoning. Accord-                   sub-symbolic methods establish correlations between
ing to Benderskaya et al [5] the symbolic AI is usually                     input and output variables. Such relations have high
associated with knowledge bases and expert systems,                         complexity, and are often formalized by functions that
and it is a continuation of the von Neumann and Turing                      map the input to the output data or the target variables.
machines.                                                                      Sub-symbolic methods represent the Connectionism
   A key characteristic of symbolic methods is their abil-                  movement that is trying to mimic a human brain and
ity to explain and reason about the reached conclusion.                     its complex network of interconnected neurons with the
Furthermore, even their intermediate steps are often                        Artificial Neural Networks (ANN). The sub-symbolic AI
explainable. The symbolic systems provide a human-                          includes statistical learning methods, such as Bayesian
understandable computation flow which makes them                            learning, deep learning, backpropagation, and genetic
easier to debug, explain and control. In particular, the                    algorithms.
Table 1                                                        4. In-between Methods
Symbolic vs Sub-symbolic methods characteristics
                                                               Despite the fundamental differences between symbolic
 Symbolic                       Sub-symbolic                   and sub-symbolic the last years there is a link between
 Symbols                        Numbers                        them with the in-between methods. Since late 1980s,
 Logical                        Associative                    there is a discussion about the need of cognitive sub-
 Serial                         Parallel                       symbolic level [11]. The in-between methods consist of
 Reasoning                      Learning                       the efforts to bridge the gap between the symbolic and
 von Neumann machines           Dynamic Systems                sub-symbolic paradigms. The idea is to create a system
 Localised                      Distributed                    which can combine the advantages of both methods: the
 Rigid and static               Flexible and adaptive
                                                               ability to learn from the environment and the ability to
 Concept composition and        Concept creation, and
 expansion                      generalization                 reason the results.
 Model abstraction              Fitting to data                   Most of the recent applications use a combination of
 Human intervention             Learning from data             symbolic and connectionist parts to create their algo-
 Small data                     Big data                       rithms. The used terminology for the range between the
 Literal/precise input          Noisy/incomplete input         symbolic and sub-symbolic varies, as can be seen in this
                                                               Section many methods are found with different names.
                                                               Therefore, we refer to them as in-between methods.
   The sub-symbolic methods are more robust against
noisy and missing data, and generally have high comput-        4.1. General characteristics
ing performance. They are easier to scale up, therefore,
they are well suitable for big datasets and large knowl-       The advantages of the in-between computations are evi-
edge graphs. Moreover, they are better for perceptual          dent and measurable to specific applications, with higher
problems, and they require less knowledge upfront.             accuracy, efficiency and knowledge readability [12]. They
   However, connectionist methods have some disadvan-          have an explanation capacity with no need for a-priori as-
tages. The most important one is the lack of interpretabil-    sumptions, and they are comprehensive cognitive models
ity in these methods. This presents a big obstacle to their    which integrate statistical learning with logical reason-
applicability in domains where explanations and inter-         ing. They also perform well with noisy data [13]. An-
pretations are key points. Further, based on the General       other advantage is that these systems during learning can
Data Protection Regulation of European Union [9], sub-         combine logical rules with data, while fine-tuning the
symbolic techniques are proving to be usually restricted       knowledge based on the input data. Overall, they seem
in critical or high-risk decision applications such as the     suitable for applications which have large amounts of het-
medical, legal or military decision applications and the       erogeneous data and need knowledge descriptions [14].
autonomous cars. Furthermore, they are highly depen-              In the in-between algorithms we find the Knowledge-
dant on the training data they process. At first glance, it    based Neural Networks (KBNN or KBANN) [16], Hybrid
might not seem like a problem, however, this results in an     Expert System (HES) [17], Connectionist Inductive Learn-
inability to extrapolate results to unseen instances or data   ing and Logic Programming (CILP) and Connectionist
which do not follow a similar distribution as the training     Temporal Logics of Knowledge (CTLK) [14], Graph Neu-
data. Additionally, due to the typically large amount of       ral Networks (GNN) [18], Tensor Product Representa-
parameters that need to be estimated in sub-symbolic           tion [19], in which the core is a neural network that is
models, they require huge computation power and huge           loosely-coupled with a symbolic problem-solver. Also,
amounts of data. Another issue arising is the availability     we find the Logic Tensor Networks [20], Neural Tensor
of high quality data for training the algorithms, which        Networks [21] for representing complex logical struc-
often are difficult to find. Data need to be correctly la-     tures, and the latter’s extension are the knowledge graph
belled and to have decent representatives of the normal        translating embedding models [22].
not to lead to biased outcomes [10].                              The applications of these methods can be found in
   Most common applications of sub-symbolic methods            many domains which combine learning and reasoning
include prediction, clustering, pattern classification and     parts according to a specific problem. However, the ex-
recognition of objects, and Natural Language Processing        isting hybrid models are non-generalizable, they cannot
(NLP) tasks. Further, we find in sub-symbolic applica-         be applied in multiple domains; each time the model is
tions the text classification and categorization, as well as   developed to answer a specific question. Also, there is
recognition of speech and text.                                no guide deciding the combinations of symbolic and sub-
                                                               symbolic parts for computation and representation [2].
                                                               Recent downstream applications tend to combine sym-
                                                               bolic and sub-symbolic methods for their computation
Figure 2: The range from symbolic to sub-symbolic as proposed by Hilario [15]



model, more often than using a strictly only one of the        fied approaches, Hilario identifies two main categories
two as can be seen in Section 5.                               the neuronal and the connectionist symbol processing,
                                                               and on the hybrid approaches the translational and func-
4.2. Existing Categorisations in                               tional hybrids respectively. The translational models
                                                               translate representations between NNs and symbolic
     Literature                                                structures. Furthermore, she creates a visual continu-
Some of the in-between methods are found in literature as      ous representation from connectionism to symbolism,
connectionist expert systems (or neural network based          in which includes the categories she is proposing in the
expert systems) [23], multi-agent systems [5], hybrid          range between sub-symbolic and symbolic techniques,
representations [24], neural-fuzzy [25, 26] and neural-        as it is illustrated in Figure 2.
symbolic (or neurosymbolic [15]) computing, learning              In continuation of Hilario’s model, McGarry et al [30]
and reasoning2 , and its sub-type neurules [27].               focuses on hybrid rule-based systems. They are propos-
   To the best of our knowledge, there is no report con-       ing the categorisation of the symbolic rules and neural
taining all the in-between methods. Moreover, there is         networks integrations into unified, transformational and
no standard categorisation or common taxonomy for the          adds the modular subcategory. The latter covers the
methods which belong in the range between the sym-             hybrid models that they consist of several ANNs and
bolic and sub-symbolic techniques. The used terminology        rule-based modules, which are coupled and integrated
varies, therefore, we refer to them as in-between methods,     with many degrees. They support that most of hybrid
and not as neural-symbolic, hybrid or unified. In the last     models are modular.
years, there is an increased interest about in-between
methods [28], and there are some review works in the
domain each one presenting a taxonomy. Most of them            5. Knowledge Graph Tasks
refer to the in-between methods as neural-symbolic ap-         There are plenty of symbolic, sub-symbolic and in-between
proaches.                                                      applications in different domains. Our main focus in this
   Garcez et al [14] present a neural-symbolic comput-         study will be knowledge graph related applications.
ing table that separates the methods to applications of           A knowledge graph (KG) consists of a set of triples
knowledge representation, learning, reasoning and ex-          𝐾 ⊆ 𝐸 × 𝑅 × (𝐸 ∪ 𝐿), where 𝐸 is a set of resources
plainability. Bader and Hitzler [13] study the dimen-          that we refer as entities, 𝐿 a set of literals, and 𝑅 a set
sions of neural-symbolic integration and propose the           of relations. Given a triple (h, r, t) (aka a statement), h is
dimensions of usage, language and interrelation. In the        known as subject, r as relation, and t as object. A KG can
neural-symbolic techniques, they identify two models,          represent any kind of information for the world such as
the hybrid and integrated (also called unified or transla-     (Anna_Karenina, writtenBy, Leo_Tolstoy) and
tional) [29]. The difference between the two is that hybrid    (Leo_Tolstoy, bornIn, Russia), which means that
models combine two or more symbolic and sub-symbolic           Anna Karenina is written by Leo Tolstoy, who was born
techniques which run in parallel, while the integrated         in Russia. The above notation will help us explain and
neural symbolic systems consist of a main sub-symbolic         analyse the following tasks.
component which uses symbolic knowledge in the pro-
cessing.
   Hilario [15] separates neurosymbolic integration to         5.1. Schema Representation
unified and hybrid approaches, each consisted of two           Schemata are present from the beginning of databases
subcategories. Both categories, unified and hybrid, have       and data management systems, and stand in for the
a similar description to Bader and Hitzler [13]. In the uni-   structure of the data and knowledge. In the last years,
    2
                                                               there is attention towards linking and structuring data
        http://www.neural-symbolic.org/
in the web. Connecting information on the web can mapping evolves from specific to generic applications.
also be achieved by schemata; an example is schema.org3 The majority of these methods focus on class alignment,
which focuses on the schema creation, representation and however, there also are works focus on relation align-
maintenance [31]. When modelling knowledge graphs, ment [39, 40, 41].
schemata can be used to prescribe high-level rules that
the graph should follow [4]. Knowledge schema or other- 5.3. Knowledge Graph Completion
wise schema representation contains a conceptual model
of the KG. A schema defines the types of entities and rela- Once a knowledge graph is created, it contains a lot of
tions which can exist in a KG, and an abstract way of com- noisy and incomplete data [42]. In order to fill the missing
bining these entities and relations in (h,r,t) triples. In our information for a constructed knowledge graph, we use
example, a schema representation could exist in the form the task of Knowledge Graph Completion (KGC). KGC,
of triples which will state that (Book, writtenBy, Author)     similar to knowledge graph identification [43], is an in-
and (Author, bornIn, Country) [32].                            telligent way of performing data cleaning. This is usually
   Schema representation is traditionally a symbolic task. addressed with filling the missing edges (link prediction),
First-order logic, ontologies, and formal knowledge repre- deduplicating entity nodes (entity resolution) and dealing
sentation languages, such as RDF(S), OWL [33], XML [34] with missing values.
as well as rules have been used for schema formulations.          Mostly in-between methods are used for KGC, with
Some of the most representative examples of schema rep- the Knowledge Graph Embeddings (KGEs) to be one of
resentation in terms of knowledge graphs construction the most powerful and commonly used techniques. KGEs
are YAGO [35] and DBpedia [36]. The two of the most aim to create a low dimensional vector representation
frequently used KGs are following a symbolic approach of the KG and model relation patterns, hence reduce the
as they are mostly rely on rule mining techniques used complexity of KG related tasks while achieving high ac-
to extract knowledge and represent it in RDF(S) terms.         curacy. We further analyse the KGC task into the specific
                                                               sub-tasks of entity resolution and link prediction.
5.2. Schema Matching
                                                               5.3.1. Entity Resolution
Different KGs use different schemata to represent the
same information which create the need for schema match- Entity resolution (ER) is also known as record linkage,
ing. Schema matching or mapping, also can be found as reference matching or duplication. It is the process of
schema alignment, is happening between two or more finding duplicated references or records in a dataset. It
KGs, when we want to perform data integration or map- is related to data integration as it is one of its founda-
ping, and it refers to the process of identifying semanti- tional problems [44]. Based on our example, we will
cally related objects. It is similar to the entity resolution, have to perform entity resolution between the entities “L.
in Section 5.3.1, with the difference that the latter cares Tolstoy” and “Leo Tolstoy” which can exist in the triples
about mapping object references, such as “L. Tolstoy” and (Anna_Karenina, writtenBy, L._Tolstoy) and (Leo_Tol-
“Leo Tolstoy”, while the schema matching works on the stoy, bornIn, Russia), and refer to the same per-
schema definitions, such as Author and Person.                 son.
   Over the last decades, many models and prototypes             Record linkage was introduced by Halbert L. Dunn in
have been introduced on schema matching. Based on 1946. In 1960s, there are statistical sub-symbolic mod-
a survey in schema matching [37], each model uses an els describing the process of entity resolution, which
input schema with most common an OWL data model, formulate the mathematical basis for many of the cur-
then a RDF, and finally a document type definition. They rent models [45]. In 1990s, machine learning techniques
process the symbolic input by using different models, are applied for this assignment and the techniques used
which can be linguistic or language based, constrain- are mostly based on the in-between methods [46]. Com-
based, and structured-based. The linguistic matchers monly, ER techniques rely on attribute similarity between
combine the symbolic input with sub-symbolic NLP algo- the entities [47]. The algorithms deployed for ER are in-
rithms [38]. The constrain-based matchers are exploiting spired by information retrieval and relational duplicate
the constrains in data features, such as the data types elimination [48].
and ranges. The structured-based matchers focus on the
database/graph structure. Both constrain and structured        5.3.2. Link Prediction
based models use mostly symbolic techniques, while             Link prediction techniques, also known as edge predic-
there is also an interesting raise in combinations of match-   tion, have been applied in many and different fields. Edge
ers (hybrid models). The application domain for schema         prediction refers to the task of adding new links to an ex-
                                                               isting graph. In practice, this can be used as a recommen-
    3
        https://schema.org/
dation system for future connections, or as completion-            6. Conclusions
correction tool that foresees the missing links between
entities.                                                          We represented the symbolic, sub-symbolic and in-between
   The link prediction is a well studied field consisted of        methods in AI and analysed the key characteristics, main
many and different approaches. A survey of link predic-            approaches, advantages and disadvantages of each tech-
tion in complex networks [49] separates the approaches             nique respectively. Further, we argued that the current,
based on the method they are using. It identifies the edge         and possibly future, area of processes is the application
prediction problem to a few techniques that belong to              of the in-between methods. We justified this belief by dis-
sub-symbolic, such as AI based ANN, probabilistic and              cussing principal downstream tasks related to knowledge
Monte Carlo algorithms. However, most of the solutions             graph.
it proposes, belong to the in-between range. The cur-
rent state-of-the-art for link prediction tasks is focused
on in-between methods. KGEs play a big role in this
                                                                   Acknowledgments
task and they can be found in different forms of transla-          This work was partially funded by the European Union’s
tional models [22], neural based KGE with logical rules by         Horizon 2020 research and innovation program under the
[50], and hierarchy-aware KGEs [51]. In link prediction            Marie Skłodowska-Curie grant agreement No 860801.
tasks, we find the triple classification, entity classification,
and the head (?, writtenBy, Leo_Tolstoy), rela-
tion (Anna_Karenina, ?, Leo_Tolstoy), and tail                     References
(Anna_Kareni-
na, writtenBy, ?) prediction respectively. We addi-                 [1] L. R. Medsker, Hybrid neural network and expert
tionally focus on the analysis of entity classification, and            systems, Springer Science & Business Media, 2012.
triple classification in the next paragraphs.                       [2] T. R. Besold, K.-U. Kühnberger, Towards integrated
                                                                        neural–symbolic systems for human-level ai: Two
   Entity Classification. It can also be found as node                  research programs helping to bridge the gaps, Bio-
classification or type prediction in the literature. Entity             logically Inspired Cognitive Architectures 14 (2015)
classification tries to predict the type or the class of an             97–110.
entity given some characteristics. In our case, the input           [3] G. Marcus, Deep learning: A critical appraisal,
triple would be (Leo_Tolsto, isA, ?), which could                       arXiv preprint arXiv:1801.00631 (2018).
give the results (1, Person, 99%) and (2, Author,                   [4] A. Hogan, E. Blomqvist, M. Cochez, C. d’Amato,
98%).                                                                   G. de Melo, C. Gutierrez, J. E. L. Gayo, S. Kirrane,
   While most of link prediction tasks are sub-symbolic                 S. Neumaier, A. Polleres, et al., Knowledge graphs,
based with combination of some symbolic parts, the en-                  arXiv preprint arXiv:2003.02320 (2020).
tity classification is more related to the schema and on-           [5] E. N. Benderskaya, S. V. Zhukova, Multidisciplinary
tology of the KG, hence, the techniques are symbolic                    trends in modern artificial intelligence: Turing’s
based [52, 53].                                                         way, in: Artificial Intelligence, Evolutionary Com-
                                                                        puting and Metaheuristics, Springer, 2013, pp. 319–
   Triple Classification. The triple classification is a                343.
binary problem which answers whether a triple (h,r,t)               [6] I. Hatzilygeroudis, J. Prentzas, Neuro-symbolic ap-
is true or not, for example the input (Anna_Karenina,                   proaches for knowledge representation in expert
writtenBy, Leo_Tolstoy)? leads to result (yes,                          systems, International Journal of Hybrid Intelligent
92%). Systems like the Trans* KGEs [22] use a density                   Systems 1 (2004) 111–126.
function to make predictions about the triple classifi-             [7] D. B. Lenat, M. Prakash, M. Shepherd, Cyc: Using
cation based on a probability function. Mayank Kejri-                   common sense knowledge to overcome brittleness
wal [54] claims that the correct metric for this task with              and knowledge acquisition bottlenecks, AI maga-
the usage of KGEs is accuracy, if the test data are bal-                zine 6 (1985) 65–65.
anced.                                                              [8] J. Cullen, A. Bryman, The knowledge acquisition
   Triple classification algorithms usually belong to in-               bottleneck: time for reassessment?, Expert Systems
between methods, with some examples using neural ten-                   5 (1988) 216–225.
sor networks [21] and time-aware [55], latent factor and            [9] E. Salami, An analysis of the general data protection
semantic matching models.                                               regulation (eu) 2016/679, Available at SSRN 2966210
                                                                        (2017).
                                                                   [10] E. Ntoutsi, P. Fafalios, U. Gadiraju, V. Iosifidis,
                                                                        W. Nejdl, M.-E. Vidal, S. Ruggieri, F. Turini, S. Pa-
                                                                        padopoulos, E. Krasanakis, et al., Bias in data-driven
     artificial intelligence systems—an introductory sur-    [23] S. I. Gallant, S. I. Gallant, Neural network learning
     vey, Wiley Interdisciplinary Reviews: Data Mining            and expert systems, MIT press, 1993.
     and Knowledge Discovery 10 (2020) e1356.                [24] M. Moreno, D. Civitarese, R. Brandao, R. Cerqueira,
[11] M. Frixione, G. Spinelli, S. Gaglio, Symbols and sub-        Effective integration of symbolic and connectionist
     symbols for representing knowledge: a catalogue              approaches through a hybrid representation, arXiv
     raisonne, in: Proceedings of the 11th international          preprint arXiv:1912.08740 (2019).
     joint conference on Artificial intelligence-Volume      [25] R. Fullér, Neural fuzzy systems (1995).
     1, 1989, pp. 3–7.                                       [26] L. Magdalena, A first approach to a taxonomy of
[12] A. d. Garcez, T. R. Besold, L. De Raedt, P. Földiak,         fuzzy-neural systems, Connectionist symbolic inte-
     P. Hitzler, T. Icard, K.-U. Kühnberger, L. C. Lamb,          gration (1997).
     R. Miikkulainen, D. L. Silver, Neural-symbolic learn-   [27] J. Prentzas, I. Hatzilygeroudis, Neurules-a type of
     ing and reasoning: contributions and challenges, in:         neuro-symbolic rules: An overview, in: Combi-
     Proceedings of the AAAI Conference on Artificial             nations of Intelligent Methods and Applications,
     Intelligence, 2015.                                          Springer, 2011, pp. 145–165.
[13] S. Bader, P. Hitzler, Dimensions of neural-symbolic     [28] A. d’Avila Garcez, Proceedings of ijcai international
     integration-a structured survey, arXiv preprint              workshop on neural-symbolic learning and reason-
     cs/0511042 (2005).                                           ing nesy 2005 (2005).
[14] A. d. Garcez, M. Gori, L. C. Lamb, L. Serafini,         [29] R. Sun, F. Alexandre, Connectionist-symbolic inte-
     M. Spranger, S. N. Tran, Neural-symbolic com-                gration: From unified to hybrid approaches, Psy-
     puting: An effective methodology for principled in-          chology Press, 2013.
     tegration of machine learning and reasoning, arXiv      [30] K. McGarry, S. Wermter, J. MacIntyre, Hybrid neu-
     preprint arXiv:1905.06088 (2019).                            ral systems: from simple coupling to fully inte-
[15] M. Hilario, An overview of strategies for neurosym-          grated neural networks, Neural Computing Surveys
     bolic integration, Connectionist-Symbolic Integra-           2 (1999) 62–93.
     tion: From Unified to Hybrid Approaches (1997)          [31] J. Ronallo, Html5 microdata and schema. org,
     13–36.                                                       Code4Lib Journal (2012).
[16] G. Agre, I. Koprinska, Case-based refinement of         [32] C. Belth, X. Zheng, J. Vreeken, D. Koutra, What
     knowledge-based neural networks, in: Proceed-                is normal, what is strange, and what is missing in
     ings of the International Conference" Intelligent            a knowledge graph: Unified characterization via
     Systems: A Semiotic Perspective, volume 2, 1996,             inductive summarization, in: Proceedings of The
     pp. 20–23.                                                   Web Conference 2020, 2020, pp. 1115–1126.
[17] S. Sahin, M. R. Tolun, R. Hassanpour, Hybrid ex-        [33] K. Sengupta, P. Hitzler, Web ontology language
     pert systems: A survey of current approaches and             (owl), Encyclopedia of Social Network Analysis
     applications, Expert systems with applications 39            and Mining (2014).
     (2012) 4609–4617.                                       [34] F. Zhang, L. Yan, Z. M. Ma, J. Cheng, Knowledge
[18] L. Lamb, A. Garcez, M. Gori, M. Prates, P. Avelar,           representation and reasoning of xml with ontology,
     M. Vardi, Graph neural networks meet neural-                 in: Proceedings of the 2011 ACM symposium on
     symbolic computing: A survey and perspective,                applied computing, 2011, pp. 1705–1710.
     arXiv preprint arXiv:2003.00330 (2020).                 [35] F. M. Suchanek, G. Kasneci, G. Weikum, Yago: A
[19] P. Smolensky, Tensor product variable binding and            large ontology from wikipedia and wordnet, Jour-
     the representation of symbolic structures in con-            nal of Web Semantics 6 (2008) 203–217.
     nectionist systems, Artificial intelligence 46 (1990)   [36] J. Lehmann, R. Isele, M. Jakob, A. Jentzsch, D. Kon-
     159–216.                                                     tokostas, P. N. Mendes, S. Hellmann, M. Morsey,
[20] I. Donadello, L. Serafini, A. D. Garcez, Logic tensor        P. van Kleef, S. Auer, C. Bizer, Dbpedia - A large-
     networks for semantic image interpretation, arXiv            scale, multilingual knowledge base extracted from
     preprint arXiv:1705.08968 (2017).                            wikipedia, Semantic Web 6 (2015) 167–195.
[21] R. Socher, D. Chen, C. D. Manning, A. Ng, Rea-          [37] E. Sutanta, R. Wardoyo, K. Mustofa, E. Winarko,
     soning with neural tensor networks for knowledge             Survey: Models and prototypes of schema match-
     base completion, in: Advances in neural informa-             ing., International Journal of Electrical & Computer
     tion processing systems, 2013, pp. 926–934.                  Engineering (2088-8708) 6 (2016).
[22] H. Cai, V. W. Zheng, K. C.-C. Chang, A comprehen-       [38] O. Unal, H. Afsarmanesh, et al., Using linguistic
     sive survey of graph embedding: Problems, tech-              techniques for schema matching., in: International
     niques, and applications, IEEE Transactions on               Conference on Software Technologies ICSOFT (2),
     Knowledge and Data Engineering 30 (2018) 1616–               2006, pp. 115–120.
     1637.                                                   [39] M. Koutraki, N. Preda, D. Vodislav, Online relation
     alignment for linked datasets, in: European Seman- [53] J. Sleeman, T. Finin, Type prediction for efficient
     tic Web Conference, Springer, 2017, pp. 152–168.               coreference resolution in heterogeneous semantic
[40] M. Koutraki, N. Preda, D. Vodislav, SOFYA: se-                 graphs, in: 2013 IEEE Seventh International Con-
     mantic on-the-fly relation alignment, in: Interna-             ference on Semantic Computing, IEEE, 2013, pp.
     tional Conference on Extending Database Technol-               78–85.
     ogy (EDBT), 2016.                                         [54] M. Kejriwal, Advanced topic: Knowledge graph
[41] R. Biswas, M. Koutraki, H. Sack, Exploiting equiva-            completion, in: Domain-Specific Knowledge Graph
     lence to infer type subsumption in linked graphs,              Construction, Springer, 2019, pp. 59–74.
     in: European Semantic Web Conference, Springer, [55] T. Jiang, T. Liu, T. Ge, L. Sha, S. Li, B. Chang, Z. Sui,
     2018, pp. 72–76.                                               Encoding temporal information for time-aware link
[42] R. West, E. Gabrilovich, K. Murphy, S. Sun, R. Gupta,          prediction, in: Proceedings of the 2016 Conference
     D. Lin, Knowledge base completion via search-                  on Empirical Methods in Natural Language Process-
     based question answering, in: Proceedings of the               ing, 2016, pp. 2350–2354.
     23rd international conference on World wide web,
     2014, pp. 515–526.
[43] J. Pujara, H. Miao, L. Getoor, W. Cohen, Knowledge
     graph identification, in: International Semantic
     Web Conference, Springer, 2013, pp. 542–557.
[44] S. Thirumuruganathan, M. Ouzzani, N. Tang, Ex-
     plaining entity resolution predictions: Where are
     we and what needs to be done?, in: Proceedings of
     the Workshop on Human-In-the-Loop Data Analyt-
     ics, 2019, pp. 1–6.
[45] I. P. Fellegi, A. B. Sunter, A theory for record linkage,
     Journal of the American Statistical Association 64
     (1969) 1183–1210.
[46] T. Ebisu, R. Ichise, Graph pattern entity ranking
     model for knowledge graph completion, arXiv
     preprint arXiv:1904.02856 (2019).
[47] V. Christophides, V. Efthymiou, T. Palpanas, G. Pa-
     padakis, K. Stefanidis, End-to-end entity reso-
     lution for big data: A survey, arXiv preprint
     arXiv:1905.06397 (2019).
[48] N. Koudas, S. Sarawagi, D. Srivastava, Record link-
     age: similarity measures and algorithms, in: Pro-
     ceedings of the 2006 ACM SIGMOD international
     conference on Management of data, 2006, pp. 802–
     803.
[49] B. Pandey, P. K. Bhanodia, A. Khamparia, D. K.
     Pandey, A comprehensive survey of edge predic-
     tion in social networks: Techniques, parameters
     and challenges, Expert Systems with Applications
     124 (2019) 164–181.
[50] M. Nayyeri, C. Xu, J. Lehmann, H. S. Yazdi, Logi-
     cenn: A neural based knowledge graphs embed-
     ding model with logical rules, arXiv preprint
     arXiv:1908.07141 (2019).
[51] Z. Zhang, J. Cai, Y. Zhang, J. Wang, Learning
     hierarchy-aware knowledge graph embeddings for
     link prediction., in: Proceedings of the Thirty-
     Fourth AAAI Conference on Artificial Intelligence,
     2020, pp. 3065–3072.
[52] H. Paulheim, C. Bizer, Type inference on noisy rdf
     data, in: International semantic web conference,
     Springer, 2013, pp. 510–525.