<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Using the Meaning of Symbol Names to Guide First-Order Logic Reasoning</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Claudia Schon</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Hochschule Trier</institution>
          ,
          <addr-line>Schneidershof, 54293 Trier</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
      </contrib-group>
      <fpage>19</fpage>
      <lpage>27</lpage>
      <abstract>
        <p>When humans evaluate the validity of a logical conclusion, they typically consider the meaning of the conclusion, taking into account the context, leading to a naturally targeted approach that avoids irrelevant inferences. For example, if humans are asked to show that all hammocks are also beds, they will certainly not draw conclusions about vehicles or weapons. However, automated theorem provers do not usually take the context of the conclusion into account when selecting inference steps. Existing heuristics for selecting the clause for the next inference step usually do not take into account the meaning of symbol names, overlooking this valuable source of information about the context of a clause. Therefore, in the example above, clauses with symbol names such as weapon or vehicle could well be found in the processed clauses. However, since these clauses are not required for the actual proof, they are not helpful to the prover and tend to distract from the actual proof task. In this paper, we present an approach that uses natural language processing techniques to align the selection of the clause for the next inference step with the meaning of the current proof goal. By mapping symbol names to natural language words and representing them as vectors, we compute similarities between symbols and the proof goal, which inform the selection of the clause for the next inference. The approach has been implemented and evaluated. The experimental results show a significant reduction in the number of clauses processed during proof construction using the proposed method.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Commonsense Reasoning</kwd>
        <kwd>Automated Reasoning</kwd>
        <kwd>heuristics for Automated Reasoning</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        When humans evaluate the validity of a logical conclusion,
they naturally consider the context and meaning of the
conclusion and avoid irrelevant inferences. This intuitive
approach is in line with Daniel Kahneman’s model of human
cognition based on two systems [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], where system 1
represents fast, instinctive and contextual thinking, while system
2 involves slower, more deliberate and logical reasoning.
In the field of automated theorem proving (ATP), current
systems [
        <xref ref-type="bibr" rid="ref2 ref3">2, 3</xref>
        ] predominantly use logic-based approaches
similar to system 2, often not using the contextual insights
corresponding to system 1.
      </p>
      <p>For example, a human being asked to prove that all
hammocks are also beds will intuitively avoid considering
unrelated concepts such as vehicles or weapons. Conversely,
automated theorem provers typically do not consider the
contextual relevance of symbol names when selecting
inference steps. Existing heuristics focus on syntactic properties
without exploiting the semantic content embedded in
symbol names, which can lead to the inclusion of irrelevant
clauses and an ineficient proof process.</p>
      <p>It is a key strength of human reasoning that we are able to
combine both intuitive (system 1) and analytical (system 2)
processes. It is therefore plausible that automated theorem
proving could similarly benefit from integrating these two
approaches, using the contextual understanding of system 1
alongside the logical strictness of system 2. Currently,
however, the meaning of symbol names, and thus the contextual
content of formulae, is not taken into account during proof
construction in automated reasoning.</p>
      <p>To address this gap, we propose an approach that
integrates statistical methods from natural language processing
into the reasoning process to improve the selection of the
clause for the next inference step of a theorem prover. We do
this by mapping symbol names to natural language words
and representing them as vectors. This allows us to measure
10th Workshop on Formal and Cognitive Reasoning (FCR-2024) at the 47th
German Conference on Artificial Intelligence (KI-2024, September 23 - 27),
Würzburg, Germany
$ c.schon@hochschule-trier.de (C. Schon)
© 2024 Copyright for this paper by its authors. Use permitted under Creative Commons License
Attribution 4.0 International (CC BY 4.0).
the semantic similarity between symbols and the goal of the
proof. We use these similarities to derive symbol weights,
which guide the selection of the clause for the next inference
step in a more goal-directed way.</p>
      <p>
        Our approach has been implemented and evaluated using
the E theorem prover [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. Experimental results show a
significant reduction in the number of clauses processed
during proof construction, indicating a more focused and
eficient proof process. This work highlights the potential
of combining logic-based and statistical methods to improve
automated reasoning, in line with Kahneman’s theory.
      </p>
      <p>The main contributions of this paper are:
• The introduction of a heuristic for determining
symbol weights based on similarity to the proof goal.
These weights can be used to guide the selection of
the clause for the next inference step in automated
theorem proving.
• An evaluation of the approach, demonstrating a
significant reduction in the number of clauses
processed during proof construction for commonsense
reasoning tasks.</p>
      <p>
        The paper is structured as follows: after discussing related
work (Sect. 2) and preliminaries (Sect. 3), we describe how
to determine symbol weights based on the similarity of the
symbol name to the proof goal in Sect. 4. In Sect. 5 we
present experimental results of our approach. Finally, we
summarize and discuss ideas for future work in Sect. 6.
The selection of the clause for the next inference step has a
very large influence on the performance of theorem provers.
A perfect selection can make a search unnecessary, and a
proof can be found directly. Most heuristics for selecting
the clause for the next inference step are based on simple
counting of symbols and the like [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. In this way, weights are
specified for symbols and then used to determine weights
for clauses, which are used to select the clause for the next
inference step.
      </p>
      <p>
        To make it possible to adjust symbol weights, the theorem
prover E [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] ofers the option of assigning a weight to each
symbol. This can be done by specifying a weight function
with FunWeight1. Although this option allows you to set
symbol weights, it does not describe how to determine them.
      </p>
      <p>
        In [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], the selection of symbol weights is investigated.
The proposed approach employs a graph neural network to
predict symbol weights based on the structure of the clause
normal form of the input problem. The system is trained
using prior proof successes, and a novel scheme is
introduced to balance positive clauses (derived from a proof) and
negative clauses (not derived from a proof) during training.
      </p>
      <p>
        Another approach is to use machine learning to derive
new clause selection strategies from successful proofs [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ].
      </p>
      <p>
        The ENIGMA [
        <xref ref-type="bibr" rid="ref10 ref7 ref8 ref9">7, 8, 9, 10</xref>
        ] method represents a key
advancement in internal guidance within ATPs by integrating
machine learning models directly into the theorem proving
process to influence decisions such as clause selection.
Utilizing techniques like gradient-boosted trees and recursive
neural networks, ENIGMA has significantly enhanced the
performance of saturation-based provers like E, particularly
in large-theory settings like the Mizar Mathematical Library
[
        <xref ref-type="bibr" rid="ref11">11</xref>
        ].
      </p>
      <p>To the best of our knowledge, none of the currently used
approaches to determining symbol weights take into
account the meaning of symbol names.</p>
      <p>
        The problem of premise selection, where relevant
formulae for a given proof task have to be selected from a
knowledge base, is related to the problem considered in
this paper. Especially in the area of large mathematical
libraries, there are many approaches to premise selection
that rely on machine learning methods [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. The ATPBOOST
method improves premise selection by framing it as a
binary classification task and using the XGBoost algorithm.
DeepMath [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] uses neural networks for premise selection.
It is worth noting that DeepMath does not use manually
created features. Another notable technique uses recurrent
neural networks to capture dependencies between premises
[
        <xref ref-type="bibr" rid="ref14">14</xref>
        ]. In addition, the work in [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] applies the transformer
architecture to premise selection, achieving superior
performance compared to earlier recurrent neural network-based
methods. Similarly, the transformer-based Magnushammer
[
        <xref ref-type="bibr" rid="ref16">16</xref>
        ] shows significant improvements in detection success
rates through efective retrieval and ranking of relevant
premises.
      </p>
      <p>
        The Semantic Web [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ] is one domain where the
significance of symbol names has been assessed, revealing that the
semantics embedded in the names of IRIs (Internationalized
Resource Identifiers) reflect a type of social semantics that
aligns with the formal meaning of the referenced resource.
      </p>
      <p>
        Regarding premise selection in commonsense knowledge
bases, there are a few approaches that consider the meaning
of symbol names. One such approach is Similarity SInE [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ],
an extension of SInE selection that uses word embeddings
to account for symbol similarity. Similarity SInE ensures
that all the formulae selected by SInE are included, and
additionally selects formulae for symbols that are similar to
those appearing in the goal. Furthermore, [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ] introduces a
vector-based technique for premise selection which uses a
word embedding to vectorize the formulae of a knowledge
base and to use vector similarity for selection purposes.
The SeVEn [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ] selection takes the idea of vectorizing a
1See the manual of E for details: https://wwwlehre.dhbw-stuttgart.de/
~sschulz/WORK/E_DOWNLOAD/V_2.4/eprover.pdf
19–27
knowledge base one step further by using a large language
model for the vectorization step. These three methods take
into account the meaning of the symbol names, but they are
not used to select the clause for the next inference step.
      </p>
      <p>We are not aware of any approaches that use the
meaning of symbols to select the clause to be used in the next
reasoning step.</p>
    </sec>
    <sec id="sec-2">
      <title>3. Preliminaries</title>
      <p>We consider the following first-order logic reasoning task
in this paper: given of a set of formulae called the
knowledge base (KB) and a goal . The task is to show that the
KB implies  which can be reduced to showing that the
conjunction of KB and ¬ is unsatisfiable. To show the
unsatisfiability, usually the conjunction of the KB and the
negated goal are transformed into an equisatisfiable clause
normal form and a contradiction is inferred from the clause
set.</p>
      <p>To achieve this contradiction, saturation-based provers
organise the execution of their inferences using a given-clause
algorithm as shown in Alg. 1. This algorithm maintains two
sets of clauses: The set of processed clauses and the set of
unprocessed clauses. Initially, all clauses are added to the
set of unprocessed clauses, and the set of processed clauses
is empty. In each step, the algorithm selects a so-called
given clause from the set of unprocessed clauses, adds it
to the set of processed clauses, and then performs all the
inferences possible with the given clause and the set of
processed clauses. The resulting new clauses are added to the
set of unprocessed clauses. This process is repeated until
either the empty clause is inferred, or the set of unprocessed
clauses is empty (or a resource or time limit is reached).</p>
      <sec id="sec-2-1">
        <title>Algorithm 1 Given Clause Algorithm</title>
        <p>Require: Initial set of clauses initial_clauses
1: processed_set ← {}
2: unprocessed_set ← initial_clauses
3: while True do
4: if empty_clause ∈ processed_set then
5: return “Proof Found”
6: end if
7: if unprocessed_set is empty then
8: return “No Proof Found”
9: end if
10: given_clause ← select_clause(unprocessed_set)
11: remove given_clause from unprocessed_set
12: add given_clause to processed_set
13: new_clauses ← generate_new_clauses(given_clause,
processed_set)
14: simplified_clauses ← simplify_clauses(new_clauses)
15: for each clause in simplified_clauses do
16: if clause ∈/ unprocessed_set then
17: add clause to unprocessed_set
18: end if
19: end for
20: end while</p>
        <p>The choice of the given clause (line 10 in Alg. 1) strongly
influences whether and how fast a proof is found. An
unskilful choice of given clause can quickly lead to a large number
of irrelevant clauses in the set of unprocessed clauses,
making it dificult to find a proof. On the other hand, a skilful
choice of the given clause can lead to only those inferences
being made that are relevant to the proof task, so that a
proof can be found directly.</p>
        <p>To select the given clause, heuristics are usually used that
assign a weight to each clause. The clause with the lowest
weight is then selected as the given clause. Common
heuristics use the age of the clause (older clauses are favoured and
receive a lower weight) or are based on a weighted symbol
count. These heuristics are often used in combination. The
ratio of clauses selected by age to clauses selected by symbol
count is called the pick-given ratio. A well-known ratio is
the pick-given ratio of 5, which means that the heuristic
selects the clause with the lowest weighted symbol count
ifve times and the oldest clause once. There are many other
heuristics for selecting clauses. See Sect. 2 for some pointers
in this regard.</p>
        <p>In the following, the set of all predicate and function
symbols in a formula  is represented as sym(F ). In a
slight abuse of notation, we use sym(KB ) to denote the set
containing all predicate and function symbols occurring in
a knowledge base KB .
3.1. Distributional Semantics
Our approach uses the meaning of the names of the function
and predicate symbols. To do this, we rely on the
distributional semantics of natural language. This is best explained
using Firth’s famous statement:</p>
      </sec>
      <sec id="sec-2-2">
        <title>You shall know a word by the company it</title>
        <p>
          keeps. [
          <xref ref-type="bibr" rid="ref21">21</xref>
          ]
In other words: If we look at very large texts, we can observe
that words that occur in a similar context are similar [
          <xref ref-type="bibr" rid="ref22">22</xref>
          ].
        </p>
        <p>
          One approach in this area are word embeddings [
          <xref ref-type="bibr" rid="ref23">23</xref>
          ], [
          <xref ref-type="bibr" rid="ref24">24</xref>
          ].
These are vector representations of words, usually learned
using neural networks on very large amounts of text. For
a given set of words  , also called a vocabulary, a word
embedding is a function,  :  → R, that assigns an
-dimensional vector to each word in the vocabulary. Since
our approach uses existing word embeddings and does not
train new ones, we refrain from going into details on the
training process of word embeddings.
        </p>
        <p>It is very interesting that words with similar meanings
are mapped to similar vectors. Usually, the similarity of
two word vectors is calculated with the help of the cosine
similarity.</p>
        <p>Definition 1 (Cosine similarity of two vectors). Let ,  ∈
R, both non-zero. The cosine similarity of  and  is defined
as:
cos_sim(, ) =</p>
        <p>· 
|||| ||||</p>
        <p>The cosine similarity of two vectors  and  takes values
between -1 and 1. For exactly opposite vectors the value is
-1, for orthogonal vectors the value is 0 and for equal vectors
the value 1. The more similar two vectors are, the greater is
their cosine similarity.</p>
        <p>In our approach, we use a trained word embedding to
determine similarities between the names of symbols in
a knowledge base and the proof task. The next section
describes how we do this.
19–27</p>
        <p>KB
extraction of symbol names
set sym(KB) of symbol names in KB</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>4. Guiding the Reasoning Process by the Meaning of Symbol Names</title>
      <p>Automated theorem provers usually calculate the weight
of a clause from the weights of the symbols occurring in
the clause. For example, the theorem prover E can be given
values for weights for function and predicate symbols. These
weights then determine which weights the clauses receive
and therefore which clauses are preferred in the reasoning
process.</p>
      <p>We now present an approach to determine symbol
weights, and thus clause weights, that uses the meaning
of symbol names. As described above, we assume that there
are multiple proof tasks to solve for a given knowledge base.
In the following, when we talk about a knowledge base or
a goal, we refer to the knowledge base or the goal in this
scenario.</p>
      <p>
        Our approach uses a word embedding. It is important
that the word embedding fits to the domain of the
knowledge base. For example if the knowlege base is Yago [
        <xref ref-type="bibr" rid="ref25">25</xref>
        ]
it would be suitable to use a word embedding which was
learnt on wikipedia articles. In the following, we refer to
the word embedding used, as introduced above, as a
function  :  → R, where  is the vocabulary of the word
embedding.
      </p>
      <p>
        Fig. 1 shows the first step of our approach, which
determines a vector representation for all function and predicate
symbols occurring in the knowledge base under
consideration. To achieve this, in a first step, all function and predicate
symbols are extracted from the knowledge base resulting in
the set sym(KB ). Next, during the vector transformation,
all extracted symbols  ∈ sym(KB ) are mapped to the
vocabulary  of a word embedding. This mapping step is
typically not trivial and highly depends on the knowledge
base in use. It builds upon an existing WordNet mapping
for the knowledge base, if available, which is then
semiautomatically extended. We do not go into detail here and
refer the reader to [
        <xref ref-type="bibr" rid="ref26">26</xref>
        ] for more information. In the
following we assume that this step is possible for all symbols
extracted from the knowledge base and that there is a mapping
function map : sym(KB ) →  which performs this step.
Once a function or predicate symbol  ∈ sym(KB ) has
been assigned a word map(s) from the vocabulary of the
word embedding used, the vector  =  (map(s))
corresponding to the word is assigned to the symbol. This vector
 is used as the vector representation of the symbol . This
step only needs to be performed once for a knowledge base.
Therefore, it can be considered as a pre-processing step.
      </p>
      <p>Since we assume that for a large number of goals it has to
be proven that they are entailed from the same knowledge
base, the efort of this preprocessing step is relativized.</p>
      <p>Fig. 2 provides an overview of the process of proving a
goal using the symbol-name heuristic. First, a vector
representation of the goal is computed for a given goal . To do
this, all function and predicate symbols are extracted from
the goal leading to the set sym(G ) and mapped to words
in the vocabulary of the word embedding used. For this,
the above mentioned function map is used. It is important
to use the same word embedding that was chosen for the
knowledge base in the pre-processing step. The next step is
to compute the vector representation for the goal: For each
symbol  ∈ sym(G ) in the goal, the vector  of the word
map(s) is looked up in the word embedding. The vectors
of all symbols in sym(G ) are summed and divided by the
number of symbols in the goal resulting in vector :
(1)
 =
∑︁</p>
      <p>∈sym(G)
|sym(G )|
In other words, the vector representation of a goal
corresponds to the average of the vector representations of all
the symbols in the goal.</p>
      <p>In the next step, weights for symbols in the knowledge
base are computed based on the vector representations of
the function and predicate symbols in the knowledge base
and the vector representation of the goal. To compute the
weight of a symbol  ∈ sym(KB ), we first compute the
similarity of its vector representation  and the vector
representation of the goal . Cosine similarity is used for
this. The values of cosine similarity are between -1 and
1. The higher the similarity of the two vectors, the higher
the value. On the other hand, symbol with a low value
are preferred by provers. Therefore, we convert the cosine
similarities into weights as follows:</p>
      <p>We assume a default weight of 1000. From this we subtract
1000 times the cosine similarity. This allows us to account
for three decimal places in the cosine similarity. This results
in the weight of a symbol as
weight (, ) = 1000 − cos_sim(, )
(2)</p>
      <p>The weights of the symbols calculated in this way are
now passed to the theorem prover together with the proof
task. The prover calculates the weights for the clauses from
the weights for the symbols and uses them to control the
proof process.</p>
      <p>A first approach would be to pass all calculated weights
for symbols to the theorem prover. Experiments have shown
that this does not necessarily lead to the best results. Our
experiments have shown that it is most advantageous to
pass only the weights of the 10 highest scoring symbols and
leave all other symbol weights at the default value of 1000.</p>
    </sec>
    <sec id="sec-4">
      <title>5. Experimental Results</title>
      <p>
        To evaluate the approach, we need a knowledge base with
function and predicate symbols that are, or can be, mapped
to natural language words. We also need a large number
of proof tasks for this knowledge base. Adimen SUMO
[
        <xref ref-type="bibr" rid="ref27">27</xref>
        ] is a first-order logic ontology based on the Suggested
Upper Merged Ontology (SUMO). Since a mapping to
WordNet synsets is available for most of the symbols of Adimen
SUMO, this ontology is interesting for our evaluation.
WordNet is an extensive lexical database of the English language,
which groups words according to their meaning and forms
synonymous groups, called synsets. For the vectorization
we use the pretrained word embedding ConceptNet
Numberbatch [
        <xref ref-type="bibr" rid="ref28">28</xref>
        ] which is well suited for the commonsense
area.
      </p>
      <p>
        The proof tasks considered in the experiments were
created to test the application of the closed world assumption
(CWA) to first-order logic ontologies [
        <xref ref-type="bibr" rid="ref29">29</xref>
        ]. Each proof task
consists of a single formula, the goal, for which it must be
shown that it can be inferred from the Adimen SUMO
ontology. We call the set of all of these proof tasks the Adimen
SUMO CWA problems in the following.
      </p>
      <p>These Adimen SUMO problems are challenging even for
highly optimised theorem provers. Usually theorem provers
use selection techniques for this kind of problems, which
try to select the formulas relevant for a given proof task
from the knowledge base.</p>
      <p>In all our experiments, we use the theorem prover E with
a timeout of 10 seconds. We call it with diferent parameters
combinations2:
formulas.</p>
      <p>Fig. 3 shows the number of proofs found for the diferent
E configurations we considered. We see that the
configuration with no parameters finds the highest number of proofs
with 129 proofs, which is not surprising because of the way
the problems were chosen. Furthermore, we see that the
number of proofs found using the configuration with the
auto parameter (121 proofs), as well as the configurations
using the similarity-based weights, results in slightly fewer
proofs being found (122 for all configurations of E using
similarity-based symbol weights without auto parameter).</p>
      <p>Since our approach aims at influencing the selection of
the clause for the next inference step in such a way that
the selected clause fits the thematic context of the goal,
the number of processed clauses is of particular interest
for the evaluation. This number is usually much larger
than the number of clauses in the actual proof. The smaller
the number of processed clauses, the more targeted the
proof. Ideally, the number of processed clauses is equal to
the number of clauses in the proof, and no unnecessary
conclusions have been drawn.</p>
      <p>Fig. 4 shows the average number of clauses processed by
the diferent E configurations we considered. We can see
that E with the auto parameter processes the most clauses.
The number of clauses processed for E without any
parameter and E using the similarity-based weight for only the
symbol most similar to the goal are comparable. Again, this
is not surprising as one would not expect a single symbol
weight to have much influence. The configurations of E
using similarity-based symbol weights for the 5 to 25
symbols most similar to the goal are comparable in terms of the
number of clauses processed. The best configuration seems
to be the one with 10 symbol weights. Compared to the
conifguration using E without parameters, the configurations
using symbol weights for 5 to 25 symbols process only about
half the number of clauses. This shows the usefulness of the
presented approach. We have also investigated whether the
combination of the auto parameter and the symbol weights
for the 10 symbols most similar to the goal has advantages.
However, we observe that the number of clauses processed
by this configuration is comparable to the use of the auto
parameter alone.</p>
      <p>Fig. 5 shows a direct comparison of the number of
processed clauses for the E configuration without using any
parameters and the E configuration using the similarity
based weights for the 10 symbols most similar to the proof
goal. We observe that the majority of the points in Fig. 5 are
above the first bisector. Thus, we see that for most problems
the number of clauses processed could be reduced by using
the similarity-based symbol weights.</p>
      <p>Fig. 6 shows a direct comparison of the number of clauses
processed for the E configuration using the auto parameter
and the E configuration using the similarity-based weights
for the 10 symbols most similar to the proof goal. The vast
majority of the points in Fig. 6 are above the first bisector.
Many are also at a large distance from it. For example, we
see a cluster of problems for which E with auto-parameter
processed around 17500 clauses, but for which the
configuration with symbol weights processed only about 7500
clauses. This illustrates that the use of similarity-based
symbol weights was able to drastically reduce the number of
clauses processed in many cases.</p>
      <p>Since we observe large diferences in the number of
clauses processed, the question arises whether the sizes of
the proofs found also difer. In Fig. 7 we see the direct
comparison of the sizes of proofs constructed with E using the
auto parameter with the sizes of proofs constructed with E
and the similarity-based symbol weights of the 10 symbols
most similar to the goal. We observe that the proofs are often
smaller when E was using similarity-based symbol weights.
In future work, we plan to analyze the diferent proofs in
more detail to determine how they difer and whether the
shorter proofs are easier to understand.</p>
      <p>Fig 8 shows the corresponding comparison, where we
compared the proof size for proofs constructed with E
without any parameters with the proof sizes of proofs
constructed with E and the similarity-based symbol weights of
the 10 symbols most similar to the goal. Here we observe
that the size of some proofs is smaller using the
configuration of E with symbol weights. However, the diferences are
not as large as in Fig 7, and in most cases the sizes do not
difer.</p>
      <p>
        In a second set of experiments, we selected formulas from
500 of the Adimen SUMO CWA problems with the SInE
selection [
        <xref ref-type="bibr" rid="ref30">30</xref>
        ] (recursion depth 3, generosity 5.0) and used
the resulting formula sets as test data for our evaluation. It
is important to note that the SInE selection is incomplete.
It is therefore possible that not all the necessary formulae
for a proof were selected. Consequently, it is unlikely that a
proof can be found for all of these 500 problems.
      </p>
      <p>Since the E configuration, which combines symbol
weights with the auto parameter, performed poorly in our
ifrst experiments, we do not include this configuration in
the following experiments.</p>
      <p>In Fig. 9 we can see that nearly the same number of proofs
can be found by the diferent configurations of E. The low
number of around 200 found proofs can be explained by the
incompleteness of the SInE selection.</p>
      <p>Fig. 10 shows the average number of clauses which were
processed during proof construction. We observe that the
average number of clauses processed by E with the auto
parameter is the highest value with 1213.85. The average
number of clauses processed by E without parameters is
slightly lower with 1103.68. In comparison, E with the 10
best similarity-based symbol weights processes on average
only 473.10 clauses, which is much less.</p>
      <p>Fig. 10 shows that the similarity-based weights lead to a
reduction in the average number of clauses processed.
However, it is not clear how this reduction is distributed across
the individual problems. Therefore, in Fig. 11 we directly
compare the number of processed clauses of E without any
parameters (y-axis) with the number of processed clauses of
E with the symbol weights of the 10 symbols most similar
to the goal (x-axis). For each proof we see a point in the
graph. We observe that, with a few exceptions, the number
of processed clauses is significantly lower for the E
configuration with goal similarity-based symbol weights than for E
without parameters.</p>
      <p>As in the previous experiments, we also compare the sizes
of the found proofs. Fig. 12 and Fig. 13 show the
corresponding results. We observe that for the problems considered
in the experiments with preselected formulae, the
diferences in proof size are not that large. There are many proofs
where the proof sizes are the same. Furthermore, for both
E configurations, there are proofs where the proof size is
smaller and proofs where the proof size is larger.
Comparing the proof sizes of E without any parameters with the
configuration using 10 similarity-based symbol weights, we
observe slightly more cases where the proof size is lower
using symbol weights. But the diferences are really small.</p>
      <p>Compared to the experiments with the non preselected
formulas (see Fig. 7 and Fig. 8), the diferences in the proof
sizes are much smaller here. A possible explanation for this
could be the pre-selection of formulae with SInE. This
preselection means that there are significantly fewer formulae
available, which considerably limits the choices when
constructing the proofs. We intend to investigate this in more
detail in future work.</p>
    </sec>
    <sec id="sec-5">
      <title>6. Summary and Future Work</title>
      <p>In this paper we have developed an approach to clause
selection in automated theorem provers that aims to make
the selection of the clause for the next inference step more
goal-directed. The approach is based on the assumption
that symbol names in knowledge bases have meaningful
names that can be mapped to natural language words. If this
assumption is correct, techniques from natural language
processing can be used to determine the similarity between the
proof goal and the symbols in the knowledge base. These
similarities are then used to derive symbol weights that
guide the selection of clauses for the next inference step.
The approach has been implemented and evaluated using
the theorem prover E. Our experimental results show a
significant reduction in the number of clauses processed
during proof construction, indicating a more goal-directed
and eficient proof process.</p>
      <p>In our experiments, we further observed that the use of
similarity-based symbol weights has an impact on proof
size (especially compared to the configuration using the
auto parameter). In future work, we would like to
analyze the diferences in the proofs we found in more detail
and investigate whether there are also diferences in the
comprehensibility of the proofs.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>D.</given-names>
            <surname>Kahneman</surname>
          </string-name>
          , Thinking, Fast and Slow, Macmillan,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>S.</given-names>
            <surname>Schulz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Cruanes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Vukmirovic</surname>
          </string-name>
          , Faster, higher,
          <source>stronger: E 2</source>
          .3, in: P. Fontaine (Ed.),
          <source>Automated Deduction - CADE 27 - 27th International Conference on Automated Deduction, Natal, Brazil, August 27-30</source>
          ,
          <year>2019</year>
          , Proceedings, volume
          <volume>11716</volume>
          of Lecture Notes in Computer Science, Springer,
          <year>2019</year>
          , pp.
          <fpage>495</fpage>
          -
          <lpage>507</lpage>
          . URL: https://doi.org/10.1007/978-3-
          <fpage>030</fpage>
          -29436-6_
          <fpage>29</fpage>
          . doi:
          <volume>10</volume>
          . 1007/978-3-
          <fpage>030</fpage>
          -29436-6\_
          <fpage>29</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>A.</given-names>
            <surname>Riazanov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Voronkov</surname>
          </string-name>
          ,
          <article-title>The design and implementation of VAMPIRE</article-title>
          ,
          <source>AI Commun</source>
          .
          <volume>15</volume>
          (
          <year>2002</year>
          )
          <fpage>91</fpage>
          -
          <lpage>110</lpage>
          . URL: http://content.iospress.com/articles/ ai-communications/
          <year>aic259</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>S.</given-names>
            <surname>Schulz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Möhrmann</surname>
          </string-name>
          ,
          <article-title>Performance of clause selection heuristics for saturation-based theorem proving</article-title>
          , in: N.
          <string-name>
            <surname>Olivetti</surname>
            ,
            <given-names>A</given-names>
          </string-name>
          . Tiwari (Eds.),
          <source>Automated Reasoning - 8th International Joint Conference, IJCAR</source>
          <year>2016</year>
          , Coimbra, Portugal, June 27 - July 2,
          <year>2016</year>
          , Proceedings, volume
          <volume>9706</volume>
          of Lecture Notes in Computer Science, Springer,
          <year>2016</year>
          , pp.
          <fpage>330</fpage>
          -
          <lpage>345</lpage>
          . URL: https:// doi.org/10.1007/978-3-
          <fpage>319</fpage>
          -40229-1_
          <fpage>23</fpage>
          . doi:
          <volume>10</volume>
          .1007/ 978-3-
          <fpage>319</fpage>
          -40229-1\_
          <fpage>23</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>F.</given-names>
            <surname>Bártek</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Suda</surname>
          </string-name>
          ,
          <article-title>How much should this symbol weigh? A gnn-advised clause selection</article-title>
          , in: R.
          <string-name>
            <surname>Piskac</surname>
            ,
            <given-names>A</given-names>
          </string-name>
          . Voronkov (Eds.),
          <source>LPAR 2023: Proceedings of 24th International Conference on Logic for Programming</source>
          ,
          <source>Artificial Intelligence and Reasoning</source>
          , Manizales, Colombia,
          <fpage>4</fpage>
          -9th
          <source>June</source>
          <year>2023</year>
          , volume
          <volume>94</volume>
          of EPiC Series in Computing, EasyChair,
          <year>2023</year>
          , pp.
          <fpage>96</fpage>
          -
          <lpage>111</lpage>
          . URL: https: //doi.org/10.29007/5f4r. doi:
          <volume>10</volume>
          .29007/5F4R.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>J.</given-names>
            <surname>Denzinger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Schulz</surname>
          </string-name>
          ,
          <article-title>Learning domain knowledge to improve theorem proving</article-title>
          , in: M. A.
          <string-name>
            <surname>McRobbie</surname>
            ,
            <given-names>J. K.</given-names>
          </string-name>
          Slaney (Eds.),
          <source>Automated Deduction - CADE-13, 13th International Conference on Automated Deduction</source>
          , New Brunswick, NJ, USA,
          <source>July 30 - August 3</source>
          ,
          <year>1996</year>
          , Proceedings, volume
          <volume>1104</volume>
          of Lecture Notes in Computer Science, Springer,
          <year>1996</year>
          , pp.
          <fpage>62</fpage>
          -
          <lpage>76</lpage>
          . URL: https://doi.org/10.1007/3-540-61511-3_
          <fpage>69</fpage>
          . doi:
          <volume>10</volume>
          .1007/3-540-61511-3\_
          <fpage>69</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>J.</given-names>
            <surname>Jakubuv</surname>
          </string-name>
          ,
          <string-name>
            <surname>J. Urban,</surname>
          </string-name>
          <article-title>ENIGMA: eficient learningbased inference guiding machine</article-title>
          , in: H.
          <string-name>
            <surname>Geuvers</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>England</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          <string-name>
            <surname>Hasan</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Rabe</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          Teschke (Eds.),
          <source>Intelligent Computer Mathematics - 10th International Conference, CICM</source>
          <year>2017</year>
          ,
          <article-title>Edinburgh</article-title>
          ,
          <string-name>
            <surname>UK</surname>
          </string-name>
          ,
          <source>July 17-21</source>
          ,
          <year>2017</year>
          , Proceedings, volume
          <volume>10383</volume>
          of Lecture Notes in Computer Science, Springer,
          <year>2017</year>
          , pp.
          <fpage>292</fpage>
          -
          <lpage>302</lpage>
          . URL: https://doi.org/10.1007/978-3-
          <fpage>319</fpage>
          -62075-6_
          <fpage>20</fpage>
          . doi:
          <volume>10</volume>
          . 1007/978-3-
          <fpage>319</fpage>
          -62075-6\_
          <fpage>20</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>J.</given-names>
            <surname>Jakubuv</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Urban</surname>
          </string-name>
          ,
          <article-title>Enhancing ENIGMA given clause guidance</article-title>
          , in: F.
          <string-name>
            <surname>Rabe</surname>
            ,
            <given-names>W. M.</given-names>
          </string-name>
          <string-name>
            <surname>Farmer</surname>
            ,
            <given-names>G. O.</given-names>
          </string-name>
          <string-name>
            <surname>Passmore</surname>
            ,
            <given-names>A</given-names>
          </string-name>
          . Youssef (Eds.),
          <source>Intelligent Computer Mathematics - 11th International Conference, CICM</source>
          <year>2018</year>
          , Hagenberg, Austria,
          <source>August 13-17</source>
          ,
          <year>2018</year>
          , Proceedings, volume
          <volume>11006</volume>
          of Lecture Notes in Computer Science, Springer,
          <year>2018</year>
          , pp.
          <fpage>118</fpage>
          -
          <lpage>124</lpage>
          . URL: https:// doi.org/10.1007/978-3-
          <fpage>319</fpage>
          -96812-4_
          <fpage>11</fpage>
          . doi:
          <volume>10</volume>
          .1007/ 978-3-
          <fpage>319</fpage>
          -96812-4\_
          <fpage>11</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>J.</given-names>
            <surname>Jakubuv</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Urban</surname>
          </string-name>
          ,
          <article-title>Hammering mizar by learning clause guidance (short paper)</article-title>
          , in: J.
          <string-name>
            <surname>Harrison</surname>
          </string-name>
          ,
          <string-name>
            <surname>J. O'Leary</surname>
            ,
            <given-names>A</given-names>
          </string-name>
          . Tolmach (Eds.),
          <source>10th International Conference on Interactive Theorem Proving, ITP 2019, September 9-12</source>
          ,
          <year>2019</year>
          , Portland,
          <string-name>
            <surname>OR</surname>
          </string-name>
          , USA, volume
          <volume>141</volume>
          of LIPIcs,
          <source>Schloss Dagstuhl - Leibniz-Zentrum für Informatik</source>
          ,
          <year>2019</year>
          , pp.
          <volume>34</volume>
          :
          <fpage>1</fpage>
          -
          <lpage>34</lpage>
          :
          <fpage>8</fpage>
          . URL: https://doi.org/10. 4230/LIPIcs.ITP.
          <year>2019</year>
          .
          <volume>34</volume>
          . doi:
          <volume>10</volume>
          .4230/LIPICS.ITP.
          <year>2019</year>
          .
          <volume>34</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>M.</given-names>
            <surname>Suda</surname>
          </string-name>
          ,
          <article-title>Improving enigma-style clause selection while learning from history</article-title>
          , in: A.
          <string-name>
            <surname>Platzer</surname>
          </string-name>
          , G. Sutclife (Eds.),
          <source>Automated Deduction - CADE 28 - 28th International Conference on Automated Deduction, Virtual Event, July 12-15</source>
          ,
          <year>2021</year>
          , Proceedings, volume
          <volume>12699</volume>
          of Lecture Notes in Computer Science, Springer,
          <year>2021</year>
          , pp.
          <fpage>543</fpage>
          -
          <lpage>561</lpage>
          . URL: https:// doi.org/10.1007/978-3-
          <fpage>030</fpage>
          -79876-5_
          <fpage>31</fpage>
          . doi:
          <volume>10</volume>
          .1007/ 978-3-
          <fpage>030</fpage>
          -79876-5\_
          <fpage>31</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>G.</given-names>
            <surname>Bancerek</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Bylinski</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Grabowski</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Kornilowicz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Matuszewski</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Naumowicz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Pak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Urban</surname>
          </string-name>
          , Mizar:
          <article-title>State-of-the-art and beyond</article-title>
          , in: M.
          <string-name>
            <surname>Kerber</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Carette</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Kaliszyk</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Rabe</surname>
          </string-name>
          , V. Sorge (Eds.), Intelligent Computer Mathematics - International Conference, CICM 2015, Washington, DC, USA, July
          <volume>13</volume>
          -
          <issue>17</issue>
          ,
          <year>2015</year>
          , Proceedings, volume
          <volume>9150</volume>
          of Lecture Notes in Computer Science, Springer,
          <year>2015</year>
          , pp.
          <fpage>261</fpage>
          -
          <lpage>279</lpage>
          . URL: https://doi.org/10.1007/978-3-
          <fpage>319</fpage>
          -20615-8_
          <fpage>17</fpage>
          . doi:
          <volume>10</volume>
          . 1007/978-3-
          <fpage>319</fpage>
          -20615-8\_
          <fpage>17</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>J.</given-names>
            <surname>Alama</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Heskes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Kühlwein</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Tsivtsivadze</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Urban</surname>
          </string-name>
          ,
          <article-title>Premise selection for mathematics by corpus analysis and kernel methods</article-title>
          ,
          <source>J. Autom. Reason</source>
          .
          <volume>52</volume>
          (
          <year>2014</year>
          )
          <fpage>191</fpage>
          -
          <lpage>213</lpage>
          . URL: https: //doi.org/10.1007/s10817-013-9286-5. doi:
          <volume>10</volume>
          .1007/ S10817-013-9286-5.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>G.</given-names>
            <surname>Irving</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Szegedy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. A.</given-names>
            <surname>Alemi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Eén</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Chollet</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Urban</surname>
          </string-name>
          ,
          <article-title>Deepmath - deep sequence models for premise selection</article-title>
          , in: D. D.
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Sugiyama</surname>
            , U. von Luxburg,
            <given-names>I. Guyon</given-names>
          </string-name>
          , R. Garnett (Eds.),
          <source>Advances in Neural Information Processing Systems 29: Annual Conference on Neural Information Processing Systems</source>
          <year>2016</year>
          , Barcelona, Spain,
          <year>2016</year>
          , pp.
          <fpage>2235</fpage>
          -
          <lpage>2243</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>B.</given-names>
            <surname>Piotrowski</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Urban</surname>
          </string-name>
          ,
          <article-title>Stateful premise selection by recurrent neural networks</article-title>
          , in: E.
          <string-name>
            <surname>Albert</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          Kovács (Eds.),
          <source>LPAR 2020: 23rd International Conference on Logic for Programming</source>
          ,
          <source>Artificial Intelligence and Reasoning</source>
          , Alicante, Spain, May
          <volume>22</volume>
          -27,
          <year>2020</year>
          , volume
          <volume>73</volume>
          of EPiC Series in Computing, EasyChair,
          <year>2020</year>
          , pp.
          <fpage>409</fpage>
          -
          <lpage>422</lpage>
          . URL: https://doi.org/10.29007/j5hd. doi:
          <volume>10</volume>
          . 29007/J5HD.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>K.</given-names>
            <surname>Prorokovic</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Wand</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Schmidhuber</surname>
          </string-name>
          ,
          <article-title>Improving stateful premise selection with transformers</article-title>
          , in: F. Kamareddine, C. S. Coen (Eds.),
          <source>Intelligent Computer Mathematics - 14th International Conference, CICM</source>
          <year>2021</year>
          , Timisoara, Romania,
          <source>July 26-31</source>
          ,
          <year>2021</year>
          , Proceedings, volume
          <volume>12833</volume>
          of Lecture Notes in Computer Science, Springer,
          <year>2021</year>
          , pp.
          <fpage>84</fpage>
          -
          <lpage>89</lpage>
          . URL: https: //doi.org/10.1007/978-3-
          <fpage>030</fpage>
          -81097-
          <issue>9</issue>
          _6. doi:
          <volume>10</volume>
          .1007/ 978-3-
          <fpage>030</fpage>
          -81097-9\_6.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>M.</given-names>
            <surname>Mikula</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Antoniak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Tworkowski</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. Q.</given-names>
            <surname>Jiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. P.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Szegedy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Kucinski</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Milos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <article-title>Magnushammer: A transformer-based approach to premise selection</article-title>
          ,
          <source>CoRR abs/2303</source>
          .04488 (
          <year>2023</year>
          ). URL: https://doi.org/10.48550/arXiv.2303.04488. doi:
          <volume>10</volume>
          . 48550/ARXIV.2303.04488. arXiv:
          <volume>2303</volume>
          .
          <fpage>04488</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17] S. de Rooij, W. Beek,
          <string-name>
            <given-names>P.</given-names>
            <surname>Bloem</surname>
          </string-name>
          , F. van
          <string-name>
            <surname>Harmelen</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Schlobach</surname>
          </string-name>
          ,
          <article-title>Are names meaningful? quantifying social meaning on the semantic web</article-title>
          ,
          <source>in: ISWC (1)</source>
          , volume
          <volume>9981</volume>
          of Lecture Notes in Computer Science,
          <year>2016</year>
          , pp.
          <fpage>184</fpage>
          -
          <lpage>199</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>U.</given-names>
            <surname>Furbach</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Krämer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Schon</surname>
          </string-name>
          ,
          <article-title>Names are not just sound and smoke: Word embeddings for axiom selection</article-title>
          ,
          <source>in: CADE</source>
          , volume
          <volume>11716</volume>
          of Lecture Notes in Computer Science, Springer,
          <year>2019</year>
          , pp.
          <fpage>250</fpage>
          -
          <lpage>268</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>C.</given-names>
            <surname>Schon</surname>
          </string-name>
          ,
          <article-title>Associative reasoning for commonsense knowledge</article-title>
          , in: D.
          <string-name>
            <surname>Seipel</surname>
            ,
            <given-names>A</given-names>
          </string-name>
          . Steen (Eds.),
          <source>KI 2023: Advances in Artificial Intelligence - 46th German Conference on AI</source>
          , Berlin, Germany,
          <source>September 26-29</source>
          ,
          <year>2023</year>
          , Proceedings, volume
          <volume>14236</volume>
          of Lecture Notes in Computer Science, Springer,
          <year>2023</year>
          , pp.
          <fpage>170</fpage>
          -
          <lpage>183</lpage>
          . URL: https://doi.org/10.1007/978-3-
          <fpage>031</fpage>
          -42608-7_
          <fpage>14</fpage>
          . doi:
          <volume>10</volume>
          . 1007/978-3-
          <fpage>031</fpage>
          -42608-7\_
          <fpage>14</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>C. S.</given-names>
            <surname>Oliver</surname>
          </string-name>
          <string-name>
            <surname>Jakobs</surname>
          </string-name>
          ,
          <article-title>Context-specific selection of commonsense knowledge using large language models</article-title>
          , in: to appear
          <source>in KI 2024, Lecture Notes in Computer Science</source>
          , Springer,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>J. R.</given-names>
            <surname>Firth</surname>
          </string-name>
          ,
          <source>Papers in Linguistics 1934 - 1951: Rep</source>
          , Oxford University Press,
          <year>1991</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>G. A.</given-names>
            <surname>Miller</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W. G.</given-names>
            <surname>Charles</surname>
          </string-name>
          ,
          <article-title>Contextual correlates of semantic similarity</article-title>
          ,
          <source>Language and Cognitive Processes</source>
          <volume>6</volume>
          (
          <year>1991</year>
          )
          <fpage>1</fpage>
          -
          <lpage>28</lpage>
          . URL: http://eric.ed.gov/ERICWebPortal/ recordDetail?accno=
          <fpage>EJ431389</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>T.</given-names>
            <surname>Mikolov</surname>
          </string-name>
          , I. Sutskever,
          <string-name>
            <given-names>K.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G. S.</given-names>
            <surname>Corrado</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Dean</surname>
          </string-name>
          ,
          <article-title>Distributed representations of words and phrases and their compositionality</article-title>
          ,
          <source>in: NIPS</source>
          ,
          <year>2013</year>
          , pp.
          <fpage>3111</fpage>
          -
          <lpage>3119</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>T.</given-names>
            <surname>Mikolov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Chen</surname>
          </string-name>
          , G. Corrado,
          <string-name>
            <given-names>J.</given-names>
            <surname>Dean</surname>
          </string-name>
          ,
          <article-title>Eficient estimation of word representations in vector space</article-title>
          ,
          <source>CoRR abs/1301</source>
          .3781 (
          <year>2013</year>
          ). URL: http://arxiv.org/abs/ 1301.3781. arXiv:
          <volume>1301</volume>
          .
          <fpage>3781</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>F. M.</given-names>
            <surname>Suchanek</surname>
          </string-name>
          , G. Kasneci, G. Weikum,
          <article-title>Yago: A large ontology from Wikipedia and WordNet</article-title>
          , Web Semant.
          <volume>6</volume>
          (
          <year>2008</year>
          )
          <fpage>203</fpage>
          -
          <lpage>217</lpage>
          . URL: http://dx.doi.org/10. 1016/j.websem.
          <year>2008</year>
          .
          <volume>06</volume>
          .001. doi:
          <volume>10</volume>
          .1016/j.websem.
          <year>2008</year>
          .
          <volume>06</volume>
          .001.
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>C.</given-names>
            <surname>Schon</surname>
          </string-name>
          ,
          <article-title>Selection strategies for commonsense knowledge</article-title>
          ,
          <year>2022</year>
          . URL: https://arxiv.org/abs/2202.09163. doi:
          <volume>10</volume>
          .48550/ARXIV.2202.09163.
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [27]
          <string-name>
            <given-names>J.</given-names>
            <surname>Álvez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Lucio</surname>
          </string-name>
          , G. Rigau,
          <article-title>Adimen-sumo: Reengineering an ontology for first-order reasoning</article-title>
          ,
          <source>Int. J. Semantic Web Inf. Syst</source>
          .
          <volume>8</volume>
          (
          <year>2012</year>
          )
          <fpage>80</fpage>
          -
          <lpage>116</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [28]
          <string-name>
            <given-names>R.</given-names>
            <surname>Speer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Chin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Havasi</surname>
          </string-name>
          ,
          <article-title>Conceptnet 5.5: An open multilingual graph of general knowledge</article-title>
          , in: AAAI, AAAI Press,
          <year>2017</year>
          , pp.
          <fpage>4444</fpage>
          -
          <lpage>4451</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          [29]
          <string-name>
            <given-names>J.</given-names>
            <surname>Álvez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I.</given-names>
            <surname>Gonzalez-Dios</surname>
          </string-name>
          ,
          <string-name>
            <surname>G.</surname>
          </string-name>
          <article-title>Rigau, Applying the closed world assumption to sumo-based FOL ontologies for efective commonsense reasoning</article-title>
          , in: G. D.
          <string-name>
            <surname>Giacomo</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Catalá</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Dilkina</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Milano</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Barro</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Bugarín</surname>
          </string-name>
          , J. Lang (Eds.),
          <source>ECAI 2020 - 24th European Conference on Artificial Intelligence</source>
          ,
          <volume>29</volume>
          <fpage>August</fpage>
          -8
          <source>September</source>
          <year>2020</year>
          , Santiago de Compostela, Spain,
          <source>August 29 - September 8, 2020 - Including 10th Conference on Prestigious Applications of Artificial Intelligence (PAIS</source>
          <year>2020</year>
          ), volume
          <volume>325</volume>
          <source>of Frontiers in Artificial Intelligence and Applications</source>
          , IOS Press,
          <year>2020</year>
          , pp.
          <fpage>585</fpage>
          -
          <lpage>592</lpage>
          . URL: https://doi.org/10.3233/FAIA200142. doi:
          <volume>10</volume>
          .3233/FAIA200142.
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          [30]
          <string-name>
            <given-names>K.</given-names>
            <surname>Hoder</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Voronkov</surname>
          </string-name>
          ,
          <article-title>Sine qua non for large theory reasoning</article-title>
          , in: N.
          <string-name>
            <surname>Bjørner</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          <string-name>
            <surname>Sofronie-Stokkermans</surname>
          </string-name>
          (Eds.),
          <source>Automated Deduction - CADE-23</source>
          , volume
          <volume>6803</volume>
          of Lecture Notes in Computer Science, Springer Berlin Heidelberg,
          <year>2011</year>
          , pp.
          <fpage>299</fpage>
          -
          <lpage>314</lpage>
          . URL: http://dx. doi.org/10.1007/978-3-
          <fpage>642</fpage>
          -22438-6_
          <fpage>23</fpage>
          . doi:
          <volume>10</volume>
          .1007/ 978-3-
          <fpage>642</fpage>
          -22438-6_
          <fpage>23</fpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>