<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Learning from Ordinal Data with Inductive Logic Programming in Description Logic</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Nunung Nurul Qomariyah</string-name>
          <email>nq516@york.ac.uk</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Dimitar Kazakov</string-name>
          <email>dimitar.kazakov@york.ac.uk</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>AI Group</institution>
          ,
          <addr-line>Computer Science</addr-line>
          ,
          <institution>University of York</institution>
          ,
          <country country="UK">UK</country>
        </aff>
      </contrib-group>
      <fpage>38</fpage>
      <lpage>50</lpage>
      <abstract>
        <p>Here we describe a Description Logic (DL) based Inductive Logic Programming (ILP) algorithm for learning relations of order. We test our algorithm on the task of learning user preferences from pairwise comparisons. The results have implications for the development of customised recommender systems for e-commerce, and more broadly, wherever DL-based representations of knowledge, such as OWL ontologies, are used. The use of DL makes for easy integration with such data, and produces hypotheses that are easy to interpret by novice users. The proposed algorithm outperforms SVM, Decision Trees and Aleph on data from two domains.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>Copyright c by the paper's authors. Copying permitted for private and academic purposes.</p>
      <p>
        In: Nicolas Lachiche, Christel Vrain (eds.): Late Breaking Papers of ILP 2017, Orleans, France, September 4-6, 2017, published at
http://ceur-ws.org
preference learning, although there are exceptions [
        <xref ref-type="bibr" rid="ref10 ref14 ref19 ref20 ref3">3, 10, 20, 19, 14</xref>
        ]. Balakrishnan and Chopra [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] published a
research study on pairwise preferences by using a Bayesian framework. Qian et al. [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ] published a similar idea,
proposing an approach to learn user preferences by using pairwise comparisons to explore each attribute in turn
through \orthogonal queries" and then applying linear SVM to approximate the preferences. Similarly, Jensen
et al. [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] apply a Gaussian Process regression model for comparisons between music tracks. Another similar
work was performed by Rokach and Kisilevich [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ] which uses lazy decision trees with pairwise comparisons at
each node.
      </p>
      <p>
        Most studies on pairwise preference learning use statistical machine learning approaches while in this paper,
we propose a logic-based approach. There is one other attempt at using ILP for preference learning [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ], which
proposes a framework for learning weak constraints in Answer Set Programming (ASP). We build on previous
work combining ILP and DL, namely, the DL-Learner [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ], which aims to nd a correct class description in a
speci c ontology from a set of positive and negative examples of individuals. It builds a hypothesis in the form of
a class description, which can contain conjunction, disjunction and existential quanti cation. DL-Learner itself
develops previous work on systems, such as YinYang [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] and DL-FOIL [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. Kietz's [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] and Konstantopoulos'
[
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] work is also relevant in this context. While DL-Learner can only learn class de nitions, we aim to learn
de nitions of relations. Here we focus on the binary relation of order, and in particular, strict order, a relation
that is transitive, anti-re exive and anti-symmetric. Our aim is to learn each user's individual preferences using
a DL-based representation, and an algorithm inspired by the Progol/Aleph family of ILP tools. We evaluate
our algorithm on two datasets, one on car preferences [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], the second on sushi preferences [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]. Both datasets
provide pairwise preference choices of multiple users.
      </p>
      <p>The rest of the paper is organized as follows: in Section 2 we explain the learning task in terms of both logic
programming and DL to make the problem clearer to readers from either background. We then propose an
algorithm based on ILP in Section 3 and provide the analysis of the evaluation in Section 4. Finally, we conclude
our work and provide our plan for further work in Section 5.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Problem Representation</title>
      <p>DLs represent the world in terms of concepts, objects and roles. Concepts can be seen as a formal de nition of
classes in the application domain, e.g. one can de ne a father as a man who has at least one child. Concepts have
two functions: (1) to describe a collection of objects and (2) to describe the properties that a class should hold.
Objects are the entities that belong to one concept or another. Roles represent binary relationships between
objects, e.g. John is Anna's father. These can also be used to describe object properties, e.g. John's age is 42.
In FOL terminology, objects correspond to unique constants (i.e. ground terms), concepts correspond to unary
predicates, and roles correspond to binary predicates. All this information is stored in the form of a knowledge
base which comprises of two components, an assertional part (ABox ) and a terminological part (TBox ).The
TBox is a nite set of General Concept Inclusions (GCI) and role inclusions. It corresponds to the notion of a
schema in a relational database. The assertional set in the knowledge base is called the ABox. It is used to state
properties of individuals. In a relational database, this is the actual data, while in the FOL it corresponds to
facts (aka ground terms).</p>
      <p>The knowledge representation in DL can be expressed using the Resource Description Framework (RDF),
which is one of the most powerful general-purpose knowledge representation languages. The syntax of RDF is
based on triples in the form of subject-predicate-object expressions. An RDF triple is a statement expressing a
relation (\predicate") between the object in its rst argument, and the object or literal in its third argument
(e.g. \Anna-has father-John"). In this section, we use Turtle1 syntax to describe RDF data2
The preference learning problem is de ned as follows:</p>
      <p>Given: a set of individuals, a class hierarchy, a mapping from individuals to classes, and a set of
preferences represented as pairs of individuals, where the rst individual is preferred over (strictly
better than) the second,
Find: a disjunction of axioms de ning the domain and range of the relation betterthan (where each
axiom is expressed as a conjunction of classes) that is complete and consistent with the given preferences.
This is a supervised learning problem as the user assigns one of two possible labels for each pair of items after
1https://www.w3.org/TR/turtle/
2Turtle stands for Terse RDF Triple Language.
considering their attributes. We then use these labels to search for a de nition of the Domain and Range class
memberships that render the relation true.</p>
      <p>We propose an approach to learn relations and apply the resulting system to learn strict order (i.e. better
than). The properties of strict order relation are anti-symmetric, which means that if item A is better than item
B, then in any case item B cannot be better than item A, unless A=B. That special case is then excluded by the
additional assumption of betterthan being anti-re exive (i.e. X cannot be seen as better than itself). It is also
transitive, which means whenever item A is better than item B, and item B is better than item C, then item
A is better than item C. The ontology reasoner can always be used to extract all pairs of items satisfying the
(b) The user annotation is translated into the object property
\betterthan".
(a) A simple car class hierarchy
preference relation as a result of applying its transitivity property. We add these inferred examples to the set of
positive examples so that the complete closure is used by the learning algorithm. We also use the anti-symmetry
property to produce all negative examples as a `mirror image' of the positive (after completing their transitive
closure), i.e., all pairs derived from a positive example through the swap of the rst and second element in the
pair.</p>
      <p>The representation of a sample preference learning problem in DL is shown in Figure 1. The class hierarchy is
given to the system as an input. We evaluate our algorithm using a simple class hierarchy as shown in Figure 1a
in order to make the solutions comparable to the Aleph representation.
2.1</p>
      <sec id="sec-2-1">
        <title>Hypothesis Language</title>
        <p>The aim of ILP is to nd a theory that is complete (it covers all the given positive examples) and consistent (it
covers no negative examples). In our algorithm learning from RDF data, we build a hypothesis for an object
property we want to learn, as in the case of the property betterthan. The fact that betterthan is an object
property is described in Turtle as shown below:</p>
        <p>myontology:betterthan rdf:type owl:ObjectProperty.</p>
        <p>Each of the possible hypotheses about this relation is then described as a pair of class de nitions, which specify
the membership of the domain D and the range R of the relation. The same hypothesis language can be described
in Aleph notation as the following mode declarations:
:- modeh(1,betterthan(+car,+car)).
:- modeb(1,hasbodytype(+car,#bodytype)).
:- modeb(1,hasfuelcons(+car,#fuelconsumption)).
:- modeb(1,hasbodytype(+car,#bodytype)).
:- modeb(1,hastransmission(+car,#transmission)).
:- modeb(1,hasenginesize(+car,#enginesize)).
2.2</p>
      </sec>
      <sec id="sec-2-2">
        <title>Background Knowledge</title>
        <p>Our algorithm represents background knowledge through classes and their membership, as shown in the example
below (which uses Turtle syntax):
myontology:Sedan rdf:type owl:Class ;
rdfs:subClassOf myontology:Car ;
owl:disjointWith myontology:Suv .
myontology:car1 rdf:type owl:NamedIndividual ,
myontology:Car ,
myontology:Manual ,
myontology:NonHybrid ,
myontology:SmallCar ,
myontology:Suv .</p>
        <p>The same background knowledge can be spelt out in Aleph as follows:
car(car1). %type predicate
bodytype(sedan). %type predicate
hasbodytype(car1,suv). %bk
hasfuelcons(car1,nonhybrid). %bk
hastransmission(car1,manual). %bk
hasenginesize(car1, small). %bk</p>
      </sec>
      <sec id="sec-2-3">
        <title>2.3 Examples</title>
        <p>In our algorithm, the set of positive examples is de ned by the user's preferences, and the following pre-processing
step that computes the transitive closure of these preferences, which are then negated to produce all negative
examples. The following Turtle code expresses that car1 is better than car3, car4, car5, and car10:
myontology:car1 myontology:betterthan myontology:car3 ,
myontology:car4 ,
myontology:car5 ,
myontology:car10 .</p>
        <p>In traditional ILP syntax, the same examples are represented as ground facts of the predicate
betterthan/2, where the arguments are of type car. So, the positive examples in Aleph are written as:
betterthan(car1,car3), and the negative, obtained by their reversal, as :-betterthan(car3,car1).
3</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Proposed Algorithm</title>
      <p>
        We propose an algorithm called APARELL (Active3 Pairwise Relation Learner) which learns the relation of
strict order in DL. We implement our algorithm in Java using the OWL API4 [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] and RDF4J API 5 to handle
DL. The algorithm is shown in Algorithm 1. We follow the four basic steps used in the Progol/Aleph greedy
learning approach:
1. Select a positive example. Each instance of the relation can be seen as a pair of object IDs.
2. Build the bottom clause. The bottom clause is constructed from the conjunction of all non-disjoint
classes of which each individual in the pair is a member.
3. Search. This step is to nd all generalisations of the bottom clause of the same, lowest possible, length
that are consistent with the data.
3The active learning algorithm feature has not been described or used in this work.
4http://owlapi.sourceforge.net/
5http://rdf4j.org/
4. Remove covered positive examples. Our algorithm is greedy in its treatment of positive training
examples. We remove all covered positive examples once all clauses found by the search step are added to
the current theory.
3.1
      </p>
      <sec id="sec-3-1">
        <title>Search and Re nement Operator</title>
        <p>We use a top-down approach similar to the one in Progol/Aleph starting from the shortest clause possible. Note
that for the speci c class of problems, namely, learning strict order, the most general, unconstrained de nition
stating that for any pair of objects, the rst is better than the second, is true in exactly 50% of the cases for
every pair hx; yi; x 6= y, as where this is true, the relation will not hold for the pair hy; xi. It will also always
be false for pairs of type hx; xi. This is why the algorithm proceeds by considering an increasing, and strictly
positive number of properties (literals) constraining either object in the relation. The bottom clause contains the
conjunction of n constraints (of type class membership) on the Domain side and the same number of constraints
again on the Range side of the relation. We evaluate all combinations of constraints, except the ones that
imply the same class membership of both arguments (i.e. X is better than Y because they both share the same
property/class membership) or those that have already been considered. An example of the re nement operator
steps starting from a positive example car1 is better than car3 is illustrated in Figure 2 (please see Section 4.1
for a description of the dataset of car preferences).</p>
        <p>We express the coverage of a clause through P and N , where P { the number of positive examples covered,
and N { the number of negative examples covered. In the case that the solution has the same score as another
alternative, Aleph will only return the rst solution found. In our algorithm, we consider all new clauses that
cover at least two positive examples and none of the negatives. (This is done to ensure consistency, and that at
least a minimum level of generalisation takes place.) The search will not stop until all the possible combinations
at each level have been considered. The resulting theory is a disjunction of clauses. Therefore, any test example
covered by one of them is classi ed as positive.</p>
        <p>(Thing) betterthan (Thing)</p>
        <p>(Car) betterthan (Car)
(Manual) betterthan (MediumCar)
(Manual) betterthan (NonHybrid)</p>
        <p>Our algorithm retains all consistent clauses generalised from a given positive example, rather than just one
of them, as Aleph would have done. (At the same time, both algorithms discard the positive example from the
training set after this generalisation step in a greedy learning manner.) This is the most important di erence
between the two algorithms and a possible explanation for any observed di erence in their performance.</p>
        <p>If a consistent clause has not been found yet, the algorithm continues to re ne the current candidate by
adding one literal to constrain either object in the relation. Similarly to Aleph, when APARELL cannot nd a
consistent generalisation, it adds the bottom clause itself to the theory.</p>
        <p>We implement our algorithm using the Closed World Assumption (CWA). For the problem of learning strict
order, it makes virtually no di erence whether our system operates under the CWA or OWA. Under the OWA,
we learn two hypotheses: one as mentioned before, the other with the positive and negative examples swapped.
Algorithm 1 APARELL
1: Input : background knowledge B,
a set of positive examples E+ = fhe1; e2i; : : : ; hen; emig,
a set of negative examples E = fhe2; e1i; : : : ; hem; enig,
maximum number of literals (depth limit ) l
a relation name r = betterthan
2: Output : A theory T represented as a set of clauses, each de ning a pair of concepts specifying the relation's
domain and range
3: set T = ; /* initialization*/
4: set x1 = ; /* the list of constraints on Domain */
5: set x2 = ; /* the list of constraints on Range */
6: set f lag = false /* whether a consistent generalisation of the bottom clause has been found */
7: while E+ is not empty do
8: set clause size = 2 /* initial value for jx1j + jx2j */
9: set x1 = ;, x2 = ;, f lag=false /* reset in every loop*/
10: select e+ 2 E+ to be generalised, and de ne he1; e2i = e+
11: /* build the bottom clause hx1; x2i for e+ as follows: */
12: set x1 = fC1; : : : ; Cmg 8Ci such that e1 2 Ci
13: set x2 = fD1; : : : ; Dng 8Di such that e2 2 Di
14: while f lag==false and clause size &lt; l do</p>
        <p>/* search (top-down) through all generalisations of the bottom clause */
15: for all x01 x1; x02 x2; jx01j + jx02j == clause size do
16: if x01 betterthan x02 is consistent and more general than e+ then
17: add the clause to T
18: set f lag = true
19: end if
20: end for
21: if f lag==false and clause size &lt; l then
22: set clause size = clause size + 1 /* increment clause size */
23: else if f lag==false and clause size == l then
24: add e+ to T and remove e+ from E+
25: end if
26: end while
27: if ag==true then
28: remove covered positive examples from E+
29: end if
30: end while
31: Return T
For the given domain (of learning strict order), the second hypothesis (i.e. the one that being swapped between
the negative and positive examples) is almost the exact negation of the rst (modulo the choice of a training
sample). The resulting coverage is therefore almost identical to the one of the CWA hypothesis.
3.2</p>
      </sec>
      <sec id="sec-3-2">
        <title>Search in a Complex Class Hierarchy</title>
        <p>In the case of a more complex class hierarchy that consists of more than one level, we consider the hypothesis
search over all the inferred class memberships of each individual. For example, if the class hierarchy and its
membership in Figure 3 are given to the system, the algorithm progresses as follows:
1. Select a positive example. As an example, we choose car 1 is better than car 3 to be generalised. Please
see Table 1 for the properties of the cars.
2. Build the bottom clause. In this step, we use the reasoner to infer all the classes which include car 1 and
car 3 as their direct and indirect membership. The bottom class of the above example is shown as follows:
(Car and EconomyCar and FamilyCar and Manual and</p>
        <p>NonHybrid and SmallCar and Suv) betterthan
(Car and Manual and MediumCar and NonHybrid and Sedan)
The above bottom clause is only used to guide the re nement search, as
SmallCar v EconomyCar v Car. We do not consider clauses referring to the membership of the intersection
between a class and its superclass, as this will be the same as the membership of the class itself. For
example, the membership of SmallCar u EconomyCar is fcar1,car8g, which is the same as the membership
of SmallCar itself.
3. Search. The search begins from the shortest clause length considered (2 literals) as shown below:</p>
        <sec id="sec-3-2-1">
          <title>Car betterthan Manual</title>
          <p>Car betterthan MediumCar
Car betterthan NonHybrid
Car betterthan Sedan
EconomyCar betterthan Car
EconomyCar betterthan Manual
...</p>
          <p>In the case that we have not found any consistent hypothesis of length 2, this parameter is increased to 3,
etc.:</p>
        </sec>
        <sec id="sec-3-2-2">
          <title>EconomyCar and FamilyCar betterthan Car EconomyCar and FamilyCar betterthan Manual ...</title>
          <p>It should be noted that the algorithm skips the evaluation of the same value comparisons like:
Car betterthan Car to speed up the search.
4. Remove covered positive examples. We count the coverage for each hypothesis we build and remove
the covered positive examples from the training dataset.
3.3</p>
        </sec>
      </sec>
      <sec id="sec-3-3">
        <title>Current State</title>
        <p>Our algorithm can handle a multi-level class hierarchy, but, in order to limit the search, we only allow conjunctions
of literals in the clauses, e ectively limiting the language to EL description logic. With this limitation, we can still
produce accurate models for our test cases, and a result in a form that is easy to interpret. The most expensive
process is checking the class membership (by using the ontology reasoner) for every possible hypothesis. This is
used for scoring the hypothesis coverage. The search in our algorithm follows breadth- rst search with limited
depth (speci ed as the l parameter).</p>
        <p>Another property of our proposed algorithm is that, unlike Aleph, we do not provide a single consistent clause
covering a given positive example, but add to the resulting theory all such clauses of the minimal possible length.
While this can produce a more accurate result, the learning process takes longer than in Aleph. For example,
given 45 training examples in which each item has 4 attributes, our algorithm takes 409 ms to nish while Aleph
performs faster in just 88 ms.
4
4.1</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Evaluation</title>
      <sec id="sec-4-1">
        <title>Datasets</title>
        <p>
          We use two publicly available preference datasets [
          <xref ref-type="bibr" rid="ref1 ref11">1, 11</xref>
          ]. Both the sushi and the car datasets have 10 items to
rank which leads to 45 preference pairs per user. We take 60 users from each dataset and perform 10-fold cross
validation for each user's individual preferences. The car dataset has 4 attributes: body type, transmission, fuel
consumption and engine size [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ], as shown in Table 1. Despite the di erence in the number of attributes in the
two datasets, we found that the maximum clause length of 4 (in Aleph and in our algorithm) is su cient to
produce consistent hypotheses.
        </p>
        <p>
          The second dataset used in the experiments is about sushi preferences6 which has been published by
Kamishima [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ]. The dataset contains user individual preferences about 10 di erent sushi types. The users
were asked to sort the sushis according to their preferences in an ascending order. For the experiment, we
generate the pairs order from each set of individual preferences. Similar to the previously mentioned car dataset, 45
pairs of sushi preferences are produced for 10 types of sushi (which represents all possible pairs). In this dataset,
there were 5000 users involved in the survey. The sushi preference dataset is quite large when compared to the
car preference dataset as it has 7 attributes describing each type of sushi. With a dataset of such size, we can
test the performance of our algorithm and evaluate how it grows.
        </p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2 Evaluation Method</title>
        <p>The goal of this evaluation is to assess the accuracy of the predictive power of our algorithm to solve the preference
learning problem. To do this, we set up four experiments as below:
1. Assess the accuracy of our algorithm compared to the three baseline algorithms, SVM, Aleph and DT, on
60 users in car dataset and sushi dataset. Here, we also perform an ANOVA test and a post-hoc test to
assess which algorithms signi cantly di er.
2. Assess the accuracy and the performance of our algorithm on a larger dataset by conducting an experiment
on 5000 users of the sushi dataset.
3. Assess the potential of our algorithm to learn relations from a more complex class hierarchy by using the
car dataset.
4. Assess the accuracy of our algorithm on training examples of di erent size, and compare it to the three
baseline algorithms.</p>
      </sec>
      <sec id="sec-4-3">
        <title>Accuracy on 60 users.</title>
        <p>
          We compare our algorithm with three other machine learning algorithms: SVM, the Matlab CART Decision
Tree (DT) learner, and Aleph. SVM is a very common statistical classi cation algorithm, which is used in many
domains. Previous work of pairwise preference learning was performed by Qian et al. [
          <xref ref-type="bibr" rid="ref19">19</xref>
          ] showing that SVM
can also be used to learn in this domain. Both DT and Aleph are learners expressing their models in logic, where
the former uses propositional logic and the latter { (a subset of) First Order Logic (namely, Horn clauses).
        </p>
        <p>We build a simple class hierarchy as explained in Section 2 for each dataset. We learn from each set of
individual preferences and evaluate the model by using 10-fold cross validation. We repeat the same test for
all users then calculate the average accuracy. The result is shown in Table 2. In this experiment, the search
stopped at a maximum length of 4 literals (this is the same as Aleph's default clause length of 4). According to
the ANOVA test with = 0:05, the result shows that there is a signi cant di erence amongst the algorithms
with a p-value of 1:14 10 21 for the car dataset and 2:97 10 3 for the sushi dataset. ANOVA is conceptually
similar to multiple two-sample t-tests, but is more conservative (results in less type I error). After performing the
ANOVA test, we nd which algorithms are signi cantly di erent by using Fisher's Least Signi cant Di erence
(LSD). The result of this post-hoc test is shown in Table 3. Please note that the result of this 10-fold cross
validation may have introduced a bias in terms of absolute performance levels due to the fact that the use of
transitivity chains (closures) may have created an overlap between training and test data. However, the same
training and test data have been used to compare all algorithms, so the results for their comparison remain
una ected (are not biased).</p>
        <p>In this experiment, our algorithm is still showing the highest accuracy amongst the other three baseline
algorithms. The average accuracy on individual preferences from 5000 sushi dataset users is shown in Table 4. In
the rst experiment, we have already evaluated the mean di erence amongst the algorithms using ANOVA and
a post-hoc test. In here, the p-value of ANOVA ( =0.05) test is 0. This is a common case in large datasets.
Performing statistical test to analyze the mean di erences in a large dataset can be problematical as p-value
tends to drop quickly to zero. From the result shown in Table 4, the low standard deviation rate can be a good
indication that the accuracy of our algorithm is excellent in any case.</p>
        <p>As an additional experiment, we evaluate the car dataset with the complex class hierarchy (see Figure 3) using
literal depth limit=4. The accuracy of our algorithm is performed using 10-fold cross validation. The result
of the experiment with the complex class hierarchy is a mean of 0.8409 and standard deviation of 0.1405. The
accuracy result shows no di erence between using a simple class hierarchy and a complex class hierarchy.</p>
      </sec>
      <sec id="sec-4-4">
        <title>Accuracy on di erent training examples size.</title>
        <p>We perform several experiments with the algorithms by varying the proportion of training examples and test it
on 10% of examples. For a more robust result, we validate each cycle with 10-fold cross validation. The result
of these experiments is shown in Figure 4. We show that our algorithm still works better even with the smallest
number of training examples.</p>
      </sec>
      <sec id="sec-4-5">
        <title>Algorithm performance.</title>
        <p>We run another test to see the algorithm performance by di erent clause length settings (see Table 5). The
evaluation is performed on both datasets by using a 90:10% training-test split. We also record the algorithm
execution times of 60 users 90% of examples (2,382 total examples in car dataset and 2400 total examples in
sushi dataset, please note that some of the positive examples are removed during the search). The algorithm
was executed on Java 8 Eclipse IDE with 8 GB Memory 1867 MHz DDR3 and 2.9 GHz Processor Intel Core
i5 machine. The result shows there is no signi cant accuracy improvement after 4 literals. Surprisingly, the
algorithm runs very slow at 2 literals for the car dataset (where the achieved accuracy is the lowest for all tested
clause lengths). The reason is in Step 4 i.e. the removal of covered positive examples (see Section 3). If we
cannot nd any consistent hypothesis when generalising a positive example, we add an exception and only remove
one example. But when the algorithm can nd consistent hypotheses, it removes more than one example which
results in much fewer positive examples thus speeding up the search. This anomaly does not occur in the sushi
dataset due to the larger number of attributes so that the possibility to nd a consistent hypothesis for clause
length 2 is higher.</p>
        <p>
          As mentioned in [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ], by using DL representation, which is variable-free syntax, our algorithm can produce a
set of solutions which are more readable compared to the First Order Logic representation. An example of a
consistent hypothesis found by our algorithm is shown below:
(Manual) betterthan (Sedan)
(Automatic and Hybrid) betterthan (MediumCar and Suv)
        </p>
        <p>The above rules simply say that: \Manual car is better than sedan car OR automatic hybrid car is better
than medium size SUV car." In comparison, Aleph produces rules, such as:
betterthan(A,B) :- hasenginesize(B,large), hasenginesize(A,small).
betterthan(A,B) :- hasfuelcons(B,nonhybrid), hasbodytype(B,sedan).</p>
        <p>The rst Aleph's rule states that: \car A is better than car B if car B has a large engine size and car A has a
small engine size." In the second rule, it says:\car A is better than car B if car B is a non hybrid car and sedan."
In other words, if the car has characteristics as a non-hybrid sedan car, it will not be better than any other car.
5</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Conclusion and Further Work</title>
      <p>In this paper, we have shown that the implementation of ILP in DL can be useful to learn a user's preference
from pairwise comparisons. In fact, the proposed algorithm has proven statistically signi cantly better than all
tested alternatives in all but one case. The exception in question is when compared to SVM on the car dataset,
where our algorithm achieves a seemingly higher mean accuracy, but the result is not statistically signi cant (in
other words, it is a draw).</p>
      <p>We are currently working to address some of the limitations of the algorithm. In terms of accuracy, we show
that our algorithm outperformed the other algorithms. However, to produce the nal theory, our algorithm
takes much longer than each of the three baseline algorithms. Currently, the only way to restrict the search
complexity is by limiting its depth. We need to work on the improvement of this aspect without reducing the
accuracy performance. The possible alternatives include the use of heuristic functions to predict the cost of
expanding a node (i.e. informed search). We are also working on expanding the re nement operator by allowing
the use of di erent operators (e.g. union or negation) with the aim of improving the accuracy.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>Abbasnejad</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sanner</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bonilla</surname>
            ,
            <given-names>E. V.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Poupart</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Learning Community-based Preferences via Dirichlet Process Mixtures of Gaussian Processes</article-title>
          .
          <source>In: Proceedings of the 23rd International Joint Conference on Arti cial Intelligence (IJCAI)</source>
          (
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <surname>Baader</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Horrocks</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Sattler</surname>
            ,
            <given-names>U.</given-names>
          </string-name>
          :
          <article-title>Description Logics</article-title>
          .
          <source>In: Handbook of Knowledge Representation</source>
          , pp.
          <volume>135</volume>
          {
          <fpage>179</fpage>
          . Springer, Heidelberg (
          <year>2009</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>Balakrishnan</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Chopra</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Two of a kind or the ratings game? Adaptive pairwise preferences and latent factor models</article-title>
          .
          <source>In: Proceedings of IEEE 10th International Conference on Data Mining (ICDM)</source>
          , pp.
          <volume>725</volume>
          {
          <issue>730</issue>
          (
          <year>2010</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <surname>Bradley</surname>
            ,
            <given-names>R. A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Terry</surname>
            ,
            <given-names>M. E.</given-names>
          </string-name>
          :
          <article-title>The rank analysis of incomplete block designs. I. The method of paired comparisons</article-title>
          .
          <source>In: Biometrika</source>
          , vol
          <volume>39</volume>
          , no.
          <issue>3-4</issue>
          , pp.
          <volume>324</volume>
          {
          <issue>345</issue>
          (
          <year>1952</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>Fanizzi</surname>
          </string-name>
          , N.,
          <string-name>
            <surname>d'Amato</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Esposito</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          :
          <article-title>DL-FOIL Concept Learning in Description Logics</article-title>
          .
          <source>In: Proceedings of Inductive Logic Programming, ser. LNCS</source>
          , vol.
          <volume>5194</volume>
          , pp.
          <volume>107</volume>
          {
          <fpage>121</fpage>
          . Springer, Heidelberg (
          <year>2008</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>Gulwani</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hernandez-Orallo</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kitzelmann</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Muggleton</surname>
            ,
            <given-names>S.H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Schmid</surname>
            ,
            <given-names>U.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Zorn</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          :
          <article-title>Inductive Programming Meets the Real World</article-title>
          .
          <source>In: Communications of the ACM</source>
          , vol.
          <volume>58</volume>
          no.
          <issue>11</issue>
          , pp.
          <volume>90</volume>
          {
          <fpage>99</fpage>
          .
          <string-name>
            <surname>ACM</surname>
          </string-name>
          , New York (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>Horridge</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Bechhofer</surname>
            ,
            <given-names>S.:</given-names>
          </string-name>
          <article-title>The OWL API: A Java API for OWL ontologies</article-title>
          .
          <source>In: Semantic Web</source>
          , vol.
          <volume>2</volume>
          no.
          <issue>1</issue>
          , pp.
          <volume>11</volume>
          {
          <fpage>21</fpage>
          , (
          <year>2011</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8] Hullermeier, E., Furnkranz, J., Cheng, W., and
          <string-name>
            <surname>Brinker</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          :
          <article-title>Label ranking by learning pairwise preferences</article-title>
          .
          <source>In: Arti cial Intelligence</source>
          , vol.
          <volume>172</volume>
          , no.
          <issue>16</issue>
          , pp.
          <year>1897</year>
          {
          <year>1916</year>
          (
          <year>2008</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <surname>Iannone</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Palmisano</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fanizzi</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          :
          <article-title>An Algorithm Based on Counterfactuals for Concept Learning in The Semantic Web</article-title>
          .
          <source>In: Applied Intelligence</source>
          , vol.
          <volume>26</volume>
          no.
          <issue>2</issue>
          , pp.
          <volume>139</volume>
          {
          <fpage>159</fpage>
          . Springer, Heidelberg (
          <year>2007</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>Jensen</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Saez</surname>
            <given-names>Gallego</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            and
            <surname>Larsen</surname>
          </string-name>
          ,
          <string-name>
            <surname>J.:</surname>
          </string-name>
          <article-title>A predictive model of music preference using pairwise comparisons</article-title>
          .
          <source>In: Proceedings of IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)</source>
          , pp.
          <year>1977</year>
          {
          <year>1980</year>
          (
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>Kamishima</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>Nantonac Collaborative Filtering: Recommendation Based on Order Responses</article-title>
          .
          <source>In: Proceedings of the Ninth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</source>
          , pp.
          <volume>583</volume>
          {
          <fpage>588</fpage>
          .
          <string-name>
            <surname>ACM</surname>
          </string-name>
          , New York (
          <year>2003</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>Kietz</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          :
          <article-title>Learnability of Description Logic Programs</article-title>
          .
          <source>In: Proceedings of Inductive Logic Programming, ser. LNCS</source>
          , vol.
          <volume>2583</volume>
          , pp.
          <volume>117</volume>
          {
          <fpage>132</fpage>
          . Springer, Heidelberg (
          <year>2002</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>Konstantopoulos</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Charalambidis</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Formulating Description Logic Learning as An Inductive Logic Programming Task</article-title>
          .
          <source>In: Proceedings of IEEE World Congress on Computational Intelligence</source>
          , pp.
          <volume>1</volume>
          {
          <issue>7</issue>
          (
          <year>2010</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <surname>Law</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Russo</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Broda</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          :
          <article-title>Learning weak constraints in answer set programming</article-title>
          .
          <source>In: Theory and Practice of Logic Programming</source>
          , vol.
          <volume>15</volume>
          no.
          <issue>4-5</issue>
          , pp.
          <volume>511</volume>
          {
          <issue>525</issue>
          (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <surname>Lehmann</surname>
            ,
            <given-names>J.:</given-names>
          </string-name>
          <article-title>DL-Learner: Learning Concepts in Description Logics</article-title>
          .
          <source>In: The Journal of Machine Learning Research</source>
          , vol.
          <volume>10</volume>
          , pp.
          <volume>2639</volume>
          {
          <fpage>2642</fpage>
          . (
          <year>2009</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <surname>Linden</surname>
            ,
            <given-names>G. D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jacobi</surname>
            ,
            <given-names>J. A.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Benson</surname>
            ,
            <given-names>E. A.</given-names>
          </string-name>
          :
          <article-title>Collaborative recommendations using item-to-item similarity mappings</article-title>
          .
          <source>In: US Patent 6</source>
          ,
          <issue>266</issue>
          ,
          <issue>649</issue>
          (
          <year>2001</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <surname>Muggleton</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Inverse Entailment and Progol</article-title>
          .
          <source>In: New Generation Computing</source>
          , vol.
          <volume>13</volume>
          no.
          <issue>3-4</issue>
          , pp.
          <volume>245</volume>
          {
          <issue>286</issue>
          (
          <year>1995</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <surname>Muggleton</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>De Raedt</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Poole</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          , Bratko, I.,
          <string-name>
            <surname>Flach</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Inoue</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Srinivasan</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>ILP Turns 20</article-title>
          .
          <source>In: Machine Learning</source>
          , vol.
          <volume>86</volume>
          no.
          <issue>1</issue>
          , pp.
          <volume>3</volume>
          {
          <fpage>23</fpage>
          . Springer, Heidelberg (
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <surname>Qian</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gao</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Jagadish</surname>
          </string-name>
          , H.:
          <article-title>Learning user preferences by adaptive pairwise comparison</article-title>
          .
          <source>In: Proceedings of the VLDB Endowment</source>
          , vol.
          <volume>8</volume>
          no.
          <issue>11</issue>
          , pp.
          <volume>1322</volume>
          {
          <issue>1333</issue>
          (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <surname>Rokach</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Kisilevich</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Initial pro le generation in recommender systems using pairwise comparison</article-title>
          .
          <source>In: IEEE Transactions on Systems, Man, and Cybernetics</source>
          , Part C:
          <article-title>Applications</article-title>
          and Reviews, vol.
          <volume>42</volume>
          no.
          <issue>6</issue>
          , pp.
          <year>1854</year>
          {
          <year>1859</year>
          (
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <surname>Srinivasan</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>The Aleph Manual</article-title>
          .
          <source>In: Technical Report. Computing Laboratory</source>
          , Oxford University (
          <year>2000</year>
          ), http://web.comlab.ox.ac.uk/oucl/research/areas/machlearn/Aleph/
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>