<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Metric Learning for Value Alignment</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Andrea Loreggia</string-name>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Nicholas Mattei</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Francesca Rossi</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>K. Brent Venable</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>IBM Research, IBM T.J. Watson Research Center</institution>
          ,
          <addr-line>Yorktown Heights, NY</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Institute for Human and Machine Cognition (IHMC)</institution>
          ,
          <addr-line>Pensacola, FL</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Tulane University, Department of Computer Science</institution>
          ,
          <addr-line>New Orleans, LA</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>University of Padova, Department of Mathematics</institution>
          ,
          <addr-line>Padova</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Preference are central to decision making by both machines and humans. Representing, learning, and reasoning with preferences is an important area of study both within computer science and across the social sciences. When we give our preferences to an AI system we expect the system to make decisions or recommendations that are consistent with our preferences but the decisions should also adhere to certain norms, guidelines, and ethical principles. Hence, when working with preferences it is necessary to understand and compute a metric (distance) between preferences - especially if we encode both the user preferences and ethical systems in the same formalism. In this paper we investigate the use of CP-nets as a formalism for representing orderings over actions for AI systems. We leverage a recently proposed metric for CP-nets and a neural network architecture, CPMETRIC, for computing this metric. Using these two tools we look at the how one can build a fast and flexible value alignment system.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>Preferences are central to individual and group decision
making by both computer systems and humans. Due to this
central role in decision making the study of representing
[Rossi et al., 2011], learning [Fürnkranz and Hüllermeier,
2010], and reasoning [Domshlak et al., 2011] with
preferences is a focus of study within computer science and in
many other disciplines including psychology and sociology
[Goldsmith and Junker, 2009]. Individuals express their
preferences in many different ways: pairwise comparisons,
rankings, approvals (likes), positive or negative examples, and
many more examples are collected in various libraries and
databases [Mattei and Walsh, 2013; Mattei and Walsh, 2017;
Bache and Lichman, 2013]. A core task in working with
preferences is understanding the relationship between
preferences. This often takes the form of a dominance query,
i.e., which item is more or most preferred, or distance
measures, i.e., which object is the closest to my stated preference.
These types of reasoning are important in many domains
including recommender systems [Fattah et al., 2018],
collective decision making [Brandt et al., 2016], and value
alignment systems [Russell et al., 2015; Loreggia et al., 2018c;
Loreggia et al., 2018b].</p>
      <p>Using a formal structure to model preferences, especially
one that directly models dependency, can be useful for
reasoning. For example, it can support reasoning based on inference
and causality, and provide more transparency and
explainability since the preferences are explicitly represented and hence
scrutable [Kambhampati, 2019]. A number of compact
preference representation languages have been developed in the
literature for representing and reasoning with preferences; see
the work of Amor et al. [2016] for a survey of compact
graphical models; we specifically focus on conditional preference
networks (CP-nets) [Boutilier et al., 2004].</p>
      <p>CP-nets are a compact graphical model used to capture
qualitative conditional preferences over features (variables)
[Boutilier et al., 2004]. Qualitative preferences are important
as there is experimental evidence that qualitative preferences
may more accurately reflect humans’ preferences in uncertain
information settings [Popova et al., 2013; Allen et al., 2015].
CP-nets are a popular formalism for specifying preferences in
the literature and have been used for a number of applications
including recommender systems and product specification
[Pu et al., 2011; Fattah et al., 2018]. Consider a car that is
described by values for all its possible features: make, model,
color, and stereo options. A CP-net consists of a dependency
graph and a set of statements of the form, “all else being equal,
I prefer x to y.” For example, in a CP-net one could say “Given
that the car is a Honda Civic, I prefer red to yellow.”, the
condition sets the context for the preference over alternatives.
These preferences are qualitative, i.e., there is no quantity
expressing the degree of preference.</p>
      <p>A CP-net induces an ordering over all possible outcomes,
i.e., all complete assignments to the set of features. This is a
partial order if the dependency graph of the CP-net is acyclic,
i.e., the conditionality of the statements does not create a cycle,
as is often assumed in work with CP-nets [Goldsmith et al.,
2008]. The size of the description of the CP-net may be
exponentially smaller than the partial order it describes. Hence,
CP-nets are called a compact representation and reasoning and
learning on the compact structure, instead of the full order,
is an important topic of research. Recent work proposes the
first formal metric to describe the distance between CP-nets
[Loreggia et al., 2018a] and the related formalism of LP-trees
[Li and Kazimipour, 2018] in a rigorous way. What is
important is not the differences in the surface features of the
CP-nets, e.g., a single statement or dependency, but rather the
distance between their induced partial orders. Even a small
difference in a CP-net could generate a very different partial
order, depending on which feature is involved in the
modification. While the metrics proposed by Loreggia et al. [2018a]
are well grounded, they are computationally hard to compute,
in general, and approximations must be used.</p>
      <p>We envision the use of CP-nets to solve one part of the
value alignment problem [Russell et al., 2015; Loreggia et
al., 2018c; Loreggia et al., 2018b]. This is part of a broader
research program we call Ethically Bounded AI, that seeks
to understand how to harness the power of AI yet prevent
these systems from making choices we do not consider ethical
[Rossi and Mattei, 2019]. In this work we envision using a
distance metric to measure the difference between an individual
agent’s preferences over actions and another ordering given by,
e.g., ethics, norms, or business values [Loreggia et al., 2018c;
Loreggia et al., 2018b].</p>
      <p>Following work in metric learning over structured
representations [Bellet et al., 2015], we wish to learn the distance
between partial orders represented compactly as CP-nets. We
do not want to work with the partial orders directly as they
may be exponentially larger than the CP-net representation.
Informally, given two CP-nets, we wish to estimate the distance
between their induced partial orders using a neural network.
The number of possible CP-nets grows extremely fast, from
481,776 for 4 binary features to over 5:24 1040 with 7
binary features [Allen et al., 2017]. However, the computation
time of the approximation algorithm proposed by Loreggia et
al. [2018a] scales linearly with the number of features, hence,
new methods must be explored. Therefore, leveraging the
inferential properties of neural networks may help us make
CP-nets more useful as a preference reasoning formalism.
Contributions. We propose using metric learning as a tool
to practically solve aspects of the value alignment problem. We
propose to model both user preferences and ethical priorities
over actions using the CP-net formalism and we demonstrate
how one can use a state of the art neural network
formulation, CPMETRIC, to quickly and accurately judge distances
between preferred actions and ethical actions. We evaluate
our models and metrics on generated CP-nets and show that
CPMETRIC leads to a speed up in computation while still
being accurate.
2</p>
    </sec>
    <sec id="sec-2">
      <title>CP-nets</title>
      <p>Conditional Preference networks (CP-nets) are a graphical
model for compactly representing conditional and
qualitative preference relations [Boutilier et al., 2004]. CP-nets are
comprised of sets of ceteris paribus preference statements
(cp-statements). For instance, the cp-statement, “I prefer red
wine to white wine if meat is served," asserts that, given two
meals that differ only in the kind of wine served and both
containing meat, the meal with red wine is preferable to the
meal with white wine. CP-nets have been extensively used
in the preference reasoning preference learning and social
choice literature as a formalism for working with
qualitative preferences [Domshlak et al., 2011; Rossi et al., 2011;
Brandt et al., 2016]. CP-nets have even been used to
compose web services [Wang et al., 2009] and other decision aid
systems [Pu et al., 2011].</p>
      <p>Formally, a CP-net has a set of features (or variables)
F = fX1; : : : ; Xng with finite domains D(X1); : : : ; D(Xn).
For each feature Xi, we are given a set of parent features
P a(Xi) that can affect the preferences over the values of Xi.
This defines a dependency graph in which each node Xi has
P a(Xi) as its immediate predecessors. An acyclic CP-net is
one in which the dependency graph is acyclic. Given this
structural information, one needs to specify the preference over the
values of each variable Xi for each complete assignment to the
parent variables, P a(Xi). This preference is assumed to take
the form of a total or partial order over D(Xi). A cp-statement
for some feature Xi that has parents P a(Xi) = fx1; : : : ; xng
and domain D(Xi) = fa1; : : : ; amg is a total ordering over
D(Xi) and has general form: x1 = v1; x2 = v2; : : : ; xn =
vn : a1 : : : am, where for each Xi 2 P a(X1) : xi = vi
is an assignment to a parent of Xi with vi 2 D(Xi). The set
of cp-statements regarding a certain variable Xi is called the
cp-table for Xi.</p>
      <p>Consider the CP-net depicted graphically in Figure 1 (left)
with features are A, B, C, and D. Figure 1 (right) gives the
full induced preference order for the CP-net. Each variable
has binary domain containing f and f if F is the name of the
feature. All cp-statements in the CP-net are: a a, b b,
(a ^ b) : c c, (a ^ b) : c c, (a ^ b) : c c, (a ^ b) : c c,
c : d d, c : d d. Here, statement a a represents
the unconditional preference for A = a over A = a, while
statement c : d d states that D = d is preferred to D = d,
given that C = c. The semantics of CP-nets depends on the
notion of a worsening flip: a change in the value of a variable
to a less preferred value according to the cp-statement for that
variable. For example, in the CP-net above, passing from abcd
to abcd is a worsening flip since c is better than c given a
and b. One outcome is preferred to or dominates another
outcome (written ) if and only if there is a chain of
worsening flips from to . This definition induces a preorder
over the outcomes, which is a partial order if the CP-net is
acyclic [Boutilier et al., 2004], as depicted in Figure 1 (right).</p>
      <p>The complexity of dominance and consistency testing in
CP-nets is an area of active study in preference reasoning
[Goldsmith et al., 2008; Rossi et al., 2011]. Finding the
optimal outcome of a CP-net is NP-hard [Boutilier et al., 2004] in
general but can be found in polynomial time for acyclic
CPnets by assigning the most preferred value for each cp-table.
Indeed, acyclic CP-nets induce a lattice over the outcomes as
(partially) depicted in Figure 1 (right). The induced preference
ordering, Figure 1 (right), can be exponentially larger than the
CP-net Figure 1 (left), which motivates learning a metric using
only the (more compact) CP-net.
3</p>
    </sec>
    <sec id="sec-3">
      <title>Preferences and Ethical Priorities</title>
      <p>In what follow we assume that we can describe a user’s
behavior through her preferences over the features of the domain.
This can be modeled as a CP-net and it gives us an ordering
over the actions the user would like to take. The example given
in Loreggia et al. [2018c] concerns a driver of a vehicle: they
a &gt; a A
b &gt; b B</p>
      <p>C
(a ^ b)
(a ^ b)
(a ^ b)
(a ^ b)
c &gt; c
c &gt; c
c &gt; c
c &gt; c</p>
      <p>D c
c
d &gt; d
d &gt; d
abcd
abcd
abcd
Least Preferred
abcd
abcd</p>
      <p>abcd
abcd
abcd
abcd</p>
      <p>abcd
abcd</p>
      <p>abcd
abcd
abcd
abcd
abcd
may want to go as fast as possible and run over certain small
animals. However, there may be some overall ethical or moral
guidelines (priorities) the system must follow. In this case we
want to evaluate the difference, or distance, between the
individual user and the society. This idea of morality as ordering
over actions was first proposed by Sen [1974].</p>
      <p>To operationalize this system we wish to describe both
preferences and priorities using the CP-net formalism and using a
notion of distance in the metric space of CP-nets. This enables
us and the system to understand whether users’ preferences
are close enough to the moral principles or not. Eventually,
when preferences deviate from the desired behavior, we can
use CP-nets, since they induce an ordering, to find a
tradeoff so that the quality of the decision with respect to the
subjective preferences does not significantly degrade when
conforming to the ethical principles [Loreggia et al., 2018c;
Loreggia et al., 2018b]. In this way we have bounded the
behavior of the system to be ethical while still being responsive
to the user preferences [Rossi and Mattei, 2019].</p>
      <p>Traditional reasoning and learning approaches in AI provide
different and complementary capabilities to an autonomous
agent. Symbolic and logical reasoning allow these agents to
manipulate symbols and perform inference, while machine
learning techniques can learn and optimize many ill-defined
problems from large amounts of data. As ongoing work, we
intend to study the use of both kinds of approaches to model,
learn, and reason with both preferences and ethical
principles. Inspired by the System 1 and System 2 theory of Daniel
Kahneman [Kahneman, 2011], we will define a dual-agent
architecture that will provide autonomous agents with the
ability to combine symbolic and accurate reasoning with data
interpretation and learning, for both preferences and ethical
principles. This will allow machines to be flexible and
contextdependent in how they handle and combine these two sources
of information for decision making [Rossi and Loreggia, 2019;
Rossi and Mattei, 2019].</p>
      <p>The combined use of deep learning techniques and
logical reasoning formalisms is an exciting research direction to
find a principled ways to develop AI systems that are both
accountable and able to explain themselves. We hope that these
approaches will be able to overcome limitations of the
“blackbox paradigm” in the machine learning discipline [Rossi and
Loreggia, 2019].
4</p>
    </sec>
    <sec id="sec-4">
      <title>Metric Learning on CP-nets</title>
      <p>Metric learning algorithms aim to learn a metric (or distance
function) over a set of training points or samples [Sohn, 2016].
The importance of metrics has grown in recent years with the
use of these functions in many different domains: from
clustering to information retrieval and from recommender systems
to preference aggregation. For instance, many clustering
algorithms like the k-Means or classification algorithm including
k-Nearest Neighbor use a distance value between points.</p>
      <p>Formally, a metric space is a pair (M; d) where M is a set
of elements and d is a function d : M M ! R where d
satisfies four criteria. Given any three elements A; B; C 2 M ,
d must satisfy
1. (1) d(A; B) 0, there must be a value for all pairs;
2. (2) d(A; B) = d(B; A), d must be symmetric;
3. (3) d(A; B) d(A; C) + d(C; B); d must satisfy the
triangle inequality; and
4. (4) d(A; B) = 0 if and only if A = B; d can be zero if
and only if the two elements are the same.</p>
      <p>Xing et al. [2002] first formalized the problem of metric
learning, i.e., learning the metric directly from samples rather
than formally specifying the function d. This approach requires
training data, meaning that we have some oracle that is able
to give the value of the metric for each pair. The success of
deep learning in many different domains [Krizhevsky et al.,
2012] has lead many researchers to apply these approaches to
the field of metric learning, resulting in a number of important
results [Bellet et al., 2015].</p>
      <p>
        In this work we focus on metric spaces (M , d) where M
is a set of CP-nets. Given this, we want to learn the distance
d which best approximates the Kendall tau distance (KTD)
[Kendall, 1938] between the induced partial orders. Informally,
the Kendall tau distance between two orderings is the number
of pairs that are discordant, i.e., not ordered in the same way,
in both orderings. This distance metric extended to partial
orders (Definition 1) was shown to be a metric on the space of
CP-nets by Loreggia et al. [2018a]. To extend the classic KTD
to CP-nets, a penalty parameter p defined for partial rankings
[Fagin et al., 2006] was extended to the case of partial orders.
Loreggia et al. [2018a] assume that all CP-nets are acyclic
and in minimal (non-degenerate) form, i.e., all arcs in the
dependency graph have a real dependency expressed in the
cp-statements, a standard assumption in the CP-net literature
        <xref ref-type="bibr" rid="ref2 ref6">(see e.g., [Allen et al., 2017; Boutilier et al., 2004])</xref>
        .
Definition 1. Given two CP-nets A and B inducing partial
orders P and Q over the same unordered set of outcomes
U : KT D(A; B) = KT (P; Q) = P8i;j2U;i6=j Kip;j (P; Q)
where i and j are two outcomes with i 6= j (i.e., iterate over
all unique pairs), we have:
1. Kip;j (P; Q) = 0 if i; j are ordered in the same way or are
incomparable in P and Q;
2. Kip;j (P; Q) = 1 if i; j are ordered inversely in P and Q;
3. Kip;j (P; Q) = p, 0:5 p &lt; 1 if i; j are ordered in P
and incomparable in Q (resp. Q; P ).
      </p>
      <p>To make this distance scale invariant, i.e., a value in [0; 1],
it is divided by jU j.</p>
      <p>CP-nets present two important challenges when used for
metric learning. The first is that we are attempting to learn a
metric via a compact representation of a partial order. We are
not learning over the partial orders induced by the CP-nets
directly, as they could be exponentially larger than the CP-nets.
The second challenge is the encoding of the graphical
structure itself. Graph learning with neural networks is still a active
and open area of research; Goyal and Ferrara [2017] give a
complete survey of recent work as well as a Python library of
implementations for many of these techniques. Most of these
works focus on finding good embeddings for the nodes of the
network and then using collections of these learned
embeddings to represent the graph for, e.g., particular segmentation
or link prediction tasks. None of these techniques have been
applied to embedding graphs for metric learning.
5</p>
    </sec>
    <sec id="sec-5">
      <title>Structure of CPMETRIC</title>
      <p>In our task the metric space is (M; d) where M is a set of
compact, graphical preferences that induce a partial order
and our goal is to learn the metric d only from the compact,
graphical representation. The key challenge is the need to
find a vector representation of not only the graph but the
cp-statement. We briefly define the representations used for
CPMETRIC here, for a complete overview, see Loreggia et
al. [2019].</p>
      <p>We represent a CP-net I over m using two matrices. First is
the adjacency matrix adjI which represents the dependency
graph of the CP-net and is a m m matrix of 0s and 1s. The
second matrix represents the list of cp-statements cptI , which
is a m 2m 1 matrix, where each row represents a variable
Xi 2 F and each column represents a complete assignment
for each of the variables in F n Xi. The list is built following
a topological ordering of variables in the CP-net. Each cell
cptI (i; j) stores the preference value for the ith variable given
the jth assignment to variables in F n Xi.</p>
      <p>The set of training examples X = fx1; : : : ; xng is made
up of pairs of CP-nets represented through their normalized
Laplacians and the cp-statements. The set of corresponding
labels Y = fy1; : : : ; yngT , where each yi 2 Y; yi 2 [0; 1]
is the normalized value of KTD between the CP-nets in xi.
Each xi 2 X is then a tuple (LA; cptA; LB; cptB)
representing a pair of CP-net (A; B) by their Laplacian, LA, and the
encoding of their cp-statements, cptA.
6</p>
    </sec>
    <sec id="sec-6">
      <title>Experiment</title>
      <p>CPMETRIC is trained to learn the KTD metric by varying the
number of features of the CP-nets n 2 f3; : : : ; 7g and using
two different autoencoders. For a complete discussion of the
performance of CPMETRIC on the standard classification and
regression tasks in comparison with I-CPD, including a
discussion of tuning hyperparameters see Loreggia et al. [2019].
In this paper we focus on what we call the comparison task
that is necessary for value alignment systems. Informally, this
is the task where given two CP-nets, say one representing the
preferences of the user and one representing a set of ethical
values or constraints, we want to decide if some third CP-net,
say the course of action to be taken, is closer to the preferences
or closer to the constraints.
6.1</p>
      <p>Data Generation and Training
For each number of features n 2 f3; : : : ; 7g we generate 1000
CP-nets uniformly at random using the generators from Allen
et al. [Allen et al., 2017]. This set of CP-nets is split into a
training-generative-set (900 CP-nets) and test-generative-set
(100 CP-nets) 10 different ways to give us 10 fold cross
validation. For each fold we compute the training and test dataset
comprised of all, e.g., 9020 , possible pairs of CP-nets from
the training-generative-set and test-generative-set, respectively,
along with the value of KTD for that pair. While we generate
the CP-nets themselves uniformly at random observe that this
creates an unbalanced set of distances – it induces a normal
distribution – and hence our sets are unbalanced. Figure 2
shows the distribution of of CP-net pairs over 20 intervals for
all CP-nets generated for n 2 f3; : : : ; 7g. While our
classification experiments are for m = 10 classes, dividing the interval
into 20 classes provides a better visualization of the challenge
of obtaining training samples at the edges of the distribution.</p>
      <p>We ran a preliminary experiment on balancing our dataset
by sub-sampling the training and test datasets. In these
small experiments, performance was much worse than
perDistribution of CP-net Pairs per Classes
140000
0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20</p>
      <p>Intervals
formance on the unbalanced dataset. Because we are
learning a metric, for each CP-net A, there is only one CP-net
B such KT D(A; B) = 1 and only one CP-net C such
KT D(A; C) = 0. Consequently, attempting to balance or
hold out CP-nets from test or train can lead to poor
performance. We conjecture that in order to improve this task we
should perform some kind of data augmentation, but this
would introduce more subjective assumptions on how and
where data should be augmented [Wong et al., 2016].</p>
      <p>All training was done on a machine with 2 x Intel(R)
Xeon(R) CPU E5-2670 @ 2.60GHz and one NVidia K20
128GB GPU. We train CPMETRIC for 70 epochs using the
Adam optimizer [Kingma and Ba, 2014]. For each number of
features of the CP-net n we use all 9020 pairs in the
trainingset. There are only 488 binary CP-nets with 3 features [Allen
et al., 2017], hence, for n = 3 the training-set is 17K samples
while for n &gt; 3 the number of samples in the training-set
is 800K. Both the Autoencoder and Siamese Autoencoder
models are trained for 100 epochs using the Adam optimizer
[Kingma and Ba, 2014] using the same training-set. Model
weights from the best performing epoch are saved and
subsequently transferred to the deep neural network used to learn
the distance function.</p>
      <p>The training and validation loss for the autoencoder is
shown in Figure 3. Observe that the loss for the CPT
representation approaches zero after only 3 epochs for both the
training and validation phases. The same trend is true for the
adjacency matrix, though the loss converges to 0:15.
For many applications we are not concerned with the true
value of the distance but rather deciding which of two
preferences is closer to a reference point. For example, in product
recommendation we may want to display the closer of two
objects and not care about computing the values [Pu et al.,
2011]. Formally, the qualitative comparison takes takes a set of
CP-nets triples (A; B; C), where A is a reference CP-net and
the task is to decide which other CP-net B or C is closer to
A. We generate uniformly at random 1000 triples of CP-nets
for each n 2 f3; : : : ; 7g. For each triple (A; B; C) we
compute both KTD(A; B) and KTD(A; C) to establish ground
truth and use our regression networks to predict the distance
between (A; B) and (A; C).</p>
      <p>Table 1 displays the accuracy, as a percentage out of 1000
trials, of our three CPMETRIC architectures versus I-CPD for
this task; Table 2 gives the average runtime per pair, averaged
over all 1000 trials. The standard deviations in Table 1 are
across the 10 folds of the training/test set. For all of our
networks we obtain an accuracy above 85% and all the networks
perform about the same on this task ( 2:0%) and the trend for
accuracy is flat across the size of the CP-nets. It is interesting
to note that on this task neither of the autoencoders were able
to significantly improve performance as they did for the
quantitative comparison tasks. While the results are inconclusive,
as all instances of CPMETRIC performed about the same, it
will be interesting to see if there are autoencoder architectures
that are more suited to the comparison task.</p>
      <p>A positive take away is that, as Table 2 shows, we achieve
a sub-linear increase in inference time for our model. I-CPD
scales linearly with the description size of the CP-net so the
neural network does, after training, offer the ability to, in a
practical amount of time, compare CP-nets of larger sizes.
This gives us hope that while the metric itself is NP-hard to
compute in a direct way, we can use the power of deep learning
to enable systems that could be practically deployed.</p>
      <p>Unfortunately the trend for accuracy is negative when the
number of features increases. However, this is also the case for
the I-CPD approximation and both metrics seem to be losing
accuracy at about the same rate. The loss in accuracy for the
neural network models could be caused by the unbalanced
nature of the training and testing datasets. Again, as shown
in Figure 2, generating the CP-nets themselves uniformly at
random does not give us a uniform distribution over distances
and correcting this may give better performance.</p>
      <p>Our conjecture for the slight disadvantage for CPMETRIC
over I-CPD has to do with the directionality of the errors.
When training the network we are optimizing for accuracy on
the regression task. However, when using this metric it does
not matter if CPMETRIC overestimates by a small amount
or underestimates by some small amount. However, when
looking at the comparison task, it may matter a lot if the
direction of our errors is random. An important direction for
future work is to try different optimization objectives when
training the network to see if this bias is the reason for the
underperformance.
7</p>
    </sec>
    <sec id="sec-7">
      <title>Conclusion</title>
      <p>In this paper we have discussed how to use CPMETRIC, a
novel neural network model to learn a metric (distance)
function between partial orders induced from a CP-net, a compact,
structured preference representation, to enable practical value
alignment systems. To our knowledge this is the first use of
neural networks to learn and measure preferences for the value
alignment problem. We feel that this is an interesting and
fruitful direction for research in the AI Safety domain as we
must develop practical and efficient tools that can be used to
effectively harness the power of AI systems.</p>
      <p>Important directions for future work include integrating
novel graph learning techniques to our networks and
extending our work to other formalisms including, e.g., PCP-nets
[Cornelio et al., 2013] and LP-trees [Li and Kazimipour, 2018].
PCP-nets are a particularly interesting direction as they have
been proposed as an efficient way to model uncertainty over
the preferences of a single or multiple agents [Cornelio et al.,
2015]. Another important extension involves setting contexts
for different preference and ethical priority encodings.
CPnets and many other preference formalisms model a particular
domain but do not give us any insight into when we may need
to pass between one context and another.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [Allen et al.,
          <year>2015</year>
          ]
          <string-name>
            <given-names>T. E.</given-names>
            <surname>Allen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Goldsmith</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Mattei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Popova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Regenwetter</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Rossi</surname>
          </string-name>
          , and
          <string-name>
            <given-names>C.</given-names>
            <surname>Zwilling</surname>
          </string-name>
          .
          <article-title>Beyond theory and data in preference modeling: Bringing humans into the loop</article-title>
          .
          <source>In Proc. 4th ADT</source>
          ,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [Allen et al.,
          <year>2017</year>
          ]
          <string-name>
            <given-names>T. E.</given-names>
            <surname>Allen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Goldsmith</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H. E.</given-names>
            <surname>Justice</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Mattei</surname>
          </string-name>
          , and
          <string-name>
            <given-names>K.</given-names>
            <surname>Raines</surname>
          </string-name>
          .
          <article-title>Uniform random generation and dominance testing for cp-nets</article-title>
          .
          <source>JAIR</source>
          ,
          <volume>59</volume>
          :
          <fpage>771</fpage>
          -
          <lpage>813</lpage>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [Amor et al.,
          <year>2016</year>
          ]
          <string-name>
            <given-names>N. B.</given-names>
            <surname>Amor</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Dubois</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Gouider</surname>
          </string-name>
          , and
          <string-name>
            <given-names>H.</given-names>
            <surname>Prade</surname>
          </string-name>
          .
          <article-title>Graphical models for preference representation: An overview</article-title>
          .
          <source>In Proceedings of the 10th International Scalable Uncertainty Management (SUM</source>
          <year>2016</year>
          ), pages
          <fpage>96</fpage>
          -
          <lpage>111</lpage>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <source>[Bache and Lichman</source>
          , 2013]
          <string-name>
            <given-names>K.</given-names>
            <surname>Bache</surname>
          </string-name>
          and
          <string-name>
            <given-names>M.</given-names>
            <surname>Lichman</surname>
          </string-name>
          .
          <source>UCI Machine Learning Repository</source>
          ,
          <year>2013</year>
          . University of California, Irvine, School of Information and Computer Sciences.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [Bellet et al.,
          <year>2015</year>
          ]
          <string-name>
            <given-names>Aurélien</given-names>
            <surname>Bellet</surname>
          </string-name>
          , Amaury Habrard, and
          <string-name>
            <given-names>Marc</given-names>
            <surname>Sebban</surname>
          </string-name>
          .
          <source>Metric Learning. Synthesis Lectures on Artificial Intelligence and Machine Learning</source>
          . Morgan &amp; Claypool Publishers,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [Boutilier et al.,
          <year>2004</year>
          ]
          <string-name>
            <given-names>C.</given-names>
            <surname>Boutilier</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Brafman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Domshlak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.H.</given-names>
            <surname>Hoos</surname>
          </string-name>
          , and
          <string-name>
            <given-names>D.</given-names>
            <surname>Poole</surname>
          </string-name>
          .
          <article-title>CP-nets: A tool for representing and reasoning with conditional ceteris paribus preference statements</article-title>
          .
          <source>Journal of Artificial Intelligence Research</source>
          ,
          <volume>21</volume>
          :
          <fpage>135</fpage>
          -
          <lpage>191</lpage>
          ,
          <year>2004</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [Brandt et al.,
          <year>2016</year>
          ]
          <string-name>
            <given-names>F.</given-names>
            <surname>Brandt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Conitzer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>U.</given-names>
            <surname>Endriss</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Lang</surname>
          </string-name>
          , and A. D. Procaccia, editors.
          <source>Handbook of Computational Social Choice</source>
          . Cambridge University Press,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [Cornelio et al.,
          <year>2013</year>
          ]
          <string-name>
            <given-names>C.</given-names>
            <surname>Cornelio</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Goldsmith</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Mattei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Rossi</surname>
          </string-name>
          , and
          <string-name>
            <given-names>K.B.</given-names>
            <surname>Venable</surname>
          </string-name>
          .
          <article-title>Updates and uncertainty in CP-nets</article-title>
          .
          <source>In Proc. 26th AUSAI</source>
          ,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [Cornelio et al.,
          <year>2015</year>
          ]
          <string-name>
            <given-names>C.</given-names>
            <surname>Cornelio</surname>
          </string-name>
          ,
          <string-name>
            <given-names>U.</given-names>
            <surname>Grandi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Goldsmith</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Mattei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Rossi</surname>
          </string-name>
          , and
          <string-name>
            <given-names>K.B.</given-names>
            <surname>Venable</surname>
          </string-name>
          .
          <article-title>Reasoning with PCP-nets in a multi-agent context</article-title>
          .
          <source>In Proc. 14th AAMAS</source>
          ,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [Domshlak et al.,
          <year>2011</year>
          ]
          <string-name>
            <given-names>C.</given-names>
            <surname>Domshlak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Hüllermeier</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Kaci</surname>
          </string-name>
          , and
          <string-name>
            <given-names>H.</given-names>
            <surname>Prade</surname>
          </string-name>
          .
          <article-title>Preferences in AI: An overview</article-title>
          .
          <source>AI</source>
          ,
          <volume>175</volume>
          (
          <issue>7</issue>
          ):
          <fpage>1037</fpage>
          -
          <lpage>1052</lpage>
          ,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [Fagin et al.,
          <year>2006</year>
          ]
          <string-name>
            <given-names>Ronald</given-names>
            <surname>Fagin</surname>
          </string-name>
          , Ravi Kumar,
          <string-name>
            <given-names>Mohammad</given-names>
            <surname>Mahdian</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Sivakumar</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Erik</given-names>
            <surname>Vee</surname>
          </string-name>
          .
          <article-title>Comparing partial rankings</article-title>
          .
          <source>SIAM J. Discret. Math.</source>
          ,
          <volume>20</volume>
          (
          <issue>3</issue>
          ):
          <fpage>628</fpage>
          -
          <lpage>648</lpage>
          ,
          <year>March 2006</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [Fattah et al.,
          <year>2018</year>
          ]
          <string-name>
            <given-names>Sheik</given-names>
            <surname>Mohammad Mostakim Fattah</surname>
          </string-name>
          , Athman Bouguettaya, and
          <string-name>
            <given-names>Sajib</given-names>
            <surname>Mistry</surname>
          </string-name>
          .
          <article-title>A CP-Net based qualitative composition approach for an IaaS provider</article-title>
          .
          <source>In International Conference on Web Information Systems Engineering</source>
          , pages
          <fpage>151</fpage>
          -
          <lpage>166</lpage>
          . Springer,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          <source>[Fürnkranz and Hüllermeier</source>
          , 2010]
          <string-name>
            <given-names>J.</given-names>
            <surname>Fürnkranz</surname>
          </string-name>
          and
          <string-name>
            <given-names>E.</given-names>
            <surname>Hüllermeier</surname>
          </string-name>
          .
          <source>Preference Learning</source>
          . Springer,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          <source>[Goldsmith and Junker</source>
          , 2009]
          <string-name>
            <given-names>J.</given-names>
            <surname>Goldsmith</surname>
          </string-name>
          and
          <string-name>
            <given-names>U.</given-names>
            <surname>Junker</surname>
          </string-name>
          .
          <article-title>Preference handling for artificial intelligence</article-title>
          .
          <source>AI Magazine</source>
          ,
          <volume>29</volume>
          (
          <issue>4</issue>
          ),
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [Goldsmith et al.,
          <year>2008</year>
          ]
          <string-name>
            <given-names>J.</given-names>
            <surname>Goldsmith</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Lang</surname>
          </string-name>
          ,
          <string-name>
            <surname>M.</surname>
          </string-name>
          <article-title>Truszczy n´ski, and</article-title>
          <string-name>
            <surname>N. Wilson.</surname>
          </string-name>
          <article-title>The computational complexity of dominance and consistency in CP-nets</article-title>
          .
          <source>Journal of Artificial Intelligence Research</source>
          ,
          <volume>33</volume>
          (
          <issue>1</issue>
          ):
          <fpage>403</fpage>
          -
          <lpage>432</lpage>
          ,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          <source>[Goyal and Ferrara</source>
          , 2017]
          <string-name>
            <given-names>Palash</given-names>
            <surname>Goyal</surname>
          </string-name>
          and
          <string-name>
            <given-names>Emilio</given-names>
            <surname>Ferrara</surname>
          </string-name>
          .
          <article-title>Graph embedding techniques, applications, and performance: A survey</article-title>
          .
          <source>CoRR, abs/1705.02801</source>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          <source>[Kahneman</source>
          , 2011]
          <string-name>
            <given-names>Daniel</given-names>
            <surname>Kahneman</surname>
          </string-name>
          .
          <article-title>Thinking, fast and slow</article-title>
          . Farrar, Straus and Giroux, New York,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          <source>[Kambhampati</source>
          , 2019]
          <string-name>
            <given-names>S.</given-names>
            <surname>Kambhampati</surname>
          </string-name>
          .
          <article-title>Synthesizing explainable behavior for human-ai collaboration</article-title>
          .
          <source>In Proc. 18th AAMAS</source>
          ,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          <source>[Kendall</source>
          , 1938]
          <string-name>
            <given-names>M. G.</given-names>
            <surname>Kendall</surname>
          </string-name>
          .
          <article-title>A new measure of rank correlation</article-title>
          .
          <source>Biometrika</source>
          ,
          <volume>30</volume>
          (
          <issue>1</issue>
          /2):
          <fpage>81</fpage>
          -
          <lpage>93</lpage>
          ,
          <year>1938</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          <source>[Kingma and Ba</source>
          , 2014]
          <string-name>
            <given-names>D. P.</given-names>
            <surname>Kingma</surname>
          </string-name>
          and
          <string-name>
            <given-names>J.</given-names>
            <surname>Ba</surname>
          </string-name>
          .
          <article-title>Adam: A method for stochastic optimization</article-title>
          .
          <source>arXiv/1412.6980</source>
          ,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [Krizhevsky et al.,
          <year>2012</year>
          ]
          <string-name>
            <given-names>A.</given-names>
            <surname>Krizhevsky</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Sutskever</surname>
          </string-name>
          , and
          <string-name>
            <given-names>G. E.</given-names>
            <surname>Hinton</surname>
          </string-name>
          .
          <article-title>Imagenet classification with deep convolutional neural networks</article-title>
          .
          <source>In Proc. Advances in Neural and Information Processing Systems (NeurIPS)</source>
          , pages
          <fpage>1097</fpage>
          -
          <lpage>1105</lpage>
          ,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          <source>[Li and Kazimipour</source>
          , 2018]
          <string-name>
            <given-names>Minyi</given-names>
            <surname>Li</surname>
          </string-name>
          and
          <string-name>
            <given-names>Borhan</given-names>
            <surname>Kazimipour</surname>
          </string-name>
          .
          <article-title>An efficient algorithm to compute distance between lexicographic preference trees</article-title>
          .
          <source>In Proc. 27th IJCAI</source>
          , pages
          <fpage>1898</fpage>
          -
          <lpage>1904</lpage>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [Loreggia et al., 2018a]
          <string-name>
            <given-names>A.</given-names>
            <surname>Loreggia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Mattei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Rossi</surname>
          </string-name>
          , and
          <string-name>
            <given-names>K. B.</given-names>
            <surname>Venable</surname>
          </string-name>
          .
          <article-title>On the distance between CP-nets</article-title>
          .
          <source>In Proc. 17th AAMAS</source>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [Loreggia et al., 2018b]
          <string-name>
            <given-names>A.</given-names>
            <surname>Loreggia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Mattei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Rossi</surname>
          </string-name>
          , and
          <string-name>
            <given-names>K. B.</given-names>
            <surname>Venable</surname>
          </string-name>
          .
          <article-title>Preferences and ethical principles in decision making</article-title>
          .
          <source>In Proceedings of the 1st AAAI/ACM Conference on AI</source>
          ,
          <string-name>
            <surname>Ethics</surname>
          </string-name>
          , and
          <string-name>
            <surname>Society</surname>
          </string-name>
          (AIES),
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [Loreggia et al., 2018c]
          <string-name>
            <given-names>A.</given-names>
            <surname>Loreggia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Mattei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Rossi</surname>
          </string-name>
          , and
          <string-name>
            <given-names>K. B.</given-names>
            <surname>Venable</surname>
          </string-name>
          .
          <article-title>Value alignment via tractable preference distance</article-title>
          . In R. V. Yampolskiy, editor,
          <source>Artificial Intelligence Safety and Security</source>
          , chapter
          <volume>18</volume>
          . CRC Press,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [Loreggia et al.,
          <year>2019</year>
          ]
          <string-name>
            <given-names>A.</given-names>
            <surname>Loreggia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Mattei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Rossi</surname>
          </string-name>
          , and
          <string-name>
            <given-names>K. B.</given-names>
            <surname>Venable</surname>
          </string-name>
          . CPMETRIC:
          <article-title>Deep siamese networks for learning distances between structured preferences</article-title>
          .
          <source>arXiv preprint arXiv:1809.08350</source>
          ,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          <source>[Mattei and Walsh</source>
          , 2013]
          <string-name>
            <given-names>N.</given-names>
            <surname>Mattei</surname>
          </string-name>
          and
          <string-name>
            <given-names>T.</given-names>
            <surname>Walsh</surname>
          </string-name>
          . PREFLIB:
          <article-title>A library for preferences</article-title>
          , HTTP://WWW.PREFLIB.ORG.
          <source>In Proc. 3rd ADT</source>
          ,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          <source>[Mattei and Walsh</source>
          , 2017]
          <string-name>
            <given-names>N.</given-names>
            <surname>Mattei</surname>
          </string-name>
          and
          <string-name>
            <given-names>T.</given-names>
            <surname>Walsh</surname>
          </string-name>
          .
          <source>A PREFLIB</source>
          .
          <article-title>ORG Retrospective: Lessons Learned and New Directions</article-title>
          . In U. Endriss, editor,
          <source>Trends in Computational Social Choice</source>
          , chapter
          <volume>15</volume>
          , pages
          <fpage>289</fpage>
          -
          <lpage>309</lpage>
          .
          <source>AI Access Foundation</source>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          [Popova et al.,
          <year>2013</year>
          ]
          <string-name>
            <given-names>A.</given-names>
            <surname>Popova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Regenwetter</surname>
          </string-name>
          , and
          <string-name>
            <given-names>N.</given-names>
            <surname>Mattei</surname>
          </string-name>
          .
          <article-title>A behavioral perspective on social choice</article-title>
          .
          <source>AMAI</source>
          ,
          <volume>68</volume>
          (
          <issue>1-3</issue>
          ):
          <fpage>135</fpage>
          -
          <lpage>160</lpage>
          ,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          [Pu et al.,
          <year>2011</year>
          ]
          <string-name>
            <given-names>P.</given-names>
            <surname>Pu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Faltings</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , and
          <string-name>
            <given-names>P.</given-names>
            <surname>Viappiani</surname>
          </string-name>
          .
          <article-title>Usability guidelines for product recommenders based on example critiquing research</article-title>
          . In F. Ricci,
          <string-name>
            <given-names>L.</given-names>
            <surname>Rokach</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Shapira</surname>
          </string-name>
          , and P. B. Kantor, editors,
          <source>Recommender Systems Handbook</source>
          , pages
          <fpage>511</fpage>
          -
          <lpage>545</lpage>
          . Springer,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          <source>[Rossi and Loreggia</source>
          , 2019]
          <string-name>
            <given-names>F.</given-names>
            <surname>Rossi</surname>
          </string-name>
          and
          <string-name>
            <given-names>A.</given-names>
            <surname>Loreggia</surname>
          </string-name>
          .
          <article-title>Preferences and ethical priorities: Thinking fast and slow in AI</article-title>
          .
          <source>In Proc. 18th AAMAS</source>
          , pages
          <fpage>3</fpage>
          -
          <lpage>4</lpage>
          ,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          <source>[Rossi and Mattei</source>
          , 2019]
          <string-name>
            <given-names>Francesca</given-names>
            <surname>Rossi</surname>
          </string-name>
          and Nicholas Mattei.
          <article-title>Building ethically bounded AI</article-title>
          .
          <source>In Proc. 33rd AAAI</source>
          ,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          [Rossi et al.,
          <year>2011</year>
          ]
          <string-name>
            <given-names>F.</given-names>
            <surname>Rossi</surname>
          </string-name>
          ,
          <string-name>
            <surname>K.B. Venable</surname>
            , and
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Walsh</surname>
          </string-name>
          .
          <article-title>A Short Introduction to Preferences: Between Artificial Intelligence and Social Choice</article-title>
          . Morgan and Claypool,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref34">
        <mixed-citation>
          [Russell et al.,
          <year>2015</year>
          ] Stuart Russell, Daniel Dewey, and Max Tegmark.
          <article-title>Research priorities for robust and beneficial artificial intelligence</article-title>
          .
          <source>AI Magazine</source>
          ,
          <volume>36</volume>
          (
          <issue>4</issue>
          ):
          <fpage>105</fpage>
          -
          <lpage>114</lpage>
          ,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref35">
        <mixed-citation>
          <source>[Sen</source>
          ,
          <year>1974</year>
          ]
          <string-name>
            <given-names>A.</given-names>
            <surname>Sen</surname>
          </string-name>
          . Choice, Ordering, and
          <string-name>
            <surname>Morality</surname>
          </string-name>
          . Blackwell,
          <year>1974</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref36">
        <mixed-citation>
          <source>[Sohn</source>
          , 2016]
          <string-name>
            <given-names>K.</given-names>
            <surname>Sohn</surname>
          </string-name>
          .
          <article-title>Improved deep metric learning with multiclass n-pair loss objective</article-title>
          .
          <source>In Advances in Neural Information Processing Systems (NeruIPS)</source>
          , pages
          <fpage>1857</fpage>
          -
          <lpage>1865</lpage>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref37">
        <mixed-citation>
          <string-name>
            <surname>[Wang</surname>
          </string-name>
          et al.,
          <year>2009</year>
          ]
          <string-name>
            <given-names>Hongbing</given-names>
            <surname>Wang</surname>
          </string-name>
          , Shizhi Shao, Xuan Zhou, Cheng Wan, and
          <string-name>
            <given-names>Athman</given-names>
            <surname>Bouguettaya</surname>
          </string-name>
          .
          <article-title>Web service selection with incomplete or inconsistent user preferences</article-title>
          .
          <source>In Proc. 7th International Conference on Service-Oriented Computing</source>
          , pages
          <fpage>83</fpage>
          -
          <lpage>98</lpage>
          . Springer,
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref38">
        <mixed-citation>
          [Wong et al.,
          <year>2016</year>
          ]
          <string-name>
            <given-names>S. C.</given-names>
            <surname>Wong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Gatt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Stamatescu</surname>
          </string-name>
          , and
          <string-name>
            <surname>M. D McDonnell</surname>
          </string-name>
          .
          <article-title>Understanding data augmentation for classification: When to warp?</article-title>
          <source>In Proc. of the 2016 International Conference on Digital Image Computing: Techniques and Applications (DICTA)</source>
          , pages
          <fpage>1</fpage>
          -
          <lpage>6</lpage>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref39">
        <mixed-citation>
          [Xing et al.,
          <year>2002</year>
          ]
          <string-name>
            <given-names>E. P.</given-names>
            <surname>Xing</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. Y.</given-names>
            <surname>Ng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. I.</given-names>
            <surname>Jordan</surname>
          </string-name>
          , and
          <string-name>
            <given-names>S. J.</given-names>
            <surname>Russell</surname>
          </string-name>
          .
          <article-title>Distance metric learning with application to clustering with side-information</article-title>
          .
          <source>In Proc. 15th NeurIPS</source>
          , pages
          <fpage>505</fpage>
          -
          <lpage>512</lpage>
          ,
          <year>2002</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>