=Paper= {{Paper |id=Vol-2157/paper6 |storemode=property |title=Evidential Group Decision Making Model with Belief-Based Preferences |pdfUrl=https://ceur-ws.org/Vol-2157/paper6.pdf |volume=Vol-2157 |authors=Aouatef Rouahi,Kais Ben Salah,Khaled Ghedira |dblpUrl=https://dblp.org/rec/conf/cade/RouahiSG18 }} ==Evidential Group Decision Making Model with Belief-Based Preferences== https://ceur-ws.org/Vol-2157/paper6.pdf
    Evidential Group Decision Making Model with
               Belief-Based Preferences

    Aouatef Rouahi1[0000−0001−8293−9991] , Kais Ben Salah2[0000−0003−4309−9226] ,
                   and Khaled Ghedira3[0000−0001−5719−1598]
        1
            ISG of Tunis, Tunis University, Tunisia rouahi.aouatef@hotmail.fr
                 2
                    FCIT, University of Jeddah, SA kaisuria@yahoo.fr
               3
                  Central University, Tunisia khaled.ghedira@isg.rnu.tn



        Abstract. In a large number of problems such as product configuration,
        automated recommendation and combinatorial auctions, decisions stem
        from agents subjective preferences and requirements in the presence of
        uncertainty. In this paper, we introduce a framework based on the belief
        function theory to deal with problems in group decision settings where
        the preferences of the agents may be uncertain, imprecise, incomplete,
        conflicting and possibly distorted. A case study is conducted on prod-
        uct recommendation to illustrate the applicability and validity of the
        proposed framework.

        Keywords: Preference · Constraints · Uncertainty · Group decision ·
        Belief Function Theory · Soft constraints.


1     Introduction
In many real world applications such as product configuration, product design
and automated recommendations, decision-making stems from the interplay of
agents’ preferences and requirements given a set of alternatives. The main task
of such applications is to find the most preferred feasible outcomes. While imper-
fection pervades our real life, when expressing preferences, agents are assumed
to (i) act with full external information so they can certainly and precisely figure
out which alternatives are better than which; (ii) act with full internal informa-
tion so they are decisive about their preferences whatever the proposed set of
alternatives in such a way they are able to rank the alternatives from the best
to the worst, so that their beliefs are often ignored or confounded with their
final preferences. Far from being primitive, agents’ preferences should depend
on their beliefs about the properties and the outcomes of the alternatives espe-
cially in those situations in which an agent may only have partial and somehow
uncertain information about alternatives he is required to provide his prefer-
ences over, i.e., ill-defined alternatives. From another side, the increasing use of
multi-agent applications demands for the combination and handling of possibly
conflicting or distorted preferences supplied by multiple agents. Once preferences
are fixed, decisions can be inferred given the constraints that determine which
alternatives are feasible. In this context, we propose an evidential approach for
2       A. Rouahi et al.

group decision where agents requirements are modeled as hard constraints that
should be imperatively met and their possibly imperfect (uncertain, imprecise,
incomplete, indecisive, and conflicting) preferences are modeled as belief-based
soft constraints using the belief function theory. Soft constraints [1], equipped
with a powerful solving machinery, provide an interesting way to model and rea-
son with quantitative preference relations and constraints. However, the notion
of preference had still not been properly addressed within soft constraints frame-
work in the sense that preferences are basically used to relax over-constrained
problems, to discriminate between different solutions or to reduce search effort.
Thus, a preference structure is not clearly defined where some specific induced
relations are implicitly stated such as the indifference relation or ignored such as
the incomparability relation. The belief function theory [3, 10, 11] offers a sound
mathematical basis that faithfully recognizes all situations ranging from com-
plete knowledge to complete ignorance. Moreover, by considering its Transferable
Belief Model (TBM) interpretation [11], we introduce a two-level preference per-
spective: (i) the evidence base in which imperfect preferences are quantified,
combined and revised; (ii) the final preferences derived from the evidence base.
In addition, the TBM allows to combine the possibly conflicting preferences
supplied by multiple agents. Our purpose in this paper is to bring into sharper
focus the interesting interplay between beliefs, preferences and constraints when
reasoning under uncertainty. We propose then a specific Branch and Bound algo-
rithm for solving such kind of problems. The remainder of this paper is organized
as follows: Section 2 reviews related work. Some preliminaries are discussed in
Section 3. We present the evidential approach and its basic components in Sec-
tion 4. In Section 5, reasoning with preferences and constraints to construct
solutions are discussed and a specific branch and bound algorithm is introduced.
Conclusions and further researches are drawn in Section 6.


2   Related Work

While intertwined preferences and constraints is thoroughly studied in the AI
literature, beliefs and preferences are rarely considered despite the multitude
of approaches to preferences except some studies of the logic of preference [6,
7] investigating preference dynamics under belief change. To the best of our
knowledge, in the constraint satisfaction field, we are the first proposing such
connection between the belief function theory and soft constraints framework.
Nevertheless, a variety of proposals has been introduced to extend soft con-
straints framework to deal with imperfect preferences. In our model, imperfect
(i.e., uncertain, imprecise, incomplete, contradictory and so on) information af-
fects the agent’s evidence base and then propagated into his preferences. The
work in [4] that considers incomplete soft constraint problems where some of
the preferences may be allowed to be missing as long as it is feasible to find an
optimal solution, otherwise, the agent will be required to add some preferences.
In this work, incompleteness is interpreted as temporary inability or unwilling-
ness to provide preferences over some alternatives. In our approach, we consider
      Evidential Group Decision Making Model with Belief-Based Preferences           3

incompleteness as a decisive undesirability to compare some alternatives with
regard to the available evidence, thus, we do not require the agent to supply fur-
ther information. Another proposal considers preference intervals [5] to model
imprecision in preference intensity. In our work, we assume that the preference
intensities are, precisely stated. However, we permit ties in the preference list to
model users’ indifference about the tied alternatives. Other work that addresses
uncertainty in soft constraints using the possibility theory is shown in [9] where
some alternatives may be ill-defined, i.e., one cannot decide their values. In our
work, we, thoroughly, address uncertainty at two levels. First, we consider the
case where the alternatives may be ill-defined so the agent cannot express his
preferences in the form of ”yes/no” but he could reply ”I somewhat prefer this
alternative”, i.e., preference intensity, or he may simply say ”I do not know” to
express his ignorance. Second, we consider the case where the preference relation
itself may be ill-defined, i.e., an agent’s true preferences may be distorted because
the agent is not decisive (hesitant) about his preferences like in matching prob-
lems or he is not willing to provide his true preferences due to privacy reasons in
the case of multi-agent settings with competing agents. Thus, in our proposal,
the truth intensity of the preference is evaluated to measure to what extent the
provided preferences by some agent meet his true preferences. This issue is too
important, especially, for critical domains such as medical or business applica-
tions. Not taking into account such issues may result in costly biased decisions
and unsatisfied users as happened with Walmart UX in 2009. Further, by means
of our two-level preference perspective, we have been able to, faithfully, capture
the different preferential positions of an agent, i.e., strict preference, indifference
and incomparability. Furthermore, our evidential approach permits to cope, sys-
tematically and consistently, with all of these decision-making ingredients in a
unifying frame.


3     Preliminaries
3.1   Soft Constraints
The problem of finding an optimal satisfying assignment of values to a finite set
of variables, each with a finite domain, given a finite set of constraints and with
respect to a finite set of agent’s preferences is often referred to as constrained
optimization problem. In this paper, we are interested in the two generic soft
constraints frameworks, namely Semiring-based CSP (SCSP) and Valued CSP
(VCSP) [1] that cover many specific others.
    In general, in a SCSP, a soft constraint or a preference relation is defined by
associating a degree from a partially ordered set with each involved alternative
indicating to which extent an alternative is preferred. Moreover two operations
with certain properties for comparing (+) and for combining (×) preference
degrees in order to select the best solution. Formally, a c-semiring is a tuple
< A, +, ×, 0, 1 > such that: A is a set, and 0, 1 ∈ A; + is commutative, asso-
ciative, idempotent, 0 is its unit element, and 1 is its absorbing element; × is
commutative, associative, distributes over +, 1 is its unit element and 0 is its
4       A. Rouahi et al.

absorbing element. Consider the relation ≤S over A such that a ≤S b iff a+b = b.
Then: ≤S is a partial order; + and × are monotone on ≤S ; 0 is its minimum
and 1 its maximum; < A, ≤S > is a lattice and, for all a, b ∈ A, a + b = lub(a, b).
Moreover, if × is idempotent, then < A, ≤S > is a distributive lattice and ×
is its glb. Given a c-semiring S =< A, +, ×, 0, 1 >, a finite set D, and an or-
dered set of variables V , a constraint is a pair < def, con > where con ⊆ V and
def : D|con| −→ A.
    In a VCSP, each soft constraint, i.e., the whole set of alternatives, is asso-
ciated with a degree from a totally ordered set indicating to which extent the
satisfaction of a given constraint is preferred. Besides one operation with certain
properties for combining (~) different constraints. Formally, a valuation struc-
ture is defined by (E, ~, , >, ⊥) where E is a set of valuations;  is a total
ordering over E; >and ⊥ are maximum and minimum elements of E given by
; ~ is a commutative, associative binary operation on E, ⊥ is its unit element
and > is its absorbing element, ~ is monotone on .
    In our approach, we have adopted both approaches at different levels. We opt
for the SCSP approach to model the preference intensities of alternatives and
for the VCSP approach to model the truth intensities of the preferences using
the belief function theory.


3.2   Belief Function Theory

The belief function theory was first initiated by [3] and then extended by [10].
Several interpretations have been introduced such as the well known TBM es-
tablished by [11].


Basic Concepts Let Θ be a frame of discernment representing a finite set of
elementary alternatives. A basic belief assignment (bba) m is the mapping from
elements of the power set 2Θ to [0, 1] such that:
                       X
                             m(θ) = 1     and      m(∅) = 0.                     (1)
                      θ∈2Θ

The basic belief mass (bbm) m(θ), assigned to some subset θ of Θ, is a positive
finite amount of support that is derived from the available pieces of evidence and
exactly given to the set θ and not to any specific subset of θ by lack of evidence.


Discounting The discounting procedure allows taking into account the reliabil-
ity of the source providing the bba m. It consists in weighting the beliefs yielded
by each source using a discounting coefficient, so that, the smaller the reliability,
the stronger the discounting. Let m be a bbm given by the source S on Θ and
let α ∈ [0, 1] be the confidence degree allocated to the source S. If the source
is not fully reliable, the provided bba is discounted into a new weaker and less
informative one denoted mα , where every lost mass is reassigned to Θ as total
     Evidential Group Decision Making Model with Belief-Based Preferences       5

ignorance:              (
                         mα (θ) = α.m(θ), ∀θ ⊂ Θ
                                                                              (2)
                         mα (Θ) = (1 − α) + α.m(Θ).


Evidence combination The evidence of two independent bbas m1 and m2
induced from two distinct sources and defined on the same frame of discernment
Θ could be combined to form a single bba m12 on Θ via the TBM Conjunctive
Rule of Combination (CRC):
                                      X
          m12 (θ) = m1 ∩ m2 (θ) =              m1 (B)m2 (C); ∀θ ⊆ Θ.       (3)
                                  B,C⊆Θ,B∩C=θ



Decision Making Given a set of alternatives Θ and a given bba m, we want
to establish an ordering over Θ based on m. Many decision-making criteria have
been developed in the literature. In our approach, we opt for decision-making
based on maximum of pignistic probability (BetP ) that offers a compromise
between pessimistic and optimistic strategies, where higher probability degree
indicates more preferred alternative. Hence, The bba m is reformed to a subjec-
tive probability measure BetP as follows:

                                1    X |A ∩ θ|.m(θ)
               BetP (A) =                           ; ∀A ∈ Θ.                 (4)
                            1 − m(∅)        |θ|
                                       θ⊆Θ



4   Evidential Constrained Optimization

An evidential constrained optimization problem ℘ is defined by the tuple (X,D,B-
Pref,Cons), involving a finite set of variables X, their associated finite domains
D and a finite set of belief-based preferences B-Pref. A belief-based preference
b-pref is a belief soft constraint defined by the tuple (S,A,B,R), where, S ⊆ X
is the scope of the preference delimiting the set of

Example 1. (revised from [4]) A travel agency is planning Alice and Bob’s hon-
eymoon. The candidate destinations are the Maldive islands and the Caribbean,
and they can decide to go by ship or by plane. For the accommodations, the
couple can choose between a standard room, a suite, or a bungalow. We have
X = {T r, Des, Acc} where the subscripts Tr, Des, Acc stand respectively for
Transport, Destination and Accommodation, with D(T r) = {p, sh} (p stands
for plane and sh for ship), D(Des) = {m, c} (m stands for Maldives, c for
Caribbean), and D(Acc) = {r, su, b} (r stands for room, su for suite and b for
bungalow). Alice and Bob have independent and distinct opinions about this de-
cision situations given the evidence held by each of them so they have different
preferences. However, they have a common budget constraint as they cannot af-
ford more than $5000 for this trip. Table 1 summarizes the information supplied
by the agency about the different costs.
6         A. Rouahi et al.

                            Table 1. Choices costs for Example 1.

                                             Destination
                                          Maldives Caribbean
                                   Plane $2112,00 $1030,00
                      Transport
                                   Ship $1346,00 $760,00
                                   room $2280,00 $3450,00
                    Accommodation suite $3825,00 $4120,00
                                 bungalow $4285,00 $4685,00

                       Table 2. The evidence bases for Example 1.

                  b-pref1 b-pref2                            b-pref3
        S         {T r}   {Des, T r}                         {Des, Acc}
        A         {p, sh} {(m, p), (m, sh), (c, p), (c, sh)} {(m, r), (m, su), (m, b),
                                                             (c, r), (c, su), (c, b)}
        B(Alice) p:0.8    (c,sh):0.6                         (c,su),(c,b),(m,su):0.7
                 sh:0.2   (m,p),(m,sh):0.4                   (c,r),(m,b):0.3
        B(Bob) (p,sh):1.0 (m,p),(c,sh):0.6                   (m,r),(c,su):0.5
                          (m,p),(m,sh),(c,p),(c,sh):0.4 (c,r),(c,b),(m,su):0.3
                                                             (m,b):0.2



4.1     Evidence Base

Beliefs Modeling Given the scope of the preference S and the related set of
alternatives A, the agent’s beliefs B over A are modeled in terms of a partial
order D induced by the bba m on 2A to [0,1]:

                              D = {(θ1 , θ2 )|m(θ1 ) ≥ m(θ2 )}.                          (5)

The instance θ1 Dθ2 stands for ”the betterness of θ1 is at least as supported as the
betterness of θ2 ”, giving the evidence held by the agent. D is reflexive, transitive
and antisymmetric as its associated strict component B(”is strictly supported
to”) is irreflexive, transitive and asymmetric, its indifference component ≡(”is
as supported as”) is reflexive, symmetric and composed of (θ, θ) pairs only, and
its associated incomparability relation ./ (”is incomparable to”) is irreflexive,
not transitive and symmetric.
    In Example 1, The evidence bases of Alice and Bob are shown in Table 2.
    By means of our two-level preference perspective, we have been able to cap-
ture all the states of the agent towards his preferences. For us, beliefs as prior
preferences count as reasons for final preferences forming and then a base for
their justification.

    – Full certainty: take Example 1, for the preference b-pref1 in Table 2, if Alice
      came to learn from a best friend that the service on the ship they are planning
      to board on is of poor quality, she will certainly prefer traveling by plane.
      This is will be translated by a certain and precise belief:(p:1.0 ; sh:0.0).
     Evidential Group Decision Making Model with Belief-Based Preferences          7

 – Partial ignorance: for the preference b-pref1 , Alice has an uncertain but
   precise belief about the betterness of the alternatives. For the preference b-
   pref2 , Bob has an uncertain and imprecise belief as he tied some alternatives
   together.
 – Total ignorance: for the preference b-pref1 , Bob is totally ignorant about
   both transport alternatives.
 – Null support: the agent has no evidence to believe that an alternative can
   be somehow good or bad such as for the preference b-pref2 , Alice evidence
   does not allow her to decide her preference for the alternative (c,p).


Preference Truth Intensity In some cases, the provided preferences by some
source may not reflect true ones for various reasons. An agent may be unable or
unwilling to definitely decide their preferences because of uncertainty or privacy
issues. Hence, the reliability of the source providing preferences and then the
truth intensities of the provided preferences should be evaluated. In spite of its
importance for decision-making, the intertwining of preferences and reliability is
rather unexplored in AI literature. When we can quantify the extent to which
provided preferences reflect true ones, we can weaken the preference relation by
discounting its corresponding evidence base using the discounting rule described
in section 3.2. Given α ∈ [0, 1], the truth intensity of some b-pref, indicating the
reliability of its source, two extreme scenarios can be met:

 – If α = 1, the provided preferences match the true ones therefore discounting
   does not affect the preference relation so that b-pref α =b-pref.
 – If α = 0, the provided preferences are totally distorted therefore discounting
   induces to the total ignorance case where all the provided information is
   discarded.

    Take Example 1, the travel agency asks both of Alice and Bob to quan-
tify their decisiveness about their provided preferences. Alice was well informed
about the alternatives in question so she was more decisive than Bob who was
a little bit hesitant (see Table 3). Finally, it is necessary to take this meta-


              Table 3. The discounted evidence bases for Example 1.

               b-pref1 b-pref2                         b-pref3
      B(Alice) p:0.8    (c,sh):0.6                     (c,su),(c,b),(m,su):0.7
      α = 1.0 sh:0.2    (m,p),(m,sh):0.4               (c,r),(m,b):0.3
      B(Bob) (p,sh):1.0 (m,p),(c,sh):0.48              (m,r),(c,su):0.4
      α = 0.8           (m,p),(m,sh),(c,p),(c,sh):0.52 (c,r),(c,b),(m,su):0.24
                                                       (m,b):0.16
                                                       (m,r),(m,su),(m,b),(c,r),
                                                       (c,su),(c,b):0.2



knowledge about preferences into account especially for those decision-making
8         A. Rouahi et al.

problems relying on agents’ preferences such as product and service bundling,
multi-item auctions, policy making and so on. Instead of distrusting, or relying
on preferences no matter how distorted they are, one may want to assess their
truth intensities in order to decide whether to rely on those preferences. In some
critical applications, decisions should depend only on undistorted preferences.
In other applications, we may tolerate distortion to some degree.

Combination of Agents’ Preferences After revising the provided prior pref-
erences given the truth intensities, we can combine them using the CRC de-
scribed in Section 3.2. The combined preferences of Alice and Bob are shown in
Table 4.

                Table 4. The combined prior preferences for Example 1.

                          b-pref1 b-pref2           b-pref3
            B(Alice, BoB) p:0.8 (c,sh):0.6          (c,su):0.28
                          sh:0.2 (m,p),(m,sh):0.208 (c,b),(m,su):0.168
                                  (m,p):0.192       (c,su),(c,b),(m,su):0.14
                                                    (c,r):0.072
                                                    (c,r),(m,b):0.06
                                                    (m,b):0.048
                                                    ∅4 :0.232




4.2     Final Preferences
Final Preferences Deriving As the agents’ prior preferences are combined,
the final preference relations are derived as a partial preorder B induced by
the BetP measures over A, the set of alternatives:

    B = {(a1 , a2 )|(BetP (a1 ) ≥ BetP (a2 ) ∧ (min(BetP (a1 ), BetP (a2 )) > 0)} (6)

The relation B is reflexive and transitive. Given the ordering B and two
alternatives a1 , a2 ∈ A, we distinguish between three relations over a1 and a2 :
    – a1 is strictly preferred to a2 , denoted by a1 B a2 , when a1 B a2 holds
      but a2 B a1 does not. The agent’s evidence provides more support 5 for a1
      over a2 . B is irreflexive, transitive and asymmetric.
    – a1 is indifferent to a2 denoted by a1 ≈B a2 , when both a1 B a2 and
      a2 B a1 hold. The agent’s evidence does not support a1 more strongly
      than a2 and does not support a2 more strongly than a1 . ≈B is reflexive,
      transitive and symmetric.
4
  The preferences will be re-normalized when deriving the final preferences using the
  pignistic probabilities (see Section 3.2).
5
  The term support denotes a non-null degree of belief; otherwise, we cannot refer to
  a zero degree of belief as a support.
       Evidential Group Decision Making Model with Belief-Based Preferences          9

 – a1 is incomparable to a2 , denoted by a1 ∼B a2 , when neither a1 B a2 nor
   a2 B a1 holds. The agent has no evidence (about a1 or a2 or both) for
   comparing a1 and a2 . ∼B is irreflexive, not transitive and symmetric.
   The derived final preference relations from evidence bases in Example 1 are
shown in Table 5.

                 Table 5. The derived final preference for Example 1.

                                b-pref1 b-pref2         b-pref3
                  R(Alice, BoB) p:0.8 (c,sh):0.6        (c,su):0.43
                                sh:0.2 (m,p):0.296      (c,b):0.17
                                        (m,sh):0.104    (m,su):0.17
                                        (c,p):0.0       (c,r):0.13
                                                        (m,b):0.1
                                                        (m,r):0.0
                  B instances    pB sh (c,sh)∼B (c,p) (c,b)≈B (m,su)




Conflicting Preferences In practice, there are often several preference re-
lations that have to be considered, each emphasizing a different facet of the
problem being addressed. In our approach, a variable may be involved in more
than one preference relation which may give rise to some conflict. A preference
set B-Pref is consistent (conflict free) if and only if it has no preference b-pref i
and b-pref j such that a1 B                B
                            i a2 and a2 j a1 for any alternatives a1 and a2 . As
the preference relation is no longer a yes-or-no issue, the notion of consistency
also becomes a matter of degree. In our approach, we assume that the more
two preference relations are far from each other, the more they are in conflict.
Thus, we propose to define the conflict between two belief-based preference re-
lations Ri and Rj using a normalized L1 metric between their respective BetP
distributions as follows:
                        P
                         (a∈D(S)) |BetPiSi ↓S (a)−BetPjSi ↓S (a)|
                                          |D(S)|                   if S 6= ∅,
          Conf (i, j) =                                                          (7)
                        0                                         if S = ∅.

Where: S = Si ∩ Sj ; D(S)6 ={×k Dk |xk ∈ S} ; |D(S)|: Cardinality of D(S) and
BetPiSi ↓S (a) = maxai ∈Ai :a↓S =a BetP (ai )
                              i


 – If Conf (i, j) = 0 the preference relations Ri and Rj are totally concordant
 – If 0 < Conf (i, j) < 1 the preference relations Ri and Rj are partially con-
   flicting
 – If Conf (i, j) = 1 the preference relations Ri and Rj are totally conflicting
6
    We use D(.) to denote the domain of a variable and the domain of a set of variables
    as well.
10      A. Rouahi et al.

Returning Example 1 : To compute the conflict degree between b-pref1 and b-
pref2 , we have S = S1 ∩ S2 = {T r}, D(S) = D(T r), |D(S)| = 2, BetP1S1 ↓S (p) =
0.8 and BetP2S2 ↓S (p) = max(0.296, 0.0) = 0.296, BetP1S1 ↓S (sh) = 0.2 and
BetP2S2 ↓S (sh) = max(0.6, 0.104) = 0.6, hence Conf(1,2)= (|0.8−0.296|+|0.2−0.6|)
                                                                     2            =
0.452. The same process is used to get Conf(1,3)=0 and Conf(2,3)=0.148.

                 Consistency(B − P ref ) = 1 − max(conf (i, j))                     (8)

In Example 1, Consistency(B-Pref )= 1 − 0.452 = 0.548, so B-Pref is partially
consistent. In our approach, as soon as two preference relations are totally con-
flicting, the preference set is totally inconsistent. Assessing conflict between two
preference relations may serve as an early detection of the problem inconsistency,
which brings cost and time savings. Furthermore, we can quantify to what extent
a given preference relation b-prefi , in a preference set B-Pref with n preferences,
is in conflict with the other (n-1) preferences by:
                                                    n
                                            1       X
                  Conf (i, B − P ref ) =                     Conf (i, j)            (9)
                                           n−1
                                                  j=1,i6=j

Returning Example 1 : Conf(1,B-Pref )=0.226; Conf(2,B-Pref )=0.3; Conf(3,B-
Pref )=0.074.

4.3   Constraints Modeling
Constraints represent limitations that winnow the set of alternatives we can opt
for in a given situation.
    In our approach, we separately deal with constraints and preferences dif-
ferently from the common soft constraints formalism where hard and soft con-
straints are coupled together and handled using the same representation. We
argue that for many real world applications, such as product configuration and
design, automated customized recommendations, a decoupled setting is the most
appealing and allows saving computation time for early discovered unfeasible
outcomes. This issue is carefully discussed in [2].                  P
    In Example 1, Alice and Bob have one budget constraint C1( i=1..2 pi ≤
$5000), where p1 and p2 are respectively the transport and the accommodation
costs. Once final preferences and constraints are given, decisions are determina-
tive.


5     Reasoning with Preferences and Constraints
Let S be a set of variables, we will use the notation ωS to denote an outcome
resulting from assigning a value to each variable in S from its equivalent domain.
We will say that an outcome is complete iff it is defined on X, otherwise it is said
to be partial. Consider a b-prefi defined on the set of variables Si , δ(i, ωSi ) =
BetPi (ωSi ) will denote the satisfaction degree of b-prefi by some outcome ωSi ∈
Ai . b-prefi is said to be satisfied by ωSi , noted ωSi |= b-prefi , iff δ(i, ωSi ) > 0.
      Evidential Group Decision Making Model with Belief-Based Preferences       11

    Solving an evidential constrained optimization problem consists in finding a
                    ∗
complete outcome ωX   , if it exists, that satisfies all the constraints in Cons and
is optimal with respect to the preferences in B-Pref.


5.1   Operations on Preferences

Projection and Combination consider two sets of variables S = {x1 , .., xl }
and Si = {xi1 , .., xim } such that Si ⊆ S ⊆ X. Let ω = (v1 , .., vl ) be any
outcome over S, the projection of ωS from S to Si denoted by ω ↓SSi is de-
fined as the outcome ωSi = (vi1 , .., vim ) with vik = vj if xik = xj . For ex-
                                             {Des,Acc}
ample, if ω{Des,Acc} = (m, su), then ω ↓{Des}          = (m). Consequently, given
a b-pref i defined on Si and some outcome ωS such that Si ⊆ S ⊆ X, then
δ(i, ωS ) = δ(i, ω ↓SSi ) will be the local satisfaction degree of b-pref i by ωS .
Hence, The global degree of joint satisfaction of the set of n belief-based prefer-
ences B-Pref defined on the set of variables X by a given complete outcome ωX
is obtained by combining the local satisfaction degrees as follows:

            δ(B − P ref, ωX ) = ⊗i δ(i, ωX ) = ⊗i δ(i, ω ↓X
                                                          Si ), ∀i = 1..n      (10)

Different combination operators ⊗ can be used that reflect various attitudes
towards preferences satisfaction:

 – Min-Combination: using the egalitarian min operator is a pessimistic ap-
   proach leading to the so-called ”drowning effect”, i.e., the worst local degree
   of satisfaction drowns all the others regardless how much the rest of the
   preferences are satisfied. For instance, consider two complete outcomes ωX
          0
   and ωX    with only two belief-based preferences, and such that ωX satisfies
                                                    0
   these b-pref (s) with degrees 0.5 and 1.0 while ωX satisfies them with degrees
                                                                    0
   0.5 and 0.5. Although ωX is obviously strictly preferable to ωX    , the global
   satisfaction degree of the two outcomes is identical since min (0.5,1.0) =
   min (0.5,0.5).
 – Max-Combination: it represents an optimistic approach, but this egalitarian
   operator suffers from the same weakness as the Min-Combination and barely
   discriminates between outcomes with the same global satisfaction degree.
 – Product-Combination: using an utilitarian operator avoids falling in the
   ”drowning effect” weakness. However, it does not discriminate between out-
   comes fully falsifying at least one preference.
 – Average-Combination: it represents an utilitarian flexible approach that of-
   fers more discriminating ordering than the former combinations tolerating
   some preferences to be falsified.

The suitability of each of these combination approaches depends on the applica-
tion nature, where in some critical applications we need a pessimistic and strict
approach that does not tolerate the falsification of any preference. However, in
other domains, an optimistic and flexible approaches are more useful.
12       A. Rouahi et al.

Extension and Estimation consider two sets of variables S = {x1 , .., xl } and
Si = {xi1 , .., xim } such that S ⊆ Si ⊆ X. Let ωS = (v1 , .., vl ) be any outcome
over S, the extension of ωS from S to Si denoted by ω ↑SSi is defined as the set
of outcomes ΩSi = {(vi1 , .., vim ) × DS−Si }. For example,if ω{Des} = (m), then
     {Des,T r}
ω ↑{Des} = Ω{Des,T r} = {(m, p), (m, sh)}. Thus, given a b-pref i defined on Si
and some outcome ωS such that S ⊆ Si ⊆ X, then the estimated satisfaction
degree of b-pref i by ωS will be δ e (i, ωS ) = M ax{δ(i, ωSi )|ωSi ∈ ΩSi }.

5.2     Constructing solutions
Given an evidential constrained optimization problem ℘ (X,D,B-Pref,Cons), ev-
ery feasible complete outcome with respect to Cons that jointly satisfies B-Pref
to a global satisfaction degree greater than 0 (whatever the used approach), is
considered as a solution.

                 ωX ∈ S(P ) ⇔ ωX |= Cons ∧ δ(B − P ref, ωX ) > 0             (11)

The global satisfaction degree induces a total preorder over the set of feasible
outcomes, so that the best outcome will be the one that, maximally, satisfies
B-Pref:
                         ∗
                        ωX = argmax δ(B − P ref, ωX )                      (12)
                               ωX ∈S(P )



5.3     PDBB Algorithm
Commonly, when solving such constrained optimization problems, Variable-Directed
Branch and Bound (VDBB) algorithm is the most widely used using two bounds:
an upper bound B and a lower bound b. It incrementally builds, by assigning a
variable with a value selected from its domain, outcomes prospected to be solu-
tions, where it early on aborts every partial outcome that cannot be extended
to construct a better solution than the one found so far using the bounds B and
b.
    At each level, instead of assigning one variable with a value from its domain,
we propose to assign multiple variables with values from the preference relation
covering those variables, hence, the Preference-Directed BB (PDBB). This idea
has been first introduced in [8] for the Constraint-Directed Backtracking and
proved to be less costly in terms of search effort.
    Initially, the PDBB selects a minimal set of m preferences, from the set of n
preferences, that covers all the model variables X, denoted B-Pref c ⊆ B − P ref .
In Example 1, the selected preferences will be b-pref2 and b-pref3 . An upper
bound B that contains the global satisfaction degree of the best solution found
so far and is initialized to the tolerated worst global satisfaction degree in order
to discard each solution giving a satisfaction degree ≤B. After that, at each
level, a preference relation Ri related to a b-pref i ∈ B-Pref c is explored by
trying each alternative from its related alternatives set Ai . The current partial
solution is then extended to that alternative so that variables in Si implied by
     Evidential Group Decision Making Model with Belief-Based Preferences       13

the preference, not previously covered, get assigned. If the constructed partial
outcome satisfies all the constraints involving the assigned variables, an exact
satisfaction degree of all preferences covering those variables is computed, joined
with an estimation of the satisfaction degree of the rest of preferences, resulted
in a lower bound b which is initialized to 1. Let the current partial outcome ωS
defined on the set of variables S and let B-Prefa the set of preferences activated
by the assignment and B-Prefa be the rest of B-Pref, the lower bound is computed
as follows:
                   b = δ(B − prefa , ωS ) ⊗ δ e (B − prefa , ωS )             (13)

    Backtracking occurs and the sub-tree below the current node is pruned if
(1) the partial outcome violates at least one constraint; (2) it satisfies all the
constraints but it cannot lead to a better solution b ≤ B. If all the problem’s
variables are assigned, a new solution is found with a satisfaction degree strictly
higher than B, so the solution is printed and the upper bound B is updated.
The algorithm terminates when no better solution can be found.
    In addition, by applying the heuristic ”alternative giving best satisfaction
degree (max-BetP) is selected first”, we ensure that best solution is early con-
structed and thus the search space is reduced (see Algorithm 1).
    We illustrate, in Fig. 1, the PDBB execution to solve the problem described
in Example 1. In this execution we adopted the product combination seen in
Section 5.1.




                Fig. 1. PDBB trace for the problem in Example 1.
14       A. Rouahi et al.


    Algorithm 1: PDBB Algorithm
     input : (X, RSc , Cons, S, ωS , B, b)
     /* Rc = {Ri | m (i=1) Si = X} is minimal; S = ∅;ωS = ∅; B = 0; b = 1          */
               ∗
     output: (ωX , B)
     while Rc is not empty do
        select and remove a relation Ri ∈ Rc ;
        S ← S ∪ Si ;
        while Ai is not empty do
            select and remove best a ∈ Ai ;
            ωS ← ωS ∪ a;
            if ωS |= Cons then
                Compute a lower bound b for ωS ;
                /* b = δ(B − prefa , ωS ) ⊗ δ e (B − prefa , ωS ) such that B-Prefa
                    is the set of preferences activated by the current
                    assignment and B-Prefa is the rest of B-Pref that are
                    not yet implied.                                                */
                if b > B then
                    if S = X then
                        B ← b;
                          ∗
                        ωX  ← ωS ;
                                 ∗
                        Print (ωX  , B);
                        if B = 1 then
                            return ”Finished”;

                    else
                        PDBB(X, Rc , Cons, S, ωS , B, b);

                else
                    PDBB(X, Rc , Cons, S, ωS − a, B, b);

            else
                PDBB(X, Rc , Cons, S, ωS − a, B, b);




    In Fig. 1, the outcomes having a red (X) are discarded because they violates
the constraints, however, the outcomes with green (X) are aborted because they
cannot lead to a better solution.
    For Example 1, the best affordable trip package for Alice and Bob is { Des-
tination: Caribbean; Transport: Ship; Accommodation: suite}.

6     Conclusion and Further Work
In this paper, we have introduced an evidential approach for constrained opti-
mization problems whereby agents, often dealing with only partial and somehow
uncertain external and internal information, seek for decisions that satisfy their
preferences based on their beliefs subject to certain constraints by extending the
soft constraints framework to the belief function theory.
     Evidential Group Decision Making Model with Belief-Based Preferences            15

    For solving such kind of problems, we have provided a specific branch and
bound algorithm which is initially proven to be less costly than the classical
branch and bound algorithm. However, a detailed study of its computational
properties and potential refinements should be conducted by introducing heuris-
tics for the order of checking the preferences. More sophisticated methods from
the constraint satisfaction machinery could be easily extended to our approach.
    Further research targets exploiting the expressiveness offered by the eviden-
tial approach in order to enlarge the scope of the issues that can be tackled
such as prioritized preferences. Preference dynamics can also be studied using
the belief revision process. We can address the bipolar preferences exploiting
the negative and positive belief notions. We also intend to introduce the weak
preference relation using thresholds. Finally, we plan to explore how our model
can be employed in decision support applications like recommender systems,
configuration problems and combinatorial auctions.


References
 1. S. Bistarelli, U. Montanari, and F. Rossi (1995). Constraint Solving over Semirings.
    In Proc. IJCAI95. Morgan Kaufman.
 2. C. Boutilier, R. I. Brafman, C. Domshlak, H. Hoos, and D. Poole (2004).
    Preference-based constrained optimization with CP-nets. Computational Intelli-
    gence, 20(2):137–157.
 3. A. P. Dempster (1967). Upper and lower probabilities induced by a multivalued
    mapping. The Annals of Mathematical Statistics, 38(2): 325-339.
 4. M. Gelain, M. S. Pini, F. Rossi and K. B. Venable (2007). Dealing with Incomplete
    Preferences in Soft Constraint Problems. In Proceedings of CP 2007: 286-300.
 5. M. Gelain, M. S. Pini, F. Rossi, K. B. Venable and N. Wilson (2010). Interval-
    valued soft constraint problems. Ann. Math. Artif. Intell. 58(3-4): 261-298.
 6. J. Lang and L. Van der Torre (2010). Preference change triggered by belief change:
    A principled approach. G. Bonanno, B. Löwe, W. van der Hoek (Eds.), Logic
    and the Foundations of Game and Decision Theory (LOFT 8), Lecture Notes in
    Computer Science. 6006: 86-111, Springer.
 7. F. Liu (2008). Changing for the better: Preference dynamics and agent diversity.
    Ph.D. dissertation, University of Amsterdam.
 8. W. Pang and S. Goodwin (1997). Constraint directed backtracking. Advanced Top-
    ics in AI. 1342: 47-56. Springer Verlag.
 9. M. S. Pini and F. Rossi (2005). Uncertainty in Soft Constraint Problems. In Pro-
    ceedings of CP 2005,3709: 865.
10. G. Shafer (1976). A Mathematical Theory of Evidence. Princeton University Press,
    Princeton, NJ.
11. P. Smets and R. Kennes (1994). The Transferable Belief Model. Artifical Intelli-
    gence, 66: 191–234.