<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Rino Falcone</string-name>
          <email>rino.falcone@istc.cnr.it</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Cristiano Castelfranchi</string-name>
          <email>cristiano.castelfranchi@istc.cnr.it</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Institute of Cognitive Sciences and Technologies, National Research Council of Italy</institution>
          ,
          <addr-line>Rome</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>-Transitivity in trust is very often considered as a quite simple property, trivially inferable from the classical transitivity defined in mathematics, logic, or grammar. In fact the complexity of the trust notion suggests evaluating the relationships with the transitivity in a more adequate way. In this paper, starting from a socio-cognitive model of trust, we analyze the different aspects and conceptual frameworks involved in this relation and show how different interpretations of these concepts produce different solutions and definitions of trust transitivity.</p>
      </abstract>
      <kwd-group>
        <kwd>-component</kwd>
        <kwd>Trust</kwd>
        <kwd>Transitivity</kwd>
        <kwd>Degree of Trust</kwd>
        <kwd>Task definition</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>INTRODUCTION
Trust is becoming a research topic of major interest not only in
Artificial Intelligence (AI) and in Multi-Agent Systems
(MAS), but also more in general in Information and
Communication Technologies (ICT). The main reason of this
fact is that the more recent developments of the “interaction”
paradigm of computation, are driving more and more towards
the development of computational entities with a strong and
well defined autonomy. These entities have to
cooperate/conflict among them in conditions of open world for
achieving their own goals.</p>
      <p>In perspective, we are going towards an interaction scenario in
which artificial entities and humans are indistinguishable from
each other. In this view, the probability that we have to
interact or cooperate with entities we do not have any personal
experience with, will be growing, and the need of attributing
trustworthiness to the potential partners becomes a
fundamental prerequisite. And to model trust in the way in
which humans have always done it, being them in the
interaction loop, is particularly relevant.</p>
      <p>
        Many different approaches and models of trust were
developed in the last 15 years [
        <xref ref-type="bibr" rid="ref10 ref2 ref3 ref4 ref5 ref6 ref7 ref8 ref9">1, 2, 3, 4, 5, 6, 7, 8, 9, 10</xref>
        ]: they
contributed to clarify many aspects and problems about trust
and trustworthiness, although many issues remain to be
addressed and some elementary but not so trivial trust
properties are left in a contradictory form.
      </p>
      <p>One of them is the problem of trust transitivity. If X trusts Y,
and Y trusts Z: What about the trust relationship betwe!en X
and Z? Different and sometimes diverging answers were!given
to this problem. The question is not only theoret!ically
relevant; it is very relevant from the practical point of view,
for the reason we have just mentioned: acting in an open
world, interacting with new people/agents.</p>
      <p>In this paper we will present an analysis of the trust
transitivity in the case in which a socio-cognitive model of trust
is taken in consideration. Through this kind of model we are
able to evaluate and partially cope with the complexity that the
concept of transitivity introduces when applied to the trust
relationship.</p>
      <p>II.</p>
    </sec>
    <sec id="sec-2">
      <title>A SOCIO-COGNITIVE MODEL OF TRUST</title>
      <p>
        In our socio-cognitive model of trust [
        <xref ref-type="bibr" rid="ref11 ref12 ref13 ref14">11, 12, 13, 14</xref>
        ] we
consider trust as a layered notion, where the various more or
less complex meanings are embedded one into the other. We
analyzed the relevant relationships among these different
layers and studied the possible transitions among them.
We developed the analysis of the mental attitude and
disposition towards the trustee (considering beliefs like
evaluations and expectations); the intention and decision
based on the previous dispositions; the act of relying upon the
trustee’s expected behaviour; finally the social interaction and
relation between trustor and trustee.
      </p>
      <p>In particular we consider trust as a relational construct
between the trustor (X), the trustee (Y), about a defined (more
or less specialized) task (τ ):</p>
      <p>Trust ( X Y C " gX )
where are also explicitly present both the X’s goal (gX, respect
to which trust is activated) and the role of the context (C) in
which th!e relationship is going to happens. In fact, the
successful performance of the task τ will satisfy the goal gX.
So the X’s mental ingredients of trust are: the goal gX, and a
set of main beliefs:
Bel( X CanY (" ))
Bel( X WillY (" ))
Bel( X ExtFactY (" ))
where:
CanY (τ ) means that Y is potentially able to do τ (in the sense
that, under the given conditions, is competent, has the internal
powers, skills, know-how, etc) (and this is believed by X);
WillY (τ ) means that, under the given conditions, Y potentially
has the attributions for being willing, persistent, available, etc.,
on the task τ (and this is believed by X);
ExtFactY (τ ) means that potentially there are a set of external
conditions either favoring or hindering Y realizing the task τ
(and this is believed by X).</p>
      <p>In our model we also consider that trust can be graded: X can
have a strong trust that Y will realize the task (maybe 0.95 in
the range (0,1)); or even just a sufficient trust that Y will
achieve it (maybe 0.6 with a threshold of 0.55; and so on with
other possible values). For this we have introduced a
quantification of the degree of trust (DoTXYτ ) and, in general,
a threshold (σ ) to be overcome from this DoTXYτ .</p>
      <p>Given the previous analysis of the main components of the
trust attitude (gX, Bel (X CanY (τ )), Bel (X WillY (τ )), Bel (X
ExtFactY (τ ))), we can say that this degree is, on its turn,
resultant from the several quantifications of these components:
DoTXY" = f (DoCX (OppY (" )), DoCX (AbilY (" )), DoCX (WillY (" )))
where: f is in general a function that preserves monotonicity;
DoCX (OppY(τ )) is the X’s degree of credibility about the
external opportunities (positively or negatively) interfering
with Y’s action in realizing the task τ ;
DoCX (AbilityY(τ )) is the X’s degree of credibility about the
Y’s ability/competence to perform τ ;
DoCX (WillingnessY(τ )) is the X’s degree of credibility about
the Y’s willingness to perform τ .</p>
      <p>We are ignoring the subjective certainty of the pertinent
beliefs (how much sure is X of its evalutative beliefs about
that specific Y’s quality, that is a meta-belief; in fact we can
say that this factor is integrated with the other). At the same
time we are ignoring for now the value of the goal (gX).</p>
    </sec>
    <sec id="sec-3">
      <title>So trivially X will trust Y about the task τ if</title>
      <p>DoTXY" &gt; #
that means that a set of analogous conditions must be realized
about the other quantitative elements (DoCX (OppY(τ )), DoCX
(AbilityY(τ )), Do!CX (WillingnessY(τ ))). We do not consider in
this paper the detailed analysis of how the degree of trust is
resulting by the more elementary components and we also
omit of considering the potential positive and negative
interferences among the components themselves.</p>
      <p>Introducing also the concept of Y’s trustworthiness degree
(TrustworthinessYX(τ)) with respect to X about the task τ , we
can say that from:</p>
      <p>DoTXY" &gt; #
it derives that</p>
      <p>Bel( X TrustworthinessYX (" ) &gt; #)
In general σ =Σ .</p>
      <p>!
!</p>
    </sec>
    <sec id="sec-4">
      <title>TRANSITIVITY IN TRUST</title>
      <p>Many authors have questioned whether the transitive property
can be applied to trust. In more specific words many of them
have presented this problem in this way:
If X trusts Y, and Y trusts Z: What about the trust relationship
between X and Z?
Their answers are different and very briefly we will analyze
some of them in §IV.</p>
      <p>We are now interested to translate this problem in our terms of
trust.</p>
      <p>First of all, we do not consider the unspecified case “X trusts
Y” because in our model an agent has to trust another agent
with respect a task (either very well defined or less defined
and abstract); this task directly derives from the goal the
trustor has to reach with the trust attribution. So we have to
transform “X trusts Y” in “X trusts Y about τ ”. And given the
graded qualification of trust we have that:</p>
      <p>DoTXY" &gt; #
this means in particular that X believes that Y is potentially
able and willing to do τ and that the external conditions in
which Y will perform its task are at least not so opposite to the
task realization!(may be also they are neutral or favorable).
So this Y’s trustworthiness with respect to X
(perceived/believed by X) is based on these specific beliefs.
At the same way “Y trusts Z” becomes “Y trusts Z about τ 1”
(about the difference between τ and τ 1, see later) with the
same particular Y’s beliefs about Z and the external
conditions.</p>
      <p>Also in this case we can say that there is a threshold to be
overcome and the condition:</p>
      <p>DoTYZ"1 &gt; #1
successfully satisfied in case of trust attribution.</p>
      <p>If we have to consider the trust relationship between “X and
Z” as a con!sequence of the previous trust relationships
between “X and Y” and between “Y and Z” we have to define
the task on which this relationship should be based (question
of assimilation between τ and τ 1) and the degree of trust that
must be overcome even from X:</p>
      <p>DoTXZ" &gt; # 2
with the consideration of the threshold σ 2.</p>
      <p>The role of the trust threshold is quite complex and can have
an overlappin!g with the ingredients of trust. We strongly
simplify in this case considering σ as dependent just from the
specific intrinsic characteristics of the trustor (those that
define an agent intrinsically: prudent, reckless, and so on)
independently from the external circumstances (on the
contrary, these factors affect the degree of trust, by affecting
the more elementary beliefs above showed).</p>
      <p>So, we can say that in this approximation (for the same agent
the trust threshold is always the same):
σ = σ 2
In the case in which all the agents are defined as having the
same intrinsic characteristics (this fact is possible in the case
of artificial entities), we can also say that:</p>
      <p>σ = σ 1 = σ 2
Moreover, as we just saw, not less important in our approach
is that trust is an expectation and a bet grounded on and
justified by certain beliefs about Y. X trusts Y on the basis of
the evaluation of Y’s "virtues/qualities", not just on the basis
of a statistical sampling, some probability.</p>
      <p>The evaluations about the needed “qualities” of Y for τ are the
mediator for the decision to trust Y. This mediation role is
fundamental also in trust transitivity.</p>
      <p>Let us now consider the case of the differences between the
tasks in the different relationships.</p>
      <p>For the trust transitivity the two tasks should be the same (τ !=
τ 1). Is this equality enough? !
Suppose for example that there are 3 agents: John, Mary and!
Peter; and suppose that John trusts Mary about “organizing
scientific meetings” (task τ ), at the same time Mary trusts
Peter about “organizing scientific meetings” (again task!τ ).
Can we deduce that, given the transitivity of trust: John trusts
Peter about “organizing scientific meetings”? Is in fact
transferable that task evaluation? Given the trust model
defined in §2 the situation is more complex and there are
possible pitfalls lurking: Mary is the central node for that trust
transfer and she plays different roles (and functions) in the
first case (when her trustworthiness is about to realize the tas!k
τ , and in the second case (when her trustworthiness is about
evaluating the Peter’s trustworthiness on the task τ ).
The situation is even clearer if we split in the example the two
kinds of competences: X trusts very much Y as medical
doctor; X knows that Y trusts Z as mechanic; will X trust Z !as
mechanic? Not necessarily at all: if X believes that Y is !a
good evaluator of mechanics he will trust Z; but, if X believes
that Y is a very naive in this domain, and is frequently!
swindled by people, he will not trust Z. In the previous
example the transition looks more plausible, natural, just
because the task τ is the same, and it is reasonable (bu!tnot
necessary) that if Y is very skilled and competent in task τ , she
will also be a good evaluator of other people on the same task.
So for considering transitivity of trust as a valid property (i!n
the classical way in which it is defined) in these types of
situations, we have to assimilate the task with the evaluation
of that task itself: !
Bel( X TrustworthinessYX (" ) &gt; # )
implies Bel( X TrustworthinessYX (eval(" )) &gt; # )
In words, X has to believe that if Y is trustworthy on the task
τ , it is also trustworthy on the meta-task of evaluating τ (in
! both the situations the X’s mental ingredients defining the trust
in Y allow to overcome the threshold).</p>
      <p>We do not consider in this paper the truthfulness of this
hypothesis (and the consequent properties both in the more
elementary mental ingredients of the interacting actors, and in
the tasks’ features): in fact, our trust model is apt to analyze in
detail this problem. However, we want to underline the
difference of the involved tasks in the relationships and the
necessity of taking into consideration these differences before
generally speaking of trust transitivity.</p>
      <p>So resuming we have the more basic case of the relationship
between trust and transitivity so defined (case A):
if
iA) DoTXY" &gt; #X (X trusts Y about τ ) and
iiA) DoTYZ" &gt; #Y (Y trusts Z about τ ) and
iiiA) Bel( X DoTYZ" &gt; #Y ) (X believes that Y trusts Z about τ )
and
ivA) Bel(X TrustworthinessYX (" ) &gt; # X ) implies
Bel(X TrustworthinessYX (eval(" )) &gt; # X )
(Y is equally trustworthy in the realization of the task and in
evaluating others performing that task; this implication also
!
takes into account the fact that X has a good consideration
about the adopted threshold by Y, "Y )
then
vA) DoTXZ" &gt; #X (X trusts Z about τ )
In the case (B) in wh!ich we assume σ =σ X=σ Y (the trust
threshold is the same for each agent and for each relation), we
have:
if
iB) DoTXY" &gt; # (X trusts Y about τ ) and
iiB) DoTYZ" &gt; # (Y trusts Z about τ ) and
iiiB) Bel( X DoTYZ" &gt; # ) (X believes that Y trusts Z about τ )
and
ivB) Bel(X TrustworthinessYX (" ) &gt; # ) implies
Bel(X TrustworthinessYX (eval(" )) &gt; # )
(Y is equally trustworthy in the realization of the task and in
evaluating others performing that task)
then!
vB) DoTXZ" &gt; # (X trusts Z about τ )
The fact that are true: DoTXY" &gt; # , DoTYZ" &gt; # , and
DoTXZ" &gt; #
does not mean that necessarily DoTXY" = DoTYZ" = DoTXZ" .
As we have seen !in the §2 these!degrees are dependent from
the internal beliefs on the different components, and they are
resulting from different sources not necessarily all common to
the involved agents. !
We have also to underline that (iiA) or (iiB) should not
necessarily be true, the import thing is that are respectively
true (iiiA) or (iiiB).</p>
      <p>In the case (C) in which the tasks are different (" # " ' ), we
have (we are also assuming that σ =σ X=σ Y):
!
!
!
!
iiiC) Bel( X DoTYZ"' &gt; # ) (X believes that Y trusts Z about τ )
and
ivC) Bel(X TrustworthinessYX (" ) &gt; # ) implies
Bel(X TrustworthinessYX (eval(" ')) &gt; # )
(Y is equally trustworthy in the realization of the task and in
evaluating others performing that specific different task)
then!
vC) DoTXZ"' &gt; # (X trusts Z about τ ’)</p>
      <p>In this last case C, the transitivity essentially depends from
the implication reported in the formula (iiiC); are there
elements in the reasons (believed by X) for trusting Y on the
task τ that (in X’s view) are sufficient also for trusting Y on the
different task τ ’ ?
A. Trust and Transitivity in the delegated subtasks
Another interesting case of the relationship between trust and
transitivity is when we have the following situation:
“X trusts Y about τ ” and, in realizing the task (τ ), Y delegates
parts of the task itself to other agents Z, W (for example the
delegated subtasks are respectively τ 1 and τ 2).</p>
      <p>Then if we suppose that X is aware of this delegation; what we
can say (on the basis of the trust relationship between X and
Y) about the trust between X and Z with respect to τ 1? And
between X and W with respect τ 2? !
!
Suppose, for example, that John trusts Mary about “organizing
a scientific meeting” and that Mary delegates Peter to
“organize the registration process”, and delegates Alice to!
“organize the sponsoring of the event”. What about the John’s!
direct trust on Peter (as organizing the registration process)
and the John’s direct trust on Alice (as organizing the
sponsoring of the event)? Are these trust relationships th!e
same of the Mary’s ones? How are they mediated by the!
John’s trust on Mary? !
We have that:
DoTJohn Mary" &gt; # and
DoTMary Peter"1 &gt; # and
DoTMary Alice"2 &gt; #
But from these assumptions does not necessarily follow that:
DoTJohn Peter"1 &gt; # and/or DoTJohn Alice"2 &gt; #
In fact, John should know how exactly the delegation of the
subtasks is realized and on what basis is founded. In a trust
relationship, as!we have seen in §2, are involved not only
qualities about abilities and skills but also qualities about
willingness, availability, and so on. So these others qualities
could be elicited by the specific relation with the trustor
(delegating agent) and in some cases are strictly related with
the interaction history among the agents (see §3.2 for a more
detailed analysis of this). In these cases is more difficult for
John to evaluate which could be the Peter’s and Alice’s
performances (and then their trustworthiness).</p>
      <p>In addition may be there is a particular interaction among the
subtasks and the main task in which Mary plays a specific role
of integration and substitution (in presence of shortcomings of
the other agents (Peter and Alice)) that is essential for the
success of the complete task. Also in this case John trusting
Peter and Alice, at the same way of Mary, should be aware to
be able of playing (in case of necessity) the same role of
integration and substitution.</p>
      <p>So we can say that for applying the trust transitivity to the
cases of subtasks delegation, we have to analyze in deep
detail: on the one hand the set relationships among task and its
subtasks (and how they are distributed among the agents in
play), and, on the other hand, how the executing agents are
motivated and activated in the task realization by the
relationship with the trustor.</p>
      <p>Resuming in the case (D) of delegated subtasks we can say
that:
if
iD) DoTXY" &gt; # and
iiD) DoTYZ"1 &gt; # where τ 1 is a subtask of τ (realizing τ 1 are
achieved parts of the result of τ ) and
iiiD) Bel( X DoTYZ"1 &gt; #) and
ivD) DoTYW"2 &gt; # where τ 2 is a subtask of τ (realizing τ 2 are
achieved parts of the result of τ ) and
vD) Bel( X DoTYW"2 &gt; #) and
viD) Bel(X TrustworthinessZX ("1) # TrustworthinessZY ("1)) and
viiD) Bel(X TrustworthinessWX (" 2) # TrustworthinessWY (" 2))
then
viiiD) DoTXZ"1 &gt; # and DoTXW"2 &gt; # .
! B. Competen!ce and Willingness in Transitivity</p>
      <p>The need for a careful qualitative consideration of the nature
of the link between the trustor and the trustee, is even more
serious.</p>
      <p>Not only it is fundamental (as we have argued) to make
explicit and do not forget the specific “task” (activity, and thus
“qualities”) X is trusting Y or Z about, but it is even necessary
to consider the different dimensions/components of the trust
disposition (evaluation), decision, and relation.
In our model (a part from the basic thought and feeling that I
have not to worry about Y, that there is no danger from Y's
side, that I do not need diffidence and a defensive attitude),
trust has two basic nucleuses:
(i) Y’s competence, ability, for correctly performing the
delegated task;
(ii) Y’s willingness to do it, to act as expected.</p>
      <p>The two dimensions (and ‘virtues’ of Y) are quite independent
on each other: Y might be very well disposed and willing to
do, but not very competent or unable; Y might be very expert
and skilled, but not very reliable: unstable, unpredictable, not
well disposed, insincere, dishonest, etc.</p>
      <p>Now, this (at least) double dimensions affect transitivity. In
fact, even assuming that the competence is rather stable (see
below) (and that Y is a good evaluator of Z’s competence) not
necessarily Z’s willingness is equally stable and transferable
from Y to X. This is a more relation-based dimension.
Y was evaluating Z’s willingness to do as expected on the
basis of their specific relation. Is Z a friend of Y? Is there a
specific benevolence, or values sharing, or gratitude and
reciprocity, or obligation and hierarchical relation, etc.? Not
necessarily the reasons that Z would have for satisfying Y’s
expectation and delegation would be present (or equally
important) towards X. X’s relation with Z might be very
different. Are the reasons/motives motivating Z towards Y,
and making him reliable, transferable or equally present
towards X? Only in this case it would be reasonable for X to
adopt Y’s trustful attitude and decision towards Z.
Only certain kinds of relations will be generalized from Y to
X; for example, if Y trusts Z only because it is an economic
exchange, only for Z’s interest in money, reasonably X can
become a new client of Z; or if Y relies on Z because Z is a
charitable person, generously helping (without any prejudice
and discrimination) poor suffering people, and X is in the
same condition of Y, than also X can trust in Z.
In sum,
•</p>
      <p>Y’s expectation about Z’s reliable behavior, in
particular about Z’s “adoption” of Y’s goal (help,
etc.) depends on the relationship between Y and Z,
and in particular on Z’s attitude towards Y (and
reasons for goal-adoption); if the relation between X
and Z, and in particular Z’s attitude towards X (and
reasons for goal-adoption), would be analogous, then
the trust “transfer” would be reasonable.</p>
      <p>In the previous analysis and examples we have partially solved
this problem considering the concept of Y’s trustworthiness
towards X about the task τ , in which the trustworthiness of the
trustee (Y) also depends from the trustor (X) and the task (τ ).
C. Trust Dynamics affects Transitivity</p>
      <p>
        Moreover, we have shown ([
        <xref ref-type="bibr" rid="ref13">13</xref>
        ], [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ]) that Z’s willingness,
and even ability, can be affected, increased, by Y’s trust and
reliance (this can affect Z’s commitment, pride, effort,...
attention, study,...). Z’s trustworthiness is improved by Y’s
trust and delegation. And Y might predict and calculate this in
her decision to rely on Z.
      </p>
      <p>However, not necessarily the effect of Y on Z’s
trustworthiness will be produced also by another trustor. Thus,
also this will affect “transitivity”: suppose that Y’s trust and
delegation to Z makes him more trustworthy, improves Z’s
willingness or ability (and Y trusts and relies on Z on the basis
of such expectation); not necessarily X’s reliance on Z would
have the same effect. Thus even if X knows that Y reasonably
trusts Z (for something) and that he is a good evaluator and
decision-maker, not necessarily X can have the same trust in Z,
since perhaps Z’s trustworthiness would not be equally
improved by X’s reliance.</p>
      <p>Trust 'dynamics' between Y and Z, is not automatically
identical to trust “dynamics” between X and Z.</p>
      <p>IV.</p>
    </sec>
    <sec id="sec-5">
      <title>TRANSITIVITY AND TRUST: RELATED WORK</title>
      <p>
        The necessity of modeling trust in the social networks is
becoming more and more important, and a set of new
definitions are emerging in different domains of computing:
computational trust, trust propagation, trust metrics, trust in
web-based social networks, and so on. Most of these concepts
are strictly linked with the goal of inferring trust relationships
between individuals that are not directly connected in the
networks. For this reason the concept of trust transitivity is
very often considered and used in these approaches.
A relevant example is given from the Josang’s approach; he
introduces the subjective logics (an attempt of overcoming the
limits of the classical logics) for taking in consideration the
uncertainty, the ignorance and the subjective characteristics of
the beliefs [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ]. Using this approach Josang addressed the
problem of trust transitivity in different works [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ], till the last
developments, see [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ], where it is recognized the intrinsic
cognitive nature of this phenomenon. However, the main
limits of this approach are that trust is in fact the trust in the
information sources; and the transitivity regards two different
tasks (referred to our formalism: τ ≠ τ 1: X has to trust the
evaluation of Y (task τ ) with respect Z as realizing another
task (task τ 1, for example as mechanic). As we showed
before, this difference is really relevant for the transitivity
phenomenon. In addition, also the first task (Y as evaluator) is
just analyzed with respect to the property of sincerity (and this
is a confirmation of the constrained view of trust phenomenon;
they write: “A’s disbelief in the recommending agent B means
that A thinks that B consistently recommends the opposite of
his real opinion about the truth value of x”; where A, B, and x
are, in our terms, respectively X, Y and τ 1). But in trusting
someone as evaluator of another agent (with respect to a
specific task), I have also to consider his competence as
evaluator, not just his sincerity. Trust is based on ascribed
qualities. Y could be completely sincere but at the same time
completely inappropriate to evaluate that task. So the limits of
this approach of an adequate treatment of the trust transitivity
are quite clear.
      </p>
      <p>
        Many other authors [
        <xref ref-type="bibr" rid="ref18 ref19">18, 19</xref>
        ], developed algorithms for
inferring trust among agents not directly connected. These
algorithms differ from each other in the way they compute trust
values and propagate those values in the networks. In any case
when trust transitivity is introduced, this phenomenon is not
analyzed with respect the complexity it contains and that we
tried to explain in the previous paragraphs.
The relationships between Transitivity and Trust (like
transitivity in partial order and equivalence relations) represent
an important element to well-understand the intimate nature of
the trust concept. Since trust is considered the glue of social
interactions, its complex nature has to be deeply investigated
if we want realize a careful and precise model to be transfered
in the artificial societies. The analysis of the trust properties is,
in this sense, a very useful tool. And transitivity (among the
other properties) is one of the more interesting one. In this
paper we analyzed the relationships between trust and
transitivity, showing in particular, how this analysis implicates
the evaluation of the tasks involved in the different
relationships, of the qualities of the agents in play in those
specific tasks and of the particular relationships among the
agents (and of their interaction histories and contexts).
      </p>
      <p>We tried to show how a too trivial simplification of the
transitive property when applied to the trust concept leads to
wrong conclusions about the model of the phenomenon we are
studying.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <string-name>
            <surname>Marsh</surname>
            ,
            <given-names>S.P.</given-names>
          </string-name>
          , (
          <year>1994</year>
          ),
          <article-title>Formalising Trust as a computational concept</article-title>
          .
          <source>PhD thesis</source>
          , University of Stirling. Available at: http://www.nr.no/abie/papers/TR133.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <surname>Jonker</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Treur</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          , (
          <year>1999</year>
          ),
          <article-title>Formal Analysis of Models for the Dynamics of Trust based on Experiences</article-title>
          ,
          <source>Autonomous Agents '99 Workshop on "Deception</source>
          ,
          <article-title>Fraud and Trust in Agent Societies"</article-title>
          , Seattle, USA, May 1, pp.
          <fpage>81</fpage>
          -
          <lpage>94</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>Barber</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Kim</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          , (
          <year>2000</year>
          ),
          <article-title>Belief Revision Process based on trust: agents evaluating reputation of information sources</article-title>
          ,
          <source>Autonomous Agents 2000 Workshop on "Deception</source>
          ,
          <article-title>Fraud and Trust in Agent Societies"</article-title>
          , Barcelona, Spain, June 4, pp.
          <fpage>15</fpage>
          -
          <lpage>26</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <surname>Jones</surname>
            ,
            <given-names>A.J.I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Firozabadi</surname>
            ,
            <given-names>B.S.</given-names>
          </string-name>
          , (
          <year>2001</year>
          ),
          <article-title>On the characterization of a trusting agent: Aspects of a formal approach</article-title>
          . In Castelfranchi,
          <string-name>
            <given-names>C.</given-names>
            ,
            <surname>Tan</surname>
          </string-name>
          ,
          <string-name>
            <surname>Y.H.</surname>
          </string-name>
          , (Eds.), Trust and Deception in Virtual Societies. Pp.
          <volume>55</volume>
          -
          <fpage>90</fpage>
          . Kluwer, Dordrecht.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>Resnick</surname>
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zeckhauser</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          , (
          <year>2002</year>
          ),
          <article-title>Trust among strangers in internet transactions: Empirical analysis of eBay's reputation systems</article-title>
          . In Baye, R. (Editor),
          <article-title>The economico of the internet and e-commerce</article-title>
          . Vol.
          <volume>11</volume>
          of Advances in Applied Microneconomics. Elsevier Science.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>Yu</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Singh</surname>
            ,
            <given-names>M.P.</given-names>
          </string-name>
          , (
          <year>2003</year>
          ),
          <article-title>Searching social networks</article-title>
          .
          <source>In Proceedings of the second international joint conference on autonomous agents and multi-agent systems (AAMAS)</source>
          .
          <source>Pp</source>
          .
          <volume>65</volume>
          -
          <fpage>72</fpage>
          . ACM Press.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>Singh</surname>
            ,
            <given-names>M.P.</given-names>
          </string-name>
          , (
          <year>2003</year>
          ),
          <article-title>Trustworthy Service Composition: challenger and Research Questions</article-title>
          . in Falcone, R.,
          <string-name>
            <surname>Barber</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Korba</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Singh</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          , (Eds.),
          <source>Trust Reputation, and Security: Theories and Practice. Lecture Notes on Artificial Intelligence, n°2631</source>
          , Springer.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <surname>Sabater</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          (
          <year>2003</year>
          ),
          <article-title>Trust and Reputation for Agent Societies</article-title>
          ,
          <source>PhD thesis</source>
          , Universitat Autonoma de Barcelona.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <surname>Hang</surname>
            ,
            <given-names>C.W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Singh</surname>
            ,
            <given-names>M.P.</given-names>
          </string-name>
          , (
          <year>2009</year>
          ),
          <article-title>Operators for Propagating Trust and their Evaluation in Socila Networks</article-title>
          .
          <source>In Proceedings of the eight international joint conference on autonomous agents and multiagent systems (AAMAS).</source>
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>Ziegler</surname>
            ,
            <given-names>C.N.</given-names>
          </string-name>
          , (
          <year>2009</year>
          ),
          <article-title>On Propagation Interpersonal Trust in Social Network</article-title>
          . In Golbeck, J., (Ed.),
          <article-title>Computing with Social Trust</article-title>
          . Human Computer Interaction Series, Springer.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>Castelfranchi</surname>
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Falcone</surname>
            <given-names>R.</given-names>
          </string-name>
          , (
          <year>1998</year>
          )
          <article-title>Principles of trust for MAS: cognitive anatomy, social importance, and quantification</article-title>
          ,
          <source>Proceedings of the International Conference of Multi-Agent Systems (ICMAS'98)</source>
          , pp.
          <fpage>72</fpage>
          -
          <lpage>79</lpage>
          , Paris, July.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>Falcone</surname>
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Castelfranchi</surname>
            <given-names>C.</given-names>
          </string-name>
          , (
          <year>2001</year>
          ).
          <article-title>Social Trust: A Cognitive Approach, in Trust and Deception in Virtual Societies by Castelfranchi C.</article-title>
          and
          <string-name>
            <surname>Yao-Hua Tan</surname>
          </string-name>
          (eds), Kluwer Academic Publishers, pp.
          <fpage>55</fpage>
          -
          <lpage>90</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>Falcone</surname>
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Castelfranchi</surname>
            <given-names>C.</given-names>
          </string-name>
          (
          <year>2001</year>
          ),
          <article-title>The socio-cognitive dynamics of trust: does trust create trust? In Trust in Cyber-societies: Integrating the Human</article-title>
          and
          <string-name>
            <surname>Artificial Perspectives R. Falcone</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Singh</surname>
            , and
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>Tan</surname>
          </string-name>
          (Eds.), LNAI 2246 Springer. pp.
          <fpage>55</fpage>
          -
          <lpage>72</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <surname>Castelfranchi</surname>
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Falcone</surname>
            <given-names>R.</given-names>
          </string-name>
          , (
          <year>2010</year>
          ),
          <source>Trust Theory: A Socio-Cognitive and Computational Model</source>
          , John Wiley and Sons,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <surname>Josang</surname>
            <given-names>A.</given-names>
          </string-name>
          ,
          <article-title>A logic for uncertain probabilities</article-title>
          ,
          <source>International journal of Uncertain, Fuzziness and Knowledge-based Systems</source>
          ,
          <volume>9</volume>
          (
          <issue>3</issue>
          ):
          <fpage>279</fpage>
          -
          <lpage>311</lpage>
          , June,
          <year>2001</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <surname>Josang</surname>
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gray</surname>
            <given-names>E.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Kinateder</surname>
            <given-names>M.</given-names>
          </string-name>
          ,
          <source>Simplification and Analysis of Transitive Trust Networks, Web Intelligence and Agent Systems</source>
          ,
          <volume>4</volume>
          (
          <issue>2</issue>
          ):
          <fpage>139</fpage>
          -
          <lpage>161</lpage>
          ,
          <year>2006</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <surname>Bhuiyan</surname>
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Josang</surname>
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Xu</surname>
            <given-names>Y.</given-names>
          </string-name>
          ,
          <article-title>An analysis of trust transitività taking base rate into account</article-title>
          ,
          <source>In: proceeding of the Sixth International Conference on Ubiquitous Intelligence and Computing</source>
          ,
          <volume>7</volume>
          -
          <issue>9</issue>
          <year>July 2009</year>
          , University of Queensland, Brisbane,
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Han</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shen</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <article-title>Transitive trust to executables generated during runtime</article-title>
          ,
          <source>second International Conference on Innovative Computing, Information and Control</source>
          ,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <surname>Golbeck</surname>
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hendler</surname>
            <given-names>J.</given-names>
          </string-name>
          ,
          <article-title>Inferring binary trust relationships in web-based social networks</article-title>
          ,
          <source>ACM Transactions on Internet Technology</source>
          ,
          <volume>6</volume>
          (
          <issue>4</issue>
          ),
          <fpage>497</fpage>
          -
          <lpage>529</lpage>
          ,
          <year>2006</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>