<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Preliminary Study Towards a Fuzzy Model for Visual Attention</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Anca Ralescu</string-name>
          <email>anca.ralescu@uc.edu</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Isabelle Bloch</string-name>
          <email>isabelle.bloch@telecom-paristech.fr</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Roberto Cesar</string-name>
          <email>cesar@ime.usp.br</email>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>. EECS Department, University of Cincinnati</institution>
          ,
          <addr-line>ML 0030, Cincinnati, OH 45221</addr-line>
          ,
          <country country="US">USA -</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>. Institut Mines Telecom, Telecom Paristech, CNRS LTCI</institution>
          ,
          <addr-line>Paris</addr-line>
          ,
          <country country="FR">France -</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>. University of Sao Paulo</institution>
          ,
          <addr-line>IME, Sao Paulo</addr-line>
          ,
          <country country="BR">Brazil -</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Attention, in particular visual attention, has been a subject of studies in various disciplines, including cognitive science, experimental psychology, and computer vision. In cognitive science and experimental psychology the objective is to develop theories that can explain the attention phenomenon of cognition. In computer vision, the objective is to inform image understanding systems by hypotheses on the human visual attention. There is, however, very little influence of studies across these two disciplines. In a departure from this state of affairs, this study seeks to develop an algorithmic approach to visual attention as part of an image understanding system, by starting with a theory of visual attention put forward in experimental psychology. In the process, it will become useful to revise some of the concepts of this theory, in particular by adopting fuzzy set based representations and the necessary calculus for them.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1 Introduction</title>
      <p>As subject of human cognition, attention has attracted a
great interest from the fields of cognitive science and
experimental psychology.</p>
      <p>
        Visual attention is a wide field, largely addressed in the
literature covering different aspects. Some works related
to the present paper are briefly reviewed, without seeking
at exhaustivity. One approach relies on Gestalt theory, and
Gestalt and computer vision models are compared by
        <xref ref-type="bibr" rid="ref7">(Desolneux, Moisan, and Morel 2003)</xref>
        . Two sets of experiments
for Gestalt detection methods are carried out and compared
to computationally predicted results. Object size and noise
are the two parameters taken into account in these
experiments. The authors indicate that the qualitative thresholds
predicted by the proposed computational approach of gestalt
detection fit the human perception.
      </p>
      <p>
        Another approach is purely computational and based on
image information. An important review on visual
attention modeling is presented by
        <xref ref-type="bibr" rid="ref2">(Borji and Itti 2013)</xref>
        . The
important aspect of saliency-based attention is specifically
addressed in this review. Nearly 65 models are reviewed
and classified in a didactical taxonomy that helps
clarifying the field. Visual saliency refers to a bottom-up
phenomenon where some scene regions are detected as more
prominent than others due to some visual features. There are
different biological and computational approaches to model
such phenomena. For instance, the center-surround
hypothesis (a common issue for the analysis of receptive fields in the
retina) is a classical model for bottom-up saliency (Gao,
Mahadevan, and Vasconcelos 2008). In such settings, Gao and
co-authors (Gao, Mahadevan, and Vasconcelos 2008)
incorporate discriminant features and decision-theoretic model
for saliency characterization. Saliency detection is important
in many different imaging and vision applications
        <xref ref-type="bibr" rid="ref16">(Yan et al.
2013; Yang et al. 2013)</xref>
        . For instance, in medical imaging,
saliency maps are useful to guide model-based image
segmentation (Fouquier, Atif, and Bloch 2012), thus merging
top-down and bottom-up approaches.
      </p>
      <p>
        The mechanism of attention has been studied intensively
in the field of psychology and cognitive science,
        <xref ref-type="bibr" rid="ref9">(Kahneman 1973)</xref>
        ,
        <xref ref-type="bibr" rid="ref12">(Treisman and Gelade 1980)</xref>
        ,
        <xref ref-type="bibr" rid="ref13">(Treisman 1988)</xref>
        ,
        <xref ref-type="bibr" rid="ref14">(Treisman 2014)</xref>
        , (Humphreys 2014),
        <xref ref-type="bibr" rid="ref3">(Bundesen, Habekost,
and Kyllingsbaek 2005)</xref>
        <xref ref-type="bibr" rid="ref5">(Bundesen, Vangkilde, and Petersen
2014)</xref>
        . In this paper we focus on the theory of visual
attention introduced in
        <xref ref-type="bibr" rid="ref6">(Bundesen 1990)</xref>
        , where visual
recognition and attentional selection are considered as the task of
perceptual categorization, basically deciding to which
category an object or element of the visual field belongs.
      </p>
      <p>
        Following the notation of
        <xref ref-type="bibr" rid="ref6">(Bundesen 1990)</xref>
        , throughout
this paper, x is an input item, e.g. image or image region, or
more generally an item to be categorized of recognized. The
collection of all items x is denoted by S. A category is
denoted by i and the collection of all categories is denoted by
R. A category can stand for an ontological category (e.g., an
object, or a scene), or for subsets in the range of a particular
attribute (e.g., red for the attribute color). Regardless of the
situation the conceptual treatment of categories and/or items
is the same. E(x; i) denotes the event/statement ”x is in
category i”. When viewed as an event, one can talk about its
probability; when viewed as a statement, one can talk about
its truth or its possibility.
      </p>
      <p>From this point on this paper is organized as follows:
Section 2 contains a brief review of TVA concepts and
mechanisms - filtering and pigeonholing. Section 3 presents
the motivation for the introduction of fuzzy sets; the fuzzy
mechanisms of filtering and pigeonholing. Conclusions and
future research are in Section 4.</p>
    </sec>
    <sec id="sec-2">
      <title>2 TVA concepts and mechanisms of attention</title>
      <p>
        In this section, we review and comment the main concepts
and modeling steps of the Theory of Visual Attention (TVA)
by
        <xref ref-type="bibr" rid="ref6">(Bundesen 1990)</xref>
        .
      </p>
      <sec id="sec-2-1">
        <title>2.1 Attentional Weight</title>
        <p>One of the main concepts introduced in TVA is that of
attentional weight defined as follows:
w(x) = X (x; i) (i) (1)</p>
        <p>i2R</p>
        <p>What are the possible interpretations of the quantities in
Equation (1)? If (x; i) is interpreted as the salience of x for
category i, then w(x) could be interpreted as the salience of
x across the family of categories R, averaged with respect
to category pertinence. From the point of view of computer
vision, (x; i) is simply the output of an operator designed
to provide information for category i.</p>
        <p>Note that pertinence of a category is (or must be)
considered with respect to a task, which could be a categorization
at a higher semantic/ontological level. Adopting this point of
view, the product (x; i) (i) can then be interpreted as the
pertinence of item x to the task with respect to which
category i had pertinence (i). More precisely, one can define
(x; Ti) = (x; i) (i)
as the pertinence of x to Ti where Ti is the task to which
category i has pertinence value (i).</p>
        <p>For example, suppose that i is the color category “red” of
the attribute color. Furthermore, suppose that the color
category “red” has pertinence (red) to the task of identifying
visually an object such as, for instance, the “flag of some
country”. Let now x be a region in an image, and (x; red)
the output of evaluating it with respect to the color “red”.
Then (x; Tred) = (x; red) (red) is the pertinence of x
to the task Tred.</p>
        <p>Taking max/min with respect to x obtains:
xmax;red = arg max (x; Tred);
x2S
the region in the input which is most pertinent to Tred, and
xmin;red = arg min (x; Tred);
x2S
the region in the input which is least pertinent to Tred.</p>
        <p>Similarly, taking max = min over categories, yields
imax = arg max (i); imin = arg
i2R</p>
        <p>min
i2R; (i)&gt;0
(i)
the most/least pertinent categories respectively. The
condition (i) &gt; 0 ensures that categories which are not pertinent
at all, i.e. with (i) = 0, are not taken into account, so the
trivial case (imin) = 0 is never obtained. Then, for fixed
x, (x; imax), (x; imin) are the strengths of evidence for x
to be in the highest/lowest pertinence category, and
(x; Tmax) = (x; imax) (imax)
(x; imin) = (x; Tmin) (imin)
are the importance of x to the task corresponding to the
category of highest/lowest pertinence value. Versions of the
following “flag example” will be used in this paper to illustrate
various points.</p>
        <p>Example 1 Let T stand for the task to determine if an
object identified in an image corresponds to a “flag of some
country”. The decision is to be based on color information
only. Assume several color categories and their respective
pertinences as shown in Table 1.
In Equation (1) only those categories i with (i) &gt; 0
contribute to w(x). This means that categories which are not
pertinent (i.e., (i) = 0) are never considered for x, even
when (x; i) is very large.</p>
        <p>To summarize, with the interpretation of (x; i) (i) as
described above, the attentional weight w(x) defined by
Equation (1) is the cumulative pertinence of x to a task T ,
obtained from strength of the sensory evidence given by x to
all categories, in proportion to their pertinence to the task
T .
2.2</p>
      </sec>
      <sec id="sec-2-2">
        <title>Hazard Function</title>
        <p>
          In
          <xref ref-type="bibr" rid="ref6">(Bundesen 1990)</xref>
          the notion of a hazard function (x; i) is
introduced as (x; i) = P rob(E(x; i)), that is, the
probability that item x is in category i (e.g., image region x is red).
It is assumed (see 2nd assumption in
          <xref ref-type="bibr" rid="ref6">(Bundesen 1990)</xref>
          ) that
is computed as:
(x; i) = (x; i) (i)w(x)
(2)
where (x; i) and w(x) are as described above1, and (i) is
introduced to indicate a bias for category i. Since is
interpreted as a probability, (x; i) 2 [0; 1], which is ensured
when (x; i); (i); w(x) 2 [0; 1], without additional
constraints on these values. Moreover, when R is an exhaustive
set of exclusive (non-overlapping) categories, then should
be normalized so that Pi2R (x; i) = 1, in order to really
satisfy its interpretation from
          <xref ref-type="bibr" rid="ref6">(Bundesen 1990)</xref>
          as a
probability. More recently, in
          <xref ref-type="bibr" rid="ref5">(Bundesen, Vangkilde, and Petersen
2014)</xref>
          (i) is decomposed as
(i) = Ap(i)u(i)
(3)
where A 2 [0; 1] is the level of alertness, and p(i) and u(i)
are respectively, the prior probability and utility of category
i. One can imagine that A also varies with the category, in
1Note that the expression of
          <xref ref-type="bibr" rid="ref6">(Bundesen 1990)</xref>
          involves a
normalized version of w, i.e. w(x)= Px2S w(x). Here we implicitly
assume that w is normalized, in order to simplify equations.
which case A in Equation (3) is replaced by an Ai. This is
justified by the fact that one may be more alert to a
category than to others. In an image processing system, A, or Ai
could be tied to the performance of the image processing
operators used. The components p(i), u(i) of (i), and hence
(i), must also be tied up to a (higher level) task T . While
p(i) may be obtained from past data and experiments on the
task T , u(i) seems to be purely subjective, and to a large
extent, its role seems to overlap with that is (i). Plugging
w(x) and (i) in (2) results in
= A (x; i)p(i)u(i) Pj2R (x; j) (j)
= Ap(i)u(i)[ (x; i)2 (i)+
+ (x; i) Pj6=i (x; j) (j)]
(4)
which suggests that the most important role in computing
(x; i) is played by the sensory evidence. In particular, ’s
largest value is obtained when A = p(i) = u(i) = 1, (i.e.
under maximum alertness, maximum prior probability, and
maximum utility), and in that case (x; i) is a function only
of the sensory evidence. Stated differently, this means that
A, p(i) and u(i) can only decrease the value of (x; i).
However, they may provide a mechanism to account for different
types of subjective information, and of ranking the values of
(x; i) when they enter its definition as shown in Equations
(2) - (4). The justification in
          <xref ref-type="bibr" rid="ref5">(Bundesen, Vangkilde, and
Petersen 2014)</xref>
          of Equation (3) is based on the fact that when
either one of A, p(i), or u(i) is null, then (i) = 0.
However, the same result holds when these quantities enter the
definition of not through a product, but through other
operations, such as the min, or more generally, t-norms.
        </p>
        <p>
          The fact that the value of (x; i) decreases when
Ap(i)u(i) 6= 1 (i.e. at least one of these three values is less
than 1, u(i) for instance) can be interpreted as follows: x will
be less probably categorized in i if, for instance, the utility
for i is low, which means that we do not really care for this
category. This also goes with the interpretation as a rate of
encoding information in the memory, according to
          <xref ref-type="bibr" rid="ref6">(Bundesen 1990)</xref>
          , even without considering time information.
        </p>
        <p>
          The two mechanisms for visual attention proposed in
          <xref ref-type="bibr" rid="ref6">(Bundesen 1990)</xref>
          , filtering and pigeonholing, are described
next.
2.3
        </p>
      </sec>
      <sec id="sec-2-3">
        <title>Filtering</title>
        <p>
          Filtering
          <xref ref-type="bibr" rid="ref6">(Bundesen 1990)</xref>
          refers to the mechanism of
selecting an item x 2 S (given a higher level task), for a target
category i. This mechanism seeks to
(F1) increase (x; i) for some category i, while
(F2) not changing the conditional probability of E(x; i)
given that x is categorized.
        </p>
        <p>Filtering can be achieved by increasing w(x) as follows:</p>
        <p>For category j 2 R assume 0(j) = a (j), where
a &gt; 1. Then w(x) of equation (1) becomes w0(x) =
Pi2R;i6=j (x; i) i + (x; j) j0 = Pi2R;i6=j (x; i) i +
(x; j)a j &gt; w(x). Therefore, (x; i) becomes 0(x; i) =
(x; i) (i)w0(x) &gt; (x; i), which satisfies condition (F1)
above. Computing now P (x is i j x is categorized) yields:
P (x is i j x is categorized) = P</p>
        <p>(x;i) (ik)2R (x;k)
=</p>
        <p>(x;i) (i)w(x)
w(x) Pk2R (x;k)
= Pk2R (x;k)
(x;i)
(5)
which does not depend on w, hence satifies condition (F2).
In Equation (5) the numerator is (x; i) since
fx is ig</p>
        <p>fx is categorizedg
and therefore</p>
        <p>P (x is i; x is categorized) = P (x is i), while the
denominator uses an assumption on non-overlapping categories to
write P (x is categorized) as Pk2R (x; k). Dropping the
constraint of non-overlapping categories is discussed later
in this study.
2.4</p>
      </sec>
      <sec id="sec-2-4">
        <title>Pigeonholing</title>
        <p>
          For fixed item x 2 S, pigeonholing
          <xref ref-type="bibr" rid="ref6">(Bundesen 1990)</xref>
          refers
to the mechanism of selecting a category i 2 R (given a
higher level task), across a set of items S. It seeks to:
(P1) increase Px2S (x; i) for category i pertinent to the
task, such that
(P2) for all j 2 R, j 6= i, Px2S (x; j) does not change
Pigeonholing can be done by increasing (i) for some i 2 R
as follows: For category i 2 R, let i0 = a i, with a &gt; 1.
Then
0(x; i)
= (x; i) i0wx = (x; i)a iwx
&gt; (x; i) iwx = (x; i):
Summing up over x 2 S obtains
        </p>
        <p>x2R
P 0(i is selected) = X (x; i) i0wx &gt; P (i is selected);
(6)
which achieves (P1). At the same time, it is clear that for any
other category j 6= i, P (j is selected) does not change, and
hence (P2) is satisfied too.</p>
        <p>Equation (6) uses the assumption that items x are
nonoverlapping, for example that they form a partition of the
image. However, this partition need not be crisp, i.e. may
allow overlapping x’s, as for example these are stated in
qualitative terms. In such cases, Equation (6) does not hold.
Dropping the constraint of non-overlapping items, discussed
later, leads to a different interpretation of (x; i).
3</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Fuzzy Mechanisms for Visual Attention</title>
      <p>We consider in this section the situations when the values
of the attentional weight and/or category pertinence are not
exact. In such situations these values may be represented as
fuzzy sets, and therefore, the computation of the
categorization of an item must resort to calculus with fuzzy sets. First,
let us see why indeed such situations may arise.</p>
      <p>Recall that in its original definition, for a given input x
and category i, the strength of sensory evidence for E(x; i),
(x; i) 2 [0; 1]. Assuming that (x; i) is the output of an
operator/test for category i on item x, this output may be
inexact because of the inexact nature of the category i. For
example, if the category i = red of the attribute color, then
for a given input pixel value x this category holds ”more or
less” and it may not be useful to commit to an exact 0=1
value.</p>
      <p>Likewise, in its original definition, the pertinence of a
category, (i) conveys its importance. Obviously, given a
collection of visual categories, and task, they may be
distinguished along their pertinence values. Moreover, several
categories may have the same, maximum importance for the
given task. As an example, consider the pertinence of color
categories for the detection of an object which is known to
have one of two possible color categories, white or yellow,
from the collection of all possible color categories. In this
case, it is useful to be able to encode</p>
      <p>(white) = (yellow) = 1;
which would be possible when is considered as a
possibility distribution on the color categories, regardless of
the number of color categories allowed. By contrast,
using a probability based approach, the cardinality of R, the
collection of categories, restricts the values assigned to
these equally possible categories, to at most 0:5. That is,
(white) = (yellow) 0:5 with equality when R =
fyellow; whiteg.
3.1</p>
      <sec id="sec-3-1">
        <title>A new definition for w(x)</title>
        <p>The departure point for the new definition for w(x) is the
interpretation of a special case of Equation (1). Let Ra =
fi 2 R j (i) = ag and consider the special case R = R0 [
R1, that is, all categories in R are either ”fully” pertinent,
(i) = 1 (i 2 R1), or not pertinent (i) = 0 (i 2 R0). Then
(1) becomes
w(x) = X
Next let max = maxi2R1 , and recall that (x; i)
w(x)
max =
max</p>
        <p>X 1 =
i2R1
maxjR1j
1. Then
jR1j;
where jR1j denotes the cardinality of the set R1. That is,
w(x) is bounded by the number of categories i with
pertinence (i) = 1. If (x; i) = 1 for all i 2 R1 then w(x) is
exactly the number of such categories.</p>
        <p>This meaning of w(x) is very natural and appealing.
Indeed, one would expect the item x to count to the extent that
it supports more categories. To generalize this notion, define
for fixed x 2 S and fixed task T</p>
        <p>(x;T )(i) = (x; i) T (i)
the degree to which category i, pertinent to task T , is
supported by the (data) item x as shown by the strength of
sensory evidence, (x; i). Therefore, (x;T ) : R ! [0; 1] is the
membership of a fuzzy set on the set of categories. 2 Then
the weight of item x is now defined as the cardinality of this
fuzzy set. That is
we(x) = Card f(i; x(i)) j i 2 Rg
(7)
2In the following, assuming only one task, T , for ease of
notation, the subscript T will be dropped, to write x(i).</p>
        <p>
          Several formulas for the cardinality of a fuzzy set have been
put forward. Here, for illustration purposes, the definition
from
          <xref ref-type="bibr" rid="ref10">(Ralescu 1986)</xref>
          is used to obtain
        </p>
        <p>
          Card (f x(i) j i 2 Rg) (k) =
x;(k) ^ (1
x;(k+1)) (8)
where x;(k) denotes, the kth largest value of x( ), and
x;(jRj+1) = 0. Thus, the cardinality defined in Equation
(7) is a fuzzy set on f0; :::; jRjg. For an exact value of w(x)
the 0:5-level set of w(x) (which is an interval), or its classic
e
cardinality can be used
          <xref ref-type="bibr" rid="ref11">(Ralescu 1995)</xref>
          .
3.2
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>A new definition for (i)</title>
        <p>
          Following the discussion from Section 2.4, define
e(i) = minfA; p(i); u(i)g
(9)
As in the case of defined in (3), e(i) = 0 whenever A = 0,
or p(i) = 0, or u(i) = 0, and the discussion of
          <xref ref-type="bibr" rid="ref6">(Bundesen
1990)</xref>
          holds: that is, category i biases the selection to the
extent that the system is alert, and category i is possible and
useful. Alternatively, (9) means that the bias for the selection
of i cannot be greater than the system alertness, the
possibility of i or its utility. Furthermore, replacing the product by
min also eliminates the possibility of values for e smaller
than each one of A, p(i), and u(i), which is the well-known
drowning effect of multiplication of positive values smaller
than 1. More importantly, it should be mentioned that the
min can handle ordinal or qualitative values, without
needing specifying precise numbers. Specifying such precise
values might be difficult when subjective assessments are made.
By contrast, in the case of such assessments, ordinal or
qualitative values are usually easily produced.
        </p>
        <p>As already mentioned, in the fuzzy set framework, the
product and min are but two particular cases of a t-norm
(conjunction operator). A, p(i), and u(i) are interpreted
respectively, as degrees of alertness, possibility (rather than
probability) of i to be selected, and utility for the category i,
and the bias for i is defined as the conjunction of these. This
interpretation makes (9) meaningful beyond a mere
computational artifice. Another choice for defining e is to select
a more general, aggregation operator, H : [0; 1] [0; 1]
[0; 1] ! [0; 1], which would allow the contribution of more
than one of A, p(i), u(i) towards e.
3.3</p>
      </sec>
      <sec id="sec-3-3">
        <title>A new definition for (x; i)</title>
        <p>
          With the new definitions, we(x), and e of w(x) and
respectively, the meaning of (x; i) also changes from a probability
to a possibility, more precisely, P ossibility(x is i):
P ossibility(x is i) = H( (x; i); e(i); we(x))
(10)
where H is again an aggregation operator, and hence the
definition of (x; i) from
          <xref ref-type="bibr" rid="ref6">(Bundesen 1990)</xref>
          is a particular
case, when H is the product.
        </p>
        <p>
          For defining H, one may rely on the huge literature on
information fusion, for which the fuzzy sets theory provides
a number of useful operators (see e.g.
          <xref ref-type="bibr" rid="ref1 ref15 ref8">(Dubois and Prade
1985; Yager 1991; Bloch 1996)</xref>
          for reviews on fuzzy
fusion operators). The large choice offered by these operators
allows modeling different combination behaviors
(conjunctive, disjunctive, compromise, etc.), with different degrees
(e.g. the min is a less severe conjunction as the product).
Operators can also behave differently depending on whether
the values to be combined are small, large, of the same order
of magnitude, or having different priorities. The operators
H could also be set differently for the three values. For
instance and w, which depend on x and i could be combined
e
using an operator H1, and the result combined with e, which
depends on i only, using another operators H2.
        </p>
        <p>4</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Conclusions and Future Work</title>
      <p>This paper discussed an attentional model developed in the
field of psychology and cognitive science set in a
probabilistic framework. The basic concepts of this model were
discussed and an alternative, fuzzy set based approach was
suggested. In the fuzzy set framework, modeling would be
easier, more natural (for instance replacing numbers by
ordinal or qualitative values), and it would allow for more
flexible ways of combining the different terms. This discussion
paves the way for a new attentional model, the complete
development of it being left for future work.</p>
      <p>5</p>
    </sec>
    <sec id="sec-5">
      <title>Acknowledgments</title>
      <p>Anca Ralescu’s contribution was partially supported by a
visit to Telecom ParisTech.
Fouquier, G.; Atif, J.; and Bloch, I. 2012. Sequential
model-based segmentation and recognition of image
structures driven by visual features and spatial relations.
Computer Vision and Image Understanding 116(1):146–165.
Gao, D.; Mahadevan, V.; and Vasconcelos, N. 2008.
The discriminant center-surround hypothesis for bottom-up
saliency. In Advances in Neural Information Processing
Systems, 497–504.</p>
      <p>Humphreys, G. W. 2014. Feature confirmation in object
perception: Feature integration theory 26 years on from the
Treisman Bartlett lecture. The Quarterly Journal of
Experimental Psychology (just-accepted):1–49.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <string-name>
            <surname>Bloch</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          <year>1996</year>
          .
          <article-title>Information Combination Operators for Data Fusion: A Comparative Review with Classification</article-title>
          .
          <source>IEEE Transactions on Systems, Man, and Cybernetics</source>
          <volume>26</volume>
          (
          <issue>1</issue>
          ):
          <fpage>52</fpage>
          -
          <lpage>67</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <string-name>
            <surname>Borji</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Itti</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <year>2013</year>
          .
          <article-title>State-of-the-art in visual attention modeling</article-title>
          .
          <source>IEEE Transactions on Pattern Analysis and Machine Intelligence</source>
          <volume>35</volume>
          (
          <issue>1</issue>
          ):
          <fpage>185</fpage>
          -
          <lpage>207</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <string-name>
            <surname>Bundesen</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Habekost</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Kyllingsbaek</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <year>2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <article-title>A neural theory of visual attention: bridging cognition and neurophysiology</article-title>
          .
          <source>Psychological review 112</source>
          <volume>(2)</volume>
          :
          <fpage>291</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <string-name>
            <surname>Bundesen</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Vangkilde</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Petersen</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <year>2014</year>
          .
          <article-title>Recent developments in a computational theory of visual attention (tva)</article-title>
          .
          <source>Vision research.</source>
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <string-name>
            <surname>Bundesen</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <year>1990</year>
          .
          <article-title>A theory of visual attention</article-title>
          .
          <source>Psychological review 97</source>
          <volume>(4)</volume>
          :
          <fpage>523</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          <string-name>
            <surname>Desolneux</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Moisan</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Morel</surname>
            ,
            <given-names>J.-M.</given-names>
          </string-name>
          <year>2003</year>
          .
          <article-title>Computational gestalts and perception thresholds</article-title>
          .
          <source>Journal of Physiology-Paris</source>
          <volume>97</volume>
          (
          <issue>2</issue>
          ):
          <fpage>311</fpage>
          -
          <lpage>324</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          <string-name>
            <surname>Dubois</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Prade</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <year>1985</year>
          .
          <article-title>A Review of Fuzzy Set Aggregation Connectives</article-title>
          .
          <source>Information Sciences</source>
          <volume>36</volume>
          :
          <fpage>85</fpage>
          -
          <lpage>121</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          <string-name>
            <surname>Kahneman</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <year>1973</year>
          .
          <article-title>Attention and Effort</article-title>
          . Prentice-Hall.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          <string-name>
            <surname>Ralescu</surname>
            ,
            <given-names>A. L.</given-names>
          </string-name>
          <year>1986</year>
          .
          <article-title>A note on rule representation in expert systems</article-title>
          .
          <source>Information Sciences</source>
          <volume>38</volume>
          (
          <issue>2</issue>
          ):
          <fpage>193</fpage>
          -
          <lpage>203</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          <string-name>
            <surname>Ralescu</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <year>1995</year>
          .
          <article-title>Cardinality, quantifiers, and the aggregation of fuzzy criteria</article-title>
          .
          <source>Fuzzy sets and systems 69</source>
          <volume>(3)</volume>
          :
          <fpage>355</fpage>
          -
          <lpage>365</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          <string-name>
            <surname>Treisman</surname>
            ,
            <given-names>A. M.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Gelade</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          <year>1980</year>
          .
          <article-title>A feature-integration theory of attention</article-title>
          .
          <source>Cognitive psychology 12</source>
          (
          <issue>1</issue>
          ):
          <fpage>97</fpage>
          -
          <lpage>136</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          <string-name>
            <surname>Treisman</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <year>1988</year>
          .
          <article-title>Features and objects: The fourteenth bartlett memorial lecture</article-title>
          .
          <source>The Quarterly Journal of Experimental Psychology</source>
          <volume>40</volume>
          (
          <issue>2</issue>
          ):
          <fpage>201</fpage>
          -
          <lpage>237</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          <string-name>
            <surname>Treisman</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <year>2014</year>
          .
          <article-title>The psychological reality of levels of processing</article-title>
          .
          <source>Levels of processing in human memory 301- 330.</source>
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          <string-name>
            <surname>Yager</surname>
            ,
            <given-names>R. R.</given-names>
          </string-name>
          <year>1991</year>
          .
          <article-title>Connectives and Quantifiers in Fuzzy Sets</article-title>
          .
          <source>Fuzzy Sets and Systems</source>
          <volume>40</volume>
          :
          <fpage>39</fpage>
          -
          <lpage>75</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          <string-name>
            <surname>Yan</surname>
            ,
            <given-names>Q.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Xu</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Shi</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Jia</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <year>2013</year>
          .
          <article-title>Hierarchical saliency detection</article-title>
          .
          <source>In IEEE Conference on Computer Vision and Pattern Recognition (CVPR)</source>
          ,
          <fpage>1155</fpage>
          -
          <lpage>1162</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          2013.
          <article-title>Saliency detection via graph-based manifold ranking</article-title>
          .
          <source>In IEEE Conference on Computer Vision and Pattern Recognition (CVPR)</source>
          ,
          <fpage>3166</fpage>
          -
          <lpage>3173</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>