<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Di erent learning architectures contribute to value prediction in human cognition</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Frederic Alexandre</string-name>
          <email>frederic.alexandre@inria.fr</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Inria Bordeaux Sud-Ouest</institution>
          ,
          <addr-line>200 Avenue de la Vieille Tour, 33405 Talence</addr-line>
          ,
          <country country="FR">France</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Institut des Maladies Neurodegeneratives, Universite de Bordeaux</institution>
          ,
          <addr-line>CNRS, UMR 5293, Bordeaux</addr-line>
          ,
          <country country="FR">France</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>LaBRI, Universite de Bordeaux</institution>
          ,
          <addr-line>Bordeaux INP, CNRS, UMR 5800, Talence</addr-line>
          ,
          <country country="FR">France</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Pavlovian conditioning plays a fundamental role in our cognitive architecture, by its capacity to bind values to stimuli. Due to the multifarious characteristics of this learning mode, many approaches in machine learning have been proposed to implement it in arti cial intelligence models. Considering the complementary properties of these models and inspired by biological evidences, we propose not to select the best model but rather to combine them, thereby forming a cognitive architecture. From a functional point of view, we report the good properties and performances of this architecture. From a methodological point of view, this work highlights the interest of de ning a cognitive function at the algorithmic level, binding its general properties to its implementation details. It is also proposed that it is a fruitful approach to decipher and organize many information extracted by modern approaches in neuroscience, towards the de nition of a global cognitive architecture.</p>
      </abstract>
      <kwd-group>
        <kwd>Value prediction</kwd>
        <kwd>Bio-inspiration</kwd>
        <kwd>Competiton</kwd>
        <kwd>Cognitive architecture</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        Having particularly in mind the brain performing a cognitive function, D. Marr
[
        <xref ref-type="bibr" rid="ref28">28</xref>
        ] has argued that, to be fully understood, an information processing system
must be examined at three di erent levels of description. The function itself
can be globally described at the computational level. The representations and
the processes used to implement the function are described at the algorithmic
level. The way of computing them are described at the implementational level.
This later level, refering to the brain, corresponds to the brain circuitry but Z.
Pylyshyn [
        <xref ref-type="bibr" rid="ref35">35</xref>
        ] has proposed a similar tri-level description in a purely cognitive
purpose and has simply de ned the lower level as the functional architecture.
      </p>
      <p>The decomposition is interesting because it makes it clear that not a unique
algorithm can carry out the function and not a unique circuit can implement the
algorithm. This decomposition is also interesting in the framework of arti cial
intelligence because it helps structuring steps to build a model. On one side is
the cognitive function to model. On the other side are formalisms available (e.g.
neuronal computations). In between are probably the more critical and
important steps: de ning the structures and the sequences of information processing
to bridge both sides (e.g. neuronal architecture and learning rules). The classical
approach in arti cial intelligence is top-down: from a cognitive function to be
modeled, an algorithm and its implementation in a certain formalism are
proposed. The approach can be also bottom-up, particularly if the problem is set
in computational neuroscience: from experimentally observed neuronal circuits
and mechanisms, it is questioned which ones are active to emulate the cognitive
function generally associated to the neuronal region.</p>
      <p>In this paper, having in mind this decomposition, I extend a series of works
performed in my team to model pavlovian conditioning from di erent points of
view, in order to propose a new view of this fundamental learning mode, with
an impact on cognitive modeling as well as arti cial intelligence and machine
learning.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Pavlovian conditioning</title>
      <p>
        Pavlovian conditioning [
        <xref ref-type="bibr" rid="ref33">33</xref>
        ] is an elementary learning process, present in most
species, by which an animal learns to predict biologically signi cant stimuli
announcing pain or pleasure (called Unconditioned Stimuli, US, e.g. some food)
by some other stimuli (called Conditional Stimuli, CS, e.g. a bell ringing before
food delivery) that anticipates its venue.
      </p>
      <p>
        This learning paradigm has been extensively studied in behavioral
neuroscience because it is easy to observe (US anticipation is associated to other
socalled pavlovian responses like salivation in the appetitive case and freezing in
the aversive case) and not so simple as it might appear: Many experiments have
demonstrated that pavlovian learning is not a simple associative process
measuring the contingencies between the US and the CS [
        <xref ref-type="bibr" rid="ref24">24</xref>
        ]. For example, when an
association is learnt and extinguished, it is more rapid to be learnt again. Some
authors have proposed that during the process, latent causes are extracted [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ].
Others have stressed the role of the context [
        <xref ref-type="bibr" rid="ref39">39</xref>
        ]. As a results, after a century
and thousands of publications, many data have been accumulated, many models
have been proposed but pavlovian learning remains not fully understood.
      </p>
      <p>
        It is also important to underline that pavlovian conditioning has a prominent
role in cognitive modeling and in the domain of Arti cial General Intelligence,
because it gives the capacity to link values to objects and events of the
environment and this capacity is important in a cognitive-oriented description of
intelligent behavior. For example, in a general description of cognition [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], one central
module is concerned with declarative memory, providing facts and knowledge
about the world, in which goals can be de ned and sensorimotor associations
can be elaborated, thus forming other central modules. All these modules
heavily rely on the elementary capacity to predict aversive and appetitive values for
CS from the innate detection of US. Some cues to elaborate the CS might come
from declarative memory and the US correspond to the main behavioral goals
for most animals. This capacity is then extended to conditioned reinforcers [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ],
corresponding to intermediate goals like, in humans, tools or money, and at the
basis of categorization [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] and of judgments [
        <xref ref-type="bibr" rid="ref22">22</xref>
        ] used in strategies and more
generally in decision making and planning.
      </p>
      <p>In this paper, we propose to study pavlovian conditioning from two di erent
points of view. On the one hand, from a top-down machine learning point of
view, we will wonder how it is possible to associate the CS and the US. In other
words, the question is to associate the perception of external stimuli like the
CS, also called exteroception, to the internal perception of pain or pleasure, also
called interoception, as it is the case with the US. We summarize some more or
less classical strategies at the algorithmic level in the next section.</p>
      <p>On the other hand in computational neuroscience, we have modeled some
neuronal structures and circuits that are reported to play a fundamental role
in pavlovian conditioning. This particularly concerns the amygdala, a primitive
neuronal structure in the medial temporal lobe and its main a erences. We
summarize in the following the main results of our study in a bottom-up way
and explain how we integrate these pieces of information towards a more global
understanding of pavlovian conditioning, at the algorithmic level and how this
can be more generally pro table for cognitive modeling.
3</p>
    </sec>
    <sec id="sec-3">
      <title>Modeling pavlovian conditioning in Machine Learning</title>
      <p>
        In the domain of machine learning, pavlovian conditioning has often been
associated to supervised learning. In its simplest aspect, it corresponds, when a
stimulus is perceived, to predict the occurrence of the US and the learning can
consequently be supervised by the fact that, some milliseconds later, the US is
perceived or not. Even if some binary models have been proposed, predicting
the magnitude of the US (for example the size of the reward or the intensity of a
painful electric shock) is the main purpose of the learning rule. In that aspect, it
is interesting to note that one of the main learning rules proposed for pavlovian
conditioning by Rescorla and Wagner [
        <xref ref-type="bibr" rid="ref36">36</xref>
        ] is similar to the main learning rule
setting the bases of supervised learning in neural networks, proposed by
Rosenblatt for the Perceptron [
        <xref ref-type="bibr" rid="ref38">38</xref>
        ]. They both set the focus on a multiplicative term,
called error for the perceptron and surprise for pavlovian learning,
corresponding to the di erence between the predicted and the computed US. This term
(null if there is no error, thus ensuring the stability of the rule) also explains
that there is no learning when the US is already predicted, including by other
CS. Importantly, this has been observed in experimental situations in a famous
paradigm called blocking, increasing the interest for this rule.
      </p>
      <p>
        Nevertheless, for the same kind of structural reasons as the perceptron, the
Rescorla-Wagner learning rule has some limitations, particularly related to the
complexity of the decision that can be taken in the input space. Concerning the
perceptron, it has been shown that its learning rule guarantees the convergence
towards a unique minimum value of error, but this results in a good quality of
the learnt solution only in case of linear separability in the input space [
        <xref ref-type="bibr" rid="ref30">30</xref>
        ].
Concerning pavlovian conditioning, it has been observed that some mammals can
answer to some combination of two variables (x and y) called positive (x AND
y) et negative (x XOR y) patterning, di erently than the corresponding
combined e ect of each variable, raising the idea to create con gural representation
(representing the co-occurrence of x and y), able to provide a di erent e ect.
This principle of con gural learning proposed in [
        <xref ref-type="bibr" rid="ref39">39</xref>
        ] resulted in that paper in
the same idea of adding hidden units as in the multilayer perceptron, to propose
a model of pavlovian conditioning able to perform such complex decision.
      </p>
      <p>
        The idea of performing error-driven learning in multilayered networks has
been made possible in Machine Learning by the back-propagation learning
algorithm which allows to make emerge internal representations in the hidden layers.
Even if some biologically plausible learning rules have also been proposed [
        <xref ref-type="bibr" rid="ref31">31</xref>
        ],
the main visible e ect of this algorithm is rather today in the huge developments
about deep learning [
        <xref ref-type="bibr" rid="ref25">25</xref>
        ], which make the learning of any complex association
a ordable by a deep network. Nevertheless, as it can be seen in the di cult
steps of parameter tuning in these networks, there is no proof of convergence in
this case and, instead of looking for the best solution, mastering this
technology rather corresponds to obtaining a satisfactory solution in a satisfactory time
length.
      </p>
      <p>
        It is also often forgotten that the only strong theoretical result about
layered network is the fact that removing one layer can require to exponentionally
increase the size of the previous layer [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ]. To tell it di erently, it is possible
to approximate any separation with a unique hidden layer and a random xed
connection from the input to the hidden layer (but possibly with a very large
hidden layer). This has for example been experimentally studied in [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], showing
that the function learnt by complex deep networks can be e ciently
approximated by a simple network with a unique (and large) hidden layer. It is only
supposed that the projection to the hidden layer is likely to make emerge the
useful internal representations, to be linearly separated for the targeted
decision. A strategy when the corresponding knowledge is available, is to directly
send the pertinent information to the hidden layer, else the solution is to use
random projections and the possibility to create pertinent information becomes
very likely when the layer is very large. In summary, the main advantage of deep
networks is to have what is called a compact representation (because requiring
less neurons), when the representation is elaborated hierarchically, as compared
to a at representation, but this is at the price of a less stable representation.
      </p>
      <p>
        In a completly di erent view, it is often neglected in Machine Learning that
associative memories can provide a kind of supervised learning. Associative
memories are these kinds of recurrent networks, like the Hop eld network [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ], that
are able to learn \by heart" some prototypes as stable states and, in a recall
phase, to consider a new input and stabilize on the closest state stored
beforehand. This process is called auto-association, since the network is shown to be
able to associate a pattern (the new input) with itself (seen as the closest
prototype learnt by heart before). This is generally considered useful in Machine
Learning when the new input has been corrupted by noise and when the recall
is a way to denoise this input.
      </p>
      <p>This process can easily be extended to hetero-association. In this case, just
consider a series of i vectors Vi as the concatenation of subvectors Xi and Yi and
consider that these vectors are learned as prototypes in an associative memory.
The principle of this learning process is that a noisy Vj will be reconstructed
to the closest Vi learned before. If, in the vector Vi, the noise corresponds to
set Yi to zero and let Xi unchanged, then the recall will reconstruct the missing
part Yi and will consequently act in hetero-association, which can be seen as a
kind of supervised learning, even if in this case there is no generalization and
the network converges towards a pattern already stored before.</p>
      <p>
        Several other characteristics of this learning process must be underlined. On
the one hand, it is interesting because it is very rapid and presenting only one
time a prototype is su cient to memorize it (as opposed to layered perceptrons
learning slowly). On the other hand, it can be considered weak because it is
subject to interference (close prototypes can create spurious states combining
these prototypes) and saturation (learning too many prototypes can lead to a
catastrophic forgetting) [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ].
4
      </p>
    </sec>
    <sec id="sec-4">
      <title>Modeling pavlovian conditioning in Computational</title>
    </sec>
    <sec id="sec-5">
      <title>Neuroscience</title>
      <p>
        Concerning the implementation of pavlovian conditioning in the brain, the
amygdala is a neural structure generally considered to play a central role in this
process [
        <xref ref-type="bibr" rid="ref26">26</xref>
        ]. The structure is composed of a set of nuclei, including the basolateral
nucleus for sensory inputs and the central nucleus expressing pavlovian motor
responses. Observing direct plastic connections between these sensory and motor
nuclei [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] is a good argument to propose a shallow layered architecture to model
this network. Particularly, the lateral part of the basolateral nucleus has been
reported to represent elementary CS, selected in its a erences from the thalamus
and primary sensory cortex [
        <xref ref-type="bibr" rid="ref26">26</xref>
        ], whereas the basal part receives a erences from
more integrated regions like associative sensory cortex and hippocampus [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ].
      </p>
      <p>
        A functional analogy with a model like the simple perceptron can be also
evoked for the amgdala, since signals corresponding to error of prediction have
been observed in the network [
        <xref ref-type="bibr" rid="ref27">27</xref>
        ]. Such error signals are known to be carried
by neuromodulators, which can act at the learning and functioning levels in the
amygdala but also in other cerebral structures. This is the case for acetylcholine,
considered to tag errors due to stochasticity in the environment [
        <xref ref-type="bibr" rid="ref42">42</xref>
        ] and acting
to balance the activity between the lateral and basal sensory nuclei [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], but also
to trigger memorization in the hippocampus [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] and shift selective attention in
the higher regions of the sensory cortex [
        <xref ref-type="bibr" rid="ref32">32</xref>
        ]. To better understand the putative
role of neuromodulation on these structures, let us rst evoke the participation
of the structures to pavlovian conditioning.
      </p>
      <p>
        The basal nucleus of the amygdala receives inputs from the associative
sensory cortex [
        <xref ref-type="bibr" rid="ref26">26</xref>
        ]. In the visual case for example and in a large region going from
the primary visual cortex to the inferotemporal cortex, this path has been
described as a layered architecture, extracting more and more abstract cues to
create categories for visual recognition [
        <xref ref-type="bibr" rid="ref41">41</xref>
        ]. Based on biological observations, it
has also been proposed that the release of acetylcholine plays a prominent role
for transmitting an error signal, as an attentional signal fed back throughout the
architecture and some models have been proposed accordingly, as a biological
implementation of a kind of error backpropagation [
        <xref ref-type="bibr" rid="ref37">37</xref>
        ].
      </p>
      <p>
        The hippocampus is also known to be a major input of the basal nucleus of
the amygdala [
        <xref ref-type="bibr" rid="ref26">26</xref>
        ], generally considered to provide information about the context
of the task, that can play an important role in discrimination. This is consistent
with the fact that the hippocampus is known to perform arbitrary binding of its
inputs, collected in the entorhinal cortex from main sensory regions of the cortex,
and to learn these associations through a recurrent network to form a so-called
episodic memory [
        <xref ref-type="bibr" rid="ref40">40</xref>
        ]. Recently, we have also proposed that the hippocampus
might in fact associate interoception and exteroception in its inputs, making it
a perfect candidate for learning the vector CS-US in heteroassociation [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ] and
recalling US when the CS is presented alone. In a more computational study
[
        <xref ref-type="bibr" rid="ref20">20</xref>
        ], we have also shown that this organization is in fact bene cial to resist to
interference.
      </p>
      <p>
        The complementary roles in learning of the cortex and the hippocampus have
already been theoretized [
        <xref ref-type="bibr" rid="ref29">29</xref>
        ]. Whereas the cortex is able to learn slowly cues
important for discrimination and generalization, the hippocampus has a di erent
learning mode, allowing to learn quickly, by heart, some speci c con guration
to be remembered. In the case of the amygdala, this should correspond either to
extract cues in the cortex to learn some features of the CS or, when this results
in an error of prediction, to learn by heart the global sensory con guration
in the hippocampus and to store it as a special case to be remembered. It is
consequently interesting to remind now that acetylcholine release triggers storing
in the hippocampus.
      </p>
      <p>
        With the aim of integrating all these elements in a unique picture, we begin to
report preliminary work we did to model pavlovian conditioning in the amygdala
[
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. Our main goal was to evaluate which of the many standard rules proposed for
pavlovian conditioning [
        <xref ref-type="bibr" rid="ref24">24</xref>
        ] was the most faithful, as compared to biological data.
What we proposed, in short, is that several populations with di erent behavior,
recently reported in the basolateral nucleus [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ], receive di erent kinds of cues,
each kind reported to be important in some classic rule of pavlovian learning.
In addition, we proposed that, consistent with the fact that several populations
with di erent excitatory and inhibitory roles have also been recently described
in the central nucleus of the amygdala [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ], the behavior actually expressed by
the amygdala is in fact the result of the competition between di erent lines of
elementary sensorimotor association in the amygdala. In conclusion, contrarily
to rules from experimental psychology, becoming more and more complex as they
want to model pavlovian conditioning, taking more and more facts into account,
we propose in [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] that the same complexity of behavior can be expressed by the
simple competition between di erent lines of association, underlying the role of
neuromodulation to favor some lines over others, depending on the context or
the recent history of behavior [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ].
      </p>
      <p>
        One step further, what we propose in this paper is not to forget that cues
arriving in the sensory input of the amygdala are themselves learnt in other
regions. This is consequently a wider view, proposing to model pavlovian
conditioning in a network of neuronal structures and to consider that critical events
in the network can trigger important learning or functioning steps in di erent
locations throughout the network. In this view, as it was already proposed in [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ],
the amygdala seen as a simple perceptron decides which US to predict from a
variety of inputs received in the basolateral nucleus. These inputs can correspond
to simple elementary CS coming from the thalamus or to more elaborated hints
or con gurations learnt in other cortical areas or in the hippocampus. Possibly,
a release of acetylcholine will give a stronger weight to complex inputs, meaning
that the environment is too noisy to rely on simple cues.
      </p>
      <p>
        This mode of learning is incremental in the sense that, at the beginning,
it can be supposed that hardly no complex cues have been extracted in the
hippocampus or in the highest levels of the cortex. Errors of prediction will
result in neuromodulation triggering learning in the hippocampus of the speci c
cases poorly considered before. Later on, if the same case occurs again, it will
trigger recall in the hippocampus that will reconstruct the corresponding US and
send it back to the amygdala. Concerning learning in the layered architecture
of the cortex and in addition to bio-plausible mechanisms proposed in [
        <xref ref-type="bibr" rid="ref37">37</xref>
        ] and
mentioned above, the phenomenon of consolidation must be also evoked. In this
process, also reported in neuroscience and described in [
        <xref ref-type="bibr" rid="ref29">29</xref>
        ], the hippocampus can
act as a supervisor to the cortex and send it some examples previously learned
by heart in episodic memory, in order to transfer it to semantic memory in the
cortex. As far as we know, no stable learning rules have been proposed so far
to implement this important learning process, but this will be very important
to consider in the future because it might allow to remove examples in the
hippocampus and prevent it from saturation.
5
      </p>
    </sec>
    <sec id="sec-6">
      <title>Discussion</title>
      <p>
        The capacity for an intelligent agent to associate a sensory state (taken in the
widest sense, including more or less abstract multimodal and contextual
information) to a value (particularly representing its possible contribution to achieving
the current goal of the agent) is very desirable to de ne in a cognitive
architecture. It can be paralleled, in natural cognition, with pavlovian conditioning
linking a state (more or less abstract stimulus, context) to the unconditioned
stimulus it predicts, annoucing a pain or a pleasure. In the living, this learning
has been shown to contribute both to declarative and implicit memories [
        <xref ref-type="bibr" rid="ref40">40</xref>
        ].
      </p>
      <p>In this paper, we have summarized pavlovian conditioning as the capacity to
associate a vector representing the sensory state with a vector representing the
anticipated internal state and we have evoked three classical strategies in
machine learning to learn this kind of association, each with its respective strengths
and weaknesses. Shallow layered networks ensure a quick processing and optimal
learning but only for simple direct associations. In contrast, deep layered
networks can learn complex associations, which is particularly useful to elaborate
internal representations, concepts and abstractions, but at the price of a very
long learning that can fall in local minima and must choose pertinent con
guations. Of completly di erent nature is the elaboration of associative memories in
recurrent networks. Prototypes can be learned by heart, possibly in one shot but
are sensitive to interference and globally an associative memory can be subject
to saturation.</p>
      <p>From this global analysis, one strategy is to look for the best compromise;
another one is to try to bene t from the strengths of all. In the previous section,
we have argued to show that the brain exploits this latter strategy. Whereas
pavlovian conditioning can be summarized as an implicit sensorimotor learning
taking place in the amygdala and associating simple cues to pavlovian responses
related to the predicted US, this cannot explain all the observed pavlovian
associations and other inputs to the amygdala carrying other kind of information
must be also considered. This is particularly the case for inputs coming from the
hippocampus and from high cortical areas and learned and processed in these
structures with di erent criteria. What we have also made clear here is that
these structures are also dependent on the processing made in the amygdala and
particularly on neuromodulation released from its errors of prediction.</p>
      <p>
        This system of combined learning occuring in di erent sites in parallel but
also for common reasons can be interpreted from a phylogenetic point of view.
In primitive animals, only simple associations can be learnt [
        <xref ref-type="bibr" rid="ref23">23</xref>
        ]. With evolution,
experience from recent episodes and association of multimodal information can
be integrated to the pavlovian association.
      </p>
      <p>
        It is of course also very tempting to extrapolate this interpretation and the
related observations to the domains of cognitive modeling and arti cial
intelligence. Proposing models based on the cooperation of memories is not new in
these domains [
        <xref ref-type="bibr" rid="ref1 ref29">1, 29</xref>
        ] and relates to the design of ensemble or multi-method
approaches. This must be more and more seriously considered, since new
technologies in experimental neuroscience allow to observe that most evolved cognitive
functions are in fact elaborated from an interplay between neuronal structures
[
        <xref ref-type="bibr" rid="ref34">34</xref>
        ]. In the present case, this gives an elegant way to introduce speci c cases
to an associative process with no loss of generalisation. Using hidden layers to
elaborate con gural representations is not mandatory, as it has been reminded
that a simple layer could perform the same task (at the price of a larger hidden
layer), but it is interesting for using internal representations for other purposes.
This hierarchy of hidden layers provides a level of abstraction (not obtained
by a shallow network) that could be exploited for other cognitive function and
particularly for declarative memory and for executive functions.
      </p>
      <p>
        We have also mentioned up to which point neuromodulation plays an
important role in associating the processing and learning phases in the neuronal
structures, and can be seen as a rare ground of coordination between the otherwise
parallel processes. In the future, we plan to better understand their functions
[
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] and we have already proposed some mechanisms to decipher some
temporal aspects of pavlovian conditioning [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ] and to go beyond, towards operant
conditioning and associated high level cognitive functions [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ].
      </p>
      <p>
        To conclude on a more methodological point of view and coming back to
Marr's levels of description [
        <xref ref-type="bibr" rid="ref28">28</xref>
        ], we have presented some top-down approaches
in Machine Learning and bottom-up approaches in computational neuroscience,
forming the state of the art on the current understanding of pavlovian
conditioning. We demonstrate here that synthetizing these analyses by the algorithmic
level is a unique way to propose new interpretations and to revisit cognitive
modeling on this topic.
      </p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Alexandre</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          :
          <article-title>Biological Inspiration for Multiple Memories Implementation and Cooperation</article-title>
          . In: In V. Kvasnicka
          <string-name>
            <given-names>P.</given-names>
            <surname>Sincak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Vascak</surname>
          </string-name>
          and R. Mesiar, editors,
          <source>International Conference on Computational Intelligence</source>
          (
          <year>2000</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Alexandre</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Carrere</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Modeling Neuromodulation as a Framework to Integrate Uncertainty in General Cognitive Architectures</article-title>
          .
          <source>In: The Ninth Conference on Arti cial General Intelligence</source>
          . New-York, United States (
          <year>Jul 2016</year>
          ), https://hal.inria.fr/hal-01342902
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Alexandre</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Carrere</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kassab</surname>
          </string-name>
          , R.: Feature,
          <article-title>Con guration, History : a bioinspired framework for information representation in neural networks</article-title>
          .
          <source>In: International Conference on Neural Computation Theory and Applications</source>
          . Rome, Italy (Oct
          <year>2014</year>
          ), https://hal.inria.fr/hal-01095036
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Anderson</surname>
            ,
            <given-names>J.R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bothell</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Byrne</surname>
            ,
            <given-names>M.D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Douglass</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lebiere</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Qin</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          :
          <article-title>An integrated theory of the mind</article-title>
          .
          <source>Psychol Rev</source>
          <volume>111</volume>
          (
          <issue>4</issue>
          ),
          <volume>1036</volume>
          {1060 (Oct
          <year>2004</year>
          ), http: //dx.doi.org/10.1037/
          <fpage>0033</fpage>
          -
          <lpage>295x</lpage>
          .
          <year>111</year>
          .4.
          <fpage>1036</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Ashby</surname>
            ,
            <given-names>F.G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Alfonso-Reese</surname>
            ,
            <given-names>L.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Turken</surname>
            ,
            <given-names>A.U.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Waldron</surname>
            ,
            <given-names>E.M.:</given-names>
          </string-name>
          <article-title>A neuropsychological theory of multiple systems in category learning</article-title>
          .
          <source>Psychological review 105(3)</source>
          ,
          <volume>442</volume>
          {481 (Jul
          <year>1998</year>
          ), http://view.ncbi.nlm.nih.gov/pubmed/9697427
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Ba</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Caruana</surname>
            , R.: Do Deep Nets Really Need to be Deep? In: Ghahramani,
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Welling</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cortes</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lawrence</surname>
            ,
            <given-names>N.D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Weinberger</surname>
            ,
            <given-names>K.Q</given-names>
          </string-name>
          . (eds.)
          <source>Proceedings of the Neural Information Processing Systems Conference, NIPS 2014</source>
          , pp.
          <volume>2654</volume>
          {
          <fpage>2662</fpage>
          . Curran Associates, Inc. (
          <year>2014</year>
          ), http://papers.nips.cc/paper/ 5484-do
          <article-title>-deep-nets-really-need-to-be-deep</article-title>
          .pdf
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Belova</surname>
            ,
            <given-names>M.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Paton</surname>
            ,
            <given-names>J.J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Morrison</surname>
            ,
            <given-names>S.E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Salzman</surname>
          </string-name>
          , C.D.:
          <article-title>Expectation modulates neural responses to pleasant and aversive stimuli in primate amygdala</article-title>
          .
          <source>Neuron</source>
          <volume>55</volume>
          (
          <issue>6</issue>
          ),
          <volume>970</volume>
          {
          <fpage>984</fpage>
          (
          <year>2007</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Calandreau</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tri</surname>
            <given-names>lie</given-names>
          </string-name>
          , P.,
          <string-name>
            <surname>Mons</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Costes</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Marien</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Marighetto</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Micheau</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ja</surname>
            <given-names>ard</given-names>
          </string-name>
          , R.,
          <string-name>
            <surname>Desmedt</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Extracellular hippocampal acetylcholine level controls amygdala function and promotes adaptive conditioned emotional response</article-title>
          .
          <source>The Journal of neuroscience : the o cial journal of the Society for Neuroscience</source>
          <volume>26</volume>
          (
          <issue>52</issue>
          ),
          <volume>13556</volume>
          {13566 (Dec
          <year>2006</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Carrere</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Alexandre</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          :
          <article-title>A pavlovian model of the amygdala and its in uence within the medial temporal lobe</article-title>
          .
          <source>Frontiers in Systems Neuroscience</source>
          <volume>9</volume>
          (
          <issue>41</issue>
          ) (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Carrere</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Alexandre</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          :
          <article-title>Modeling the sensory roles of noradrenaline in action selection</article-title>
          .
          <source>In: The Sixth Joint IEEE International Conference Developmental Learning and Epigenetic Robotics (IEEE ICDL-EPIROB)</source>
          . Cergy-Pontoise / Paris, France (
          <year>Sep 2016</year>
          ), https://hal.inria.fr/hal-01401882
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Ciocchi</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Herry</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Grenier</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wol</surname>
            ,
            <given-names>S.B.E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Letzkus</surname>
            ,
            <given-names>J.J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vlachos</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ehrlich</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sprengel</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Deisseroth</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stadler</surname>
            ,
            <given-names>M.B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Muller</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Luthi</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Encoding of conditioned fear in central amygdala inhibitory circuits</article-title>
          .
          <source>Nature</source>
          <volume>468</volume>
          (
          <issue>7321</issue>
          ),
          <volume>277</volume>
          {282 (Nov
          <year>2010</year>
          ), http://dx.doi.org/10.1038/nature09559
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>French</surname>
            ,
            <given-names>R.M.</given-names>
          </string-name>
          :
          <article-title>Semi-distributed representations and catastrophic forgetting in connectionist networks</article-title>
          .
          <source>Connection Science</source>
          <volume>4</volume>
          (
          <issue>3-4</issue>
          ),
          <volume>365</volume>
          {
          <fpage>377</fpage>
          (
          <year>1992</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Gershman</surname>
            ,
            <given-names>S.J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Blei</surname>
            ,
            <given-names>D.M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Niv</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          :
          <article-title>Context, learning, and extinction</article-title>
          .
          <source>Psychological review 117(1)</source>
          ,
          <volume>197</volume>
          {209 (Jan
          <year>2010</year>
          ), http://view.ncbi.nlm.nih.gov/pubmed/ 20063968
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Goosens</surname>
            ,
            <given-names>K.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Maren</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Contextual and Auditory Fear Conditioning are Mediated by the Lateral, Basal, and Central Amygdaloid Nuclei in Rats</article-title>
          .
          <source>Learning &amp; Memory</source>
          <volume>8</volume>
          (
          <issue>3</issue>
          ),
          <volume>148</volume>
          {155 (May
          <year>2001</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Hasselmo</surname>
            ,
            <given-names>M.E.</given-names>
          </string-name>
          :
          <article-title>The role of acetylcholine in learning and memory</article-title>
          .
          <source>Curr Opin Neurobiol</source>
          <volume>16</volume>
          (
          <issue>6</issue>
          ),
          <volume>710</volume>
          {715 (Dec
          <year>2006</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>Hastad</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Goldmann</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>On the power of small-depth threshold circuits</article-title>
          .
          <source>Computational Complexity</source>
          <volume>1</volume>
          ,
          <issue>113</issue>
          {
          <fpage>129</fpage>
          (
          <year>1991</year>
          ), https://doi.org/10.1007/BF01272517
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <surname>Herry</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ciocchi</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Senn</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Demmou</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Muller</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Luthi</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Switching on and o fear by distinct neuronal circuits</article-title>
          .
          <source>Nature</source>
          <volume>454</volume>
          (
          <issue>7204</issue>
          ),
          <volume>600</volume>
          {606 (Jul
          <year>2008</year>
          ), http://dx.doi.org/10.1038/nature07166
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          18.
          <string-name>
            <surname>Hop</surname>
            <given-names>eld</given-names>
          </string-name>
          , J.J.:
          <article-title>Neural networks and physical systems with emergent collective computational abilities</article-title>
          .
          <source>In: Proceedings of the National Academy of Sciences, USA</source>
          . pp.
          <volume>2554</volume>
          {
          <issue>2558</issue>
          (
          <year>1982</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          19.
          <string-name>
            <surname>Kassab</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Alexandre</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          :
          <article-title>Integration of exteroceptive and interoceptive information within the hippocampus: a computational study</article-title>
          .
          <source>Frontiers in Systems Neuroscience</source>
          <volume>9</volume>
          (
          <issue>87</issue>
          ) (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          20.
          <string-name>
            <surname>Kassab</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Alexandre</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          :
          <article-title>A Modular Network Architecture Resolving Memory Interference through Inhibition</article-title>
          . In: et al.,
          <string-name>
            <surname>J.M.</surname>
          </string-name>
          (ed.)
          <source>Computational Intelligence, Studies in Computational Intelligence</source>
          , vol.
          <volume>669</volume>
          , pp.
          <volume>407</volume>
          {
          <fpage>422</fpage>
          . Springer (
          <year>2016</year>
          ), https://hal.inria.fr/hal-01251022
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          21.
          <string-name>
            <surname>Kaushik</surname>
            ,
            <given-names>P.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Carrere</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Alexandre</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Raju</surname>
            ,
            <given-names>S.B.</given-names>
          </string-name>
          :
          <article-title>A biologically inspired neuronal model of reward prediction error computation</article-title>
          .
          <source>In: 2017 International Joint Conference on Neural Networks, IJCNN</source>
          <year>2017</year>
          ,
          <article-title>Anchorage</article-title>
          ,
          <string-name>
            <surname>AK</surname>
          </string-name>
          , USA, May
          <volume>14</volume>
          -19,
          <year>2017</year>
          . pp.
          <volume>3577</volume>
          {
          <issue>3584</issue>
          (
          <year>2017</year>
          ), https://doi.org/10.1109/IJCNN.
          <year>2017</year>
          .7966306
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          22. L.,
          <string-name>
            <surname>A.G.</surname>
          </string-name>
          :
          <article-title>Human contingency judgments : Rule based or associative</article-title>
          .
          <source>Psychological Bulletin</source>
          <volume>114</volume>
          (
          <issue>3</issue>
          ),
          <volume>435</volume>
          {
          <fpage>448</fpage>
          (
          <year>1993</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          23.
          <string-name>
            <surname>Laberge</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Muhlenbrock-Lenter</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Grunwald</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Roth</surname>
          </string-name>
          , G.:
          <article-title>Evolution of the Amygdala: New Insights from Studies in Amphibians</article-title>
          .
          <source>Brain Behav Evol</source>
          <volume>67</volume>
          (
          <issue>4</issue>
          ),
          <volume>177</volume>
          {
          <fpage>187</fpage>
          (
          <year>2006</year>
          ), http://dx.doi.org/10.1159/000091119
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          24.
          <string-name>
            <surname>Le Pelley</surname>
            ,
            <given-names>M.E.</given-names>
          </string-name>
          :
          <article-title>The role of associative history in models of associative learning: a selective review and a hybrid model</article-title>
          .
          <source>The Quarterly Journal of Experimental Psychology</source>
          <volume>57</volume>
          (
          <issue>3</issue>
          ),
          <volume>193</volume>
          {243 (Jul
          <year>2004</year>
          ), http://dx.doi.org/10.1080/ 02724990344000141
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          25.
          <string-name>
            <surname>LeCun</surname>
          </string-name>
          , Y.,
          <string-name>
            <surname>Bengio</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hinton</surname>
          </string-name>
          , G.:
          <article-title>Deep learning</article-title>
          .
          <source>Nature</source>
          <volume>521</volume>
          (
          <issue>7553</issue>
          ),
          <volume>436</volume>
          {444 (May
          <year>2015</year>
          ), http://dx.doi.org/10.1038/nature14539
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          26.
          <string-name>
            <surname>LeDoux</surname>
          </string-name>
          , J.:
          <article-title>The amygdala</article-title>
          .
          <source>Current Biology</source>
          <volume>17</volume>
          (
          <issue>20</issue>
          ),
          <source>R868{R874 (Oct</source>
          <year>2007</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          27.
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>S.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>McNally</surname>
            ,
            <given-names>G.P.:</given-names>
          </string-name>
          <article-title>The conditions that promote fear learning: Prediction error and Pavlovian fear conditioning</article-title>
          .
          <source>Neurobiology of Learning and Memory</source>
          <volume>108</volume>
          (
          <issue>0</issue>
          ),
          <volume>14</volume>
          {
          <fpage>21</fpage>
          (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          28.
          <string-name>
            <surname>Marr</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>Vision: A Computational Investigation into the Human Representation and Processing of Visual Information. Henry Holt and Co</article-title>
          ., Inc., New York, NY, USA (
          <year>1982</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          29.
          <string-name>
            <surname>McClelland</surname>
            ,
            <given-names>J.L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>McNaughton</surname>
            ,
            <given-names>B.L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>O'Reilly</surname>
            ,
            <given-names>R.C.</given-names>
          </string-name>
          :
          <article-title>Why there are complementary learning systems in the hippocampus and neocortex: insights from the successes and failures of connectionist models of learning and memory</article-title>
          .
          <source>Psychological review 102(3) (Jul</source>
          <year>1995</year>
          ), http://view.ncbi.nlm.nih.gov/pubmed/7624455
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          30.
          <string-name>
            <surname>Minsky</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Papert</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Perceptrons: An Introduction to Computational Geometry</article-title>
          . MIT Press, Cambridge (
          <year>1969</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          31.
          <string-name>
            <given-names>O</given-names>
            <surname>'Reilly</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.C.</given-names>
            ,
            <surname>Munakata</surname>
          </string-name>
          ,
          <string-name>
            <surname>Y.</surname>
          </string-name>
          :
          <article-title>Computational Explorations in Cognitive Neuroscience: Understanding the Mind by Simulating the Brain</article-title>
          . The MIT Press,
          <volume>1</volume>
          <fpage>edn</fpage>
          .
          <source>(Sep</source>
          <year>2000</year>
          ), http://www.worldcat.org/isbn/0262650541
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          32.
          <string-name>
            <surname>Pauli</surname>
            ,
            <given-names>W.M.</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>O</given-names>
            <surname>'Reilly</surname>
          </string-name>
          ,
          <string-name>
            <surname>R.C.</surname>
          </string-name>
          :
          <article-title>Attentional control of associative learning{a possible role of the central cholinergic system</article-title>
          .
          <source>Brain Research</source>
          <volume>1202</volume>
          ,
          <volume>43</volume>
          {53 (Apr
          <year>2008</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          33.
          <string-name>
            <surname>Pavlov</surname>
            ,
            <given-names>I.P.</given-names>
          </string-name>
          :
          <article-title>Conditioned Re exes (V.Anrep, trans</article-title>
          .). London: Oxford University Press" (
          <year>1927</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref34">
        <mixed-citation>
          34.
          <string-name>
            <surname>Paz</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bauer</surname>
            ,
            <given-names>E.P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pare</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>Measuring correlations and interactions among four simultaneously recorded brain regions during learning</article-title>
          .
          <source>Journal of neurophysiology 101(5)</source>
          ,
          <volume>2507</volume>
          {
          <fpage>2515</fpage>
          (
          <year>2009</year>
          ), http://dx.doi.org/10.1152/jn.91259.2008
        </mixed-citation>
      </ref>
      <ref id="ref35">
        <mixed-citation>
          35.
          <string-name>
            <surname>Pylyshyn</surname>
            ,
            <given-names>Z.W.</given-names>
          </string-name>
          :
          <article-title>Computation and Cognition</article-title>
          . MIT Press (
          <year>1984</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref36">
        <mixed-citation>
          36.
          <string-name>
            <surname>Rescorla</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wagner</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>A theory of pavlovian conditioning: Variations in the e ectiveness of reinforcement and nonreinforcement</article-title>
          .
          <source>In: Classical Conditioning II: Current Research and Theory</source>
          , pp.
          <volume>64</volume>
          {
          <fpage>99</fpage>
          .
          <string-name>
            <surname>Appleton Century Crofts</surname>
          </string-name>
          (
          <year>1972</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref37">
        <mixed-citation>
          37.
          <string-name>
            <surname>Roelfsema</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          R., van
          <string-name>
            <surname>Ooyen</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Attention-gated reinforcement learning of internal representations for classi cation</article-title>
          .
          <source>Neural computation</source>
          <volume>17</volume>
          (
          <issue>10</issue>
          ),
          <volume>2176</volume>
          {
          <fpage>2214</fpage>
          (
          <year>2005</year>
          ), http://dx.doi.org/10.1162/0899766054615699
        </mixed-citation>
      </ref>
      <ref id="ref38">
        <mixed-citation>
          38.
          <string-name>
            <surname>Rosenblatt</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          :
          <article-title>The perceptron: a probabilistic model for information storage and organization in the brain</article-title>
          . In: Anderson,
          <string-name>
            <given-names>J.A.</given-names>
            ,
            <surname>Rosenfeld</surname>
          </string-name>
          , E. (eds.) Neurocomputing: Fundations of Research (
          <year>1989</year>
          ), pp.
          <volume>89</volume>
          {
          <fpage>92</fpage>
          . The MIT Press (
          <year>1958</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref39">
        <mixed-citation>
          39.
          <string-name>
            <surname>Schmajuk</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>DiCarlo</surname>
          </string-name>
          , J.:
          <article-title>Stimulus con guration, classical conditioning and the hippocampus</article-title>
          .
          <source>Psychological Review</source>
          <volume>99</volume>
          ,
          <issue>268</issue>
          {
          <fpage>305</fpage>
          (
          <year>1992</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref40">
        <mixed-citation>
          40.
          <string-name>
            <surname>Squire</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          :
          <article-title>Declarative and nondeclarative memory: multiple brain systems supporting learning and memory</article-title>
          .
          <source>Journal of cognitive neuroscience 4</source>
          (
          <issue>3</issue>
          ),
          <volume>232</volume>
          {
          <fpage>243</fpage>
          (
          <year>1992</year>
          ), http://view.ncbi.nlm.nih.gov/pubmed/23964880
        </mixed-citation>
      </ref>
      <ref id="ref41">
        <mixed-citation>
          41.
          <string-name>
            <surname>Thorpe</surname>
            ,
            <given-names>S.J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fabre-Thorpe</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Seeking Categories in the Brain</article-title>
          .
          <source>Science</source>
          <volume>291</volume>
          (
          <issue>5502</issue>
          ),
          <volume>260</volume>
          {263 (Jan
          <year>2001</year>
          ), http://dx.doi.org/10.1126/science.1058249
        </mixed-citation>
      </ref>
      <ref id="ref42">
        <mixed-citation>
          42.
          <string-name>
            <surname>Yu</surname>
            ,
            <given-names>A.J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dayan</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          : Uncertainty, Neuromodulation, and
          <string-name>
            <surname>Attention</surname>
          </string-name>
          .
          <source>Neuron</source>
          <volume>46</volume>
          (
          <year>2005</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>