=Paper=
{{Paper
|id=Vol-1895/AIC16_paper5
|storemode=property
|title=Modelling Auditory Spatial Attention
with Constraints
|pdfUrl=https://ceur-ws.org/Vol-1895/paper5.pdf
|volume=Vol-1895
|authors=Edward J. Golob,Kristen Brent Venable,Maxwell T. Anderson,Jesse A. Benzell,Jaelle Scheuerman
|dblpUrl=https://dblp.org/rec/conf/aic/GolobVABS16
}}
==Modelling Auditory Spatial Attention
with Constraints==
Modelling auditory spatial attention with constraints Edward J. Golob1 , K. Brent Venable2 , Maxwell T. Anderson1 , Jesse A. Benzell1 , and Jaelle Scheuerman2 1 Department of Psychology Tulane University egolob@tulane.edu,manders7@tulane.edu,jbenzell@tulane.edu 2 Department of Computer Science Tulane University kvenabl@tulane.edu,jscheuer@tulane.edu Abstract. It is well-established that spatial attention can be allocated as a gra- dient that diminishes from a central focus. In this paper we consider auditory attention and we develop a model for how it is distributed in space following basic ideas of top-down and bottom-up attentional control from verbal models [6, 12]. There are three main components of our model: a goal map, a saliency map, and a priority map. The goal map models the distribution of attention which is allocated by choice (top-down component). The saliency map, as the name suggests, models attention related to the saliency of auditory stimuli (bottom-up component) and the priority map synthesizes the other two maps in an overall distribution of the attentional bias. We model the three maps and their interac- tion using the well established AI framework of constraint satisfaction problems. We study several hypotheses on the maps and we contrast the results in terms of data obtained running different kinds of experiments. Our computational model, is to the best of our knowledge is the first which targets specifically the auditory system. Our constraint-based approach is very flexible in terms of embedding and testing different hypotheses on the components and constraint propagation techniques allow both to focus on single components as well as to consider the system dynamics as whole. The predictions arising from our model well fit the experimental data, are cognitive plausible and provide new interesting insights to the mechanism of attention control. 1 Introduction and motivation Audition is distinguished from the other senses by the ability to panoramically monitor the environment for things happening at a distance, behind obstructions, and out of sight. These considerations make the auditory system particularly useful for shifting attention to events that are important to survival and reproduction. For example, hearing the sound of a snapped twig from a predator hiding in the brush can quickly elicit a fight or flight response. The overall goal of this project is to better understand at the cognitive and neural levels of analysis how auditory attention is allocated over space. The focus is on the interplay between top-down and bottom-up spatial attention biases that govern shifting attention to distractors during performance of a simple spatial attention task. We take an interdisciplinary approach by using behavioral and neural stimulation methods to test and refine a computational model of auditory spatial attention. Previous work has established the idea of an auditory attention gradient [27, 33]. Both reports gave subjects a cue on where to attend that changed across trials, unlike natural situations where attention is engaged for longer times. Our task mimics every- day life and connects to the ecological significance of the auditory system in orienting attention to occasional unexpected, but potentially important, environmental events. We have strong preliminary data showing the novel result that when spatial attention shifts are examined over a wide range (180◦ ) reaction time slows following spatial shifts but then speeds-up at the distant location that was tested (180◦ from the currently attended location). Our recent work has defined behavioral and neural measures of auditory attention gradients [17]. The behavioral tasks in this proposal mirror everyday life by having sub- jects attend to one location for at least several minutes, as when conversing or listening to music. Auditory processing of distractors at different regions of space were probed and attentional gradients using EEG measures were defined relative to the current fo- cus of attention. Variables such as task demands, stimulus properties, normal aging, and neural stimulation of important cortical nodes of the hypothesized network were examined. We are now in a position to construct an explicit computational model of auditory spatial attention, which will be used to test existing hypotheses and make new predictions subject to experimental testing. Our aim is to develop a rigorous theory of auditory spatial attention that relates to current work on dorsal and ventral neural attention systems. We foresee that our work will help advance the understanding of basic issues in attention regardless of modality, such as top-down and bottom up interactions, capacity limitations, vigilance, and will inform debate on supramodal attention processing in the brain. There are also multiple applications, such as topics in human factors, improved audio communication systems, and brain-computer interface control using spatial attention. Our lab studies aspects of hearing that are particularly important to humans, such sound location, speech, and music, and how auditory processing is affected by atten- tion, short-term memory, and action planning and execution. The common denomina- tor is that these studies contribute to an emerging framework termed cognitive hearing science, which examines the role of the auditory system in higher-level cognition and action [4]. In addition to addressing basic science issues we also use the auditory sys- tem as a model to better understand the cognitive and neurobiological changes that accompany normal aging, Alzheimer’s disease, and speech fluency disorders. We combine these novel parametric behavioral measures to map-out auditory at- tention over space with a computational model to explain how specific top-down and bottom-up mechanisms jointly determine the shape of auditory spatial attention gradi- ents. Recent modeling work focuses on saliency, particularly when there is more than one sound happening at the same time [38, 13, 29]. Kayser and colleagues have a model of acoustic saliency based on non-spatial features (e.g. intensity, envelope), but did not examine spatial features. In contrast, we use soft constraint AI computational methods to model auditory spatial processing and overlapping top-down and bottom-up interac- tions. 2 Background In this sections we provide a brief background on psychology literature related to spatial attention and its computational models. We also give some fundamental information concerning constraints. 2.1 Spatial Attention Almost all attention models from the inception of Psychology as a formal science have distinguished attention that is directed by personal choice from attention that is directed to an event by virtue of it having a salient property, such as a loud sound [21, 30]. This dichotomy is intuitive and has many names in the literature (e.g. top-down/bottomup, endogenous/exogenous, controlled/automatic [11]. Here we use the terms top-down and bottom-up. Top-down control regulates information flow based on the current situation and goals in short-term memory by generating a task set to bias processing towards information useful for goal attainment. Bottom-up refers to attention capture that is not guided by the top-down task set. Although the top-down and bottom-up distinction is meaningful, as a practical matter they are highly interactive [15]. The difficulty of cleanly separating the two processes motivates us to use a computational model, which can examine topdown and bottom-up functions in isolation. Next, we briefly review work on auditory spatial attention at a cognitive level of analysis, and draw from the visual literature when needed to present major points relevant to auditory spatial atten- tion. Attention can be expressed as a spatial gradient relative to an attended location [26, 10]. Gradients are presumably a byproduct of limited perceptual input capacity, although limitations in behavioral output may also be relevant [2]. The spatial extent of attentional processing is variable [37], and can be modified by directly cuing different size areas [18], or manipulations of perceptual or memory loads [22]. Splitting atten- tion between locations and multi-object tracking are also possible [5, 9]. The ability to deliver attentional benefits rapidly diminishes over time, a phenomenon called the vigi- lance decrement [25]. This is important because in everyday life attention is commonly deployed over relatively long time periods (e.g. conversation, listening to music). Auditory spatial cuing decreases reaction times to subsequent targets at a cued lo- cation relative to uncued locations [32, 36, 39, 33]. Both Mondor and Zatorre (1995) and Rorden and Driver (2001) found that target reaction times increased monotonically with greater distance between the cued and target locations. Visual studies suggest that gradients may have a more complex shape, with reaction times increasing and then de- creasing away from the cued location [28, 8](Mexican-hat shape). This is, as we will see, similar to our preliminary findings in the auditory modality, but the auditory results have a much larger spatial range. 2.2 Constraints and Computational models of auditory attention Computational models of cognitive processes are beneficial because they require an explicit theory, can reveal hidden assumptions or logical inconsistencies, and simula- tions can establish proof-of-principle much faster than pilot experiments [20, 23]. Our model uses basic ideas of top-down and bottom-up attention control from prominent verbal models [6, 12]. The novelty of our approach is the application to auditory spatial attention, which is not dealt with in detail in the general models. Our model is distin- guished by focusing on auditory spatial attention and how it emerges from top-down and bottom-up interactions, which has general importance because a balance must be struck between top-down goal focus and bottom-up receptivity to unexpected events or ideas (stability-flexibility dilemma, [24]). Moreover, the models of attention mentioned above are designed as ad hoc mathematical descriptions of the considered phenomena, while we opt to cast our model into a more general artificial intelligence setting. Constraint programming [34] is a powerful paradigm for modeling and solving combinatorial search problems currently applied with success to many domains, such as scheduling, planning, vehicle routing, configuration, networks, and bioinformatics. The basic idea in constraint programming is that the user states the constraints and a general-purpose constraint solver is used to solve them. Constraint solvers take a real- world problem, represented in terms of decision variables and constraints, and find an assignment to all the variables that satisfies the constraints. Constraints concern subsets of variables and define which simultaneous assignments to those variables are allowed. For example, in scheduling activities in a company, the decision variables might be the starting times and the durations of the activities and the resources needed to perform them, and the constraints might be on the availability of the resources and on their use by a limited number of activities at a time. Solutions are found by searching the solution space either systematically, as with backtracking or branch and bound algorithms, or use forms of local search which may be incomplete, that is there is no guarantee they will return a solution. Systematic meth- ods often interleave search and inference, where inference consists of propagating the information contained in one constraint to other constraints via shared variables. The rich variety of finely-tuned algorithms available for constraint problems has made the effort of translating real world problems into this framework an efficient solv- ing approach. Constraints have been used before in the context of human cognition for example to model skilled behavior [1, 7] and learning [14]. Recently an implementation of the cognitive architecture ACT-R [3] based on constraint handling rules, which are a closely related to constraints, has been proposed in [16]. To the best of our knowledge, it is however the first time they are employed at this level of cognitive modeling and in the context of attention. Casting our model into a well-established AI framework will, on one side, facilitate future generalization to other aspects concerning attention as well as enable an easy embedding in cognitive architectures such as ACT-R, for example. 3 The computational model Figure 1 depicts the overall hypothesis on the interplay between top-down and bottom- up spatial attention processing. There are three main components: goal map, saliency map, and priority map. Each map is a 1-D vector of attentional bias in normalized units (0-1) across the semicircular horizontal frontal plane (from -90◦ on the far left to +90◦ on the far right, 2◦ increments, as shown in Figure 2). The given inputs to the model are (1) attended location, which is a goal map parameter and (2) sound location, which is input to the saliency map. The output is a priority map representation of attentional bias across the 180◦ semi-circle (in 2◦ increments). Areas of greater attentional bias are assumed to relate to measurable data by having faster reaction times, more sensitive sensory thresholds, and increased accuracy relative to locations with less bias. We emphasize that this is a model of information processing at the cognitive level. It is designed to help interpret behavioral results and inspire new experiments to test and refine the model. It is not intended to model how neural activity relates to attention. The gray boxes show inputs and outputs that interface with other cognitive functions. Fig. 1. Computational Model Schematic. Our computational model adopts a constraint-based approach to cast the interactions among the three maps into a constraint solving problem that can be efficiently solved with the rich algorithmic machinery which has been developed for constraints[34]. Con- straint models have three main components: variables, a domain of possible values for each variable, and the constraints. Constraints are defined by the relevant variables and by specifying the simultaneous variable assignments that satisfy the constraint. In our model, there is one input variable corresponding to attended location (A) with the do- main being locations (2◦ increments) in the semicircle {-90,-88,...,0,...,88,90}. In Figure 3 we depict (partially) the constraint graph of our model, where variables correspond to nodes and constraints to edges. We remark how this constraint-based representation is very flexible in terms of mod- eling different hypothesis on the attentional bias distributions and on the interaction of the maps. In our initial setting, for example, we don’t have any interaction between the goal and the saliency map. However, we could model it in a straight forward way sim- ply by defining new constraints that connect variables of the two maps. Moreover, this 0 We divide the semicircle into 90 slots of -‐90 +90 2 degrees Fig. 2. Map of attentional bias. highly decoupled model, which has one variable per location, allows for the implemen- tation of local phenomena, as, for example, we foresee may occur in case of habituation to stimuli coming from a fixed location. More broadly, the ability to model various types of interactions among these subprocesses is relevant to issue of modularity in human cognitive systems. In what follows we denote with VGi , VSi and VPi respectively, the i-th variable of the goal, saliency and priority map. We note that i ranges in {-90,...,+90}, and the domain to quantify attentional bias uses normalized units (0-1, in .01 increments). The goal map indexes top-down attention bias, and is a function of the central executive in verbal models. It models top-down, voluntary focus of attention to a location, and has a pro- gressive, symmetrical decrease in attentional bias away from the attended location. We currently consider three options for modeling the attentional bias in the goal map given that location A = a is (voluntarily) attended. We express them as a sets of constraints each of which is defined over variable A and VGi . In what follows we indicate the tuple of values which is allowed by the constraint. – Standard Gaussian Distribution: −|a−i|2 2∗d2 (A = a, VGi = GG e G ) where dG is the standard deviation of the goal map and G is the height of its peak. – Modified Gaussian distribution with inhibition: −|a−i|2 −|a−i|2 2 2 (A = a, VGi = γ(GG e 2∗dG1 )) + γ(GG − GG e 2∗dG2 )), notice that this obtained as the sum of a Gaussian and an inverted Gaussian. G is the maximum of the two functions and γ is a parameter that we use to weight the Goal Map Variables Constraints Priority Map Variables A"en%on Loca%on Saliency Map Variables Fig. 3. Variables and constraints representing the three maps and their interconnections. For clarity, only the constraints relative to the variables corresponding to the [-90◦ , -88◦ ] location are shown. components. We also have two standard deviations for the components which are denoted by dG1 and dG2 . In this way we obtain the desired shape, which has a peak at the attended location, then dips down to an area of lower attentional bias and then increases and stabilizes as we move far away from the attended location (see Figure 4 (a)). – Constant function: (A = a, VGi = k), where k is a constant value. The different shapes for the goal map when the attended location is 0◦ are shown in Figure 4 (a). We note that in Figure 3 the standard Gaussian distribution is shown on top of the variables corresponding to the goal map. Similarly we consider two options for the saliency map, which models how attention is allocated to a stimulus given how salient its characteristics are. Again, in our model, this amounts to defining constraints between variable A and each saliency map variable. The two options are: – Inverted Gaussian distribution: −|a−i|2 2 (A = a, VSi = GS − GS e 2∗dS ) where dS is the standard deviation for the saliency map, and G is its maximum value. – Constant function: (A = a, VSi = k), Goal Maps Saliency Maps Inverted Gaussian Constant -90° 0° +90° -90° 0° +90° (b) (a) Fig. 4. Candidate shapes for the goal map (a) and saliency map (b) when the attended location is A=0◦ . where k is a constant value. The two options for the saliency map are shown in Figure 4 (b). In Figure 3, the saliency map is shown as an inverted Gaussian distribution. Finally the priority map is defined as a weighted sum of the contributions of the goal and saliency map: (VGi = u, VSi = v, VPi = αu + βv) with α and β between 0 and 1. We elaborate on the cognitive interpretation of these hypothesis in Section 5. 4 Behavioral experiments We developed a behavioral task to map attentional gradients and to test the above model. It is a hybrid of our spatial target detection task [31, 17] and work on distraction from changing a task-irrelevant stimulus feature [35]. White noise is presented from the 5 locations in the frontal plane (−90◦ , −45◦ , 0◦ , +45◦ , +90◦ ), and subjects respond in each trial by discriminating a non-spatial feature (amplitude modulation (AM) rate, 25 or 75 Hz). The slow AM rate sounds like a deck of cards being shuffled while the faster rate is perceived as a buzz. Most stimuli come from a standard location (p = .84) but sometimes shift to a distractor location (p = .04/location). Separate blocks have the standard at −90◦ , 0◦ , or + 90◦ (counterbalanced). Figure 5 plots reaction times x location for each standard condition in absolute space (A), as well as the deviant location relative to the standard location (B). There were two main results. First, all conditions had slower responses to distractors vs. standards (p < .001), indicating attention shift costs. The reaction time x location function is more prominent for the left vs. the right standard (p < .01), suggesting that it is faster to shift auditory attention from right-to-left than from left-to-right. The 0◦ standard has an in- crease at near ±45◦ locations, similar to the left standard, but a decrease for the ±90◦ Fig. 5. Basic Attention Task: Reaction Time Results & Modeling locations, similar to the right standard (p < .001). Accuracy was very high (> 95%). The basic results were replicated in new subjects (n = 12, p < .01). (C). Second, in each condition reaction times sped-up for the farthest distractor location (p < .001). This was seen in each subject’s first block, so is not due to carry-over effects from pre- vious standard locations. The faster responses at far distractors cannot be accounted for by a graded reduction in bias from the attended location (goal map alone). Instead, the heightened bias to far distractors is modeled by the saliency map. A control condition in new subjects and equal probability at all locations had no differences in reaction time (n = 20, p = .83), ruling out accounts based on perceptual differences among locations. In D the inverse of the reaction times is given to show how attention bias is theorized to relate to reaction time (greater bias → faster reaction times). 5 Results We have considered two combinations of options for the goal and saliency map and we have compared how well the emerging priority map fits the behavioral experiments data. The combinations that we have considered are – Hypothesis 1: standard Gaussian centered for the goal map and inverted Gaussian for the saliency map, both centered at the attention location; – Hypothesis 2: modified Gaussian with inhibition for the goal map and inverted Gaussian for the saliency map, both centered at the attention location;. We recall that the goal map models the voluntary focus of attention. Both options that we consider model a peak of attention around the attended location and then a sym- metrical region of lower attentional bias away from the attended location. In addition, the modified Gaussian assumes an area of inhibited attention around the peak. As far as the saliency map, the inverted Gaussian shape is consistent with results which we have observed in our experiments above. which suggest that bottom-up attentional bias progressively increases away from the attended location. We have, in addition, assumed the same maximum level G = GG = GS for the goal and the saliency map. This corresponds to saying that peak attention levels gen- erated by the top-down and bottom-up component are the same, which is reasonable, in particular if the components are thought of as independent of each other. Another similar constraint which we plan to consider in the future is fixing the overall amount of attentional bias to be constant across the maps. We also assume α = β , thus taking the view that the top-down and bottom com- ponents equally contribute to the overall attentional bias. We call this parameter δ. We have fitted the data by using a stochastic local search approach on the parameters of the functions, which we recall are: G, dG , dS and δ for hypothesis 1 and G, dG1 , dG2 ,dS , γ and δ for hypothesis 2. The evaluation function we have used is the sum of squared errors: X E(p) = (dx − p(x))2 x∈{−90◦ ,−45◦ ,0◦ ,+45◦ ,+90◦ } where dx is the bias associated to location x in the experimental data and p(x) is the value associated to x by the priority map. We performed the fit for all three attended locations, that is, -90◦ , 0◦ and +90◦ . The best results for all three locations have been obtained by Hypothesis 1 with fitting values (1 − E(p)) equal to: 0.943 for 0◦ , 0.739 for +90◦ and 0.904 for -90◦ . The corresponding fitting values for Hypothesis 2 are instead: 0.903 for 0◦ , 0.501 for +90◦ and 0.850 for -90◦ . As it can be seen Hypothesis 1 outperforms Hypothesis 2 in particular in the +90◦ case. This is in part due to the asymmetry between +90◦ and - 90◦ data. However, Hypothesis 1 fits better in both cases despite being symmetric. In Figure 6 we show all the maps corresponding to the best fit superposed with the data. 6 Future directions We have a rich agenda of future directions. Impact on goal map of short-term memory load. We hypothesize that, relative to no load, the addition of a short-term memory load, such as first memorizing three words and then performing several trials of the attention task, increases reaction times to distractors near the attended location, decreases reaction times at far locations, and increases inter-trial variability. Future experiments will assess the role of short-term memory loads on the goal map. We conjecture that memory load should impair top- down control as task-specific information and load both rely on short-term memory. The rationale is that the predicted reaction time effects would be in range of the goal map but not at far locations (relative to the attended location) which are mediated by the saliency map. Load effects will be modeled by introducing a probability distribution over the goal map options and assuming that memory load results in some trials with equal attentional biases across locations (constant function). This should also increase reaction time variability. h h h Fig. 6. Fitting results. Loudness and attention gradient. We plan to investigate whether intensity changes de- crease reaction times near the standard location, with progressively smaller increases at more distant locations. Loud sounds induce automatic orienting, which our model would represent with the saliency map attentional bias. Saliency, and attentional ori- enting, can also follow from the absence of an expected sound, such as an engine un- expectedly stopping (or give a better example). The study will test whether attentional bias due to changes in stimulus intensity follows from saliency (increases or decreases bias attention) or loudness (only increases bias attention). Behavioral experiments will distinguish between an alternative hypothesis that intensity changes slow responding due to shifting attention. We expect locations near the standard to receive little ben- efit from saliency due to proximity to the goal map focus, and would be enhanced by intensity-based saliency bias. At far locations the saliency map is already tuned to bottom-up inputs, and has less to gain by intensity changes. Intensity effects are ex- pected to vary by standard location (−90◦ > 0◦ > +90◦ ). The basic task will be used with manipulations of stimulus intensity. Intensity effects will be modeled by the intro- duction of new variable having as values the intensity levels of the stimulus. We will modify the inverted Gaussian option for the saliency map with a new parameter obtain- −|a−i|2 2∗d2 ing K(G − Ge S ), where the new parameter is denoted by K. Moreover, we will add new constraints connecting the location variable A, the new intensity variable and the saliency map variables. Location probability and attention gradients. Finally we plan to understand if, as we conjecture, the probability of a stimulus at a given location is negatively associated with reaction time. It has been shown that attentional bias is strongly dependent on expectations (Itti and Baldi, 2009) and that the degree of expectation depends on base rate, with unlikely events having large evoked potentials [19]. Our preliminary data show that if distractors are improbable reaction times increase and then decrease across locations, but if equiprobable they do not differ among locations. Two experiments will test the role of stimulus probability in attention gradients. One will manipulate distractor base rate in equal increments (p = .04, .12, .20), separate blocks. The other experiment will maintain the usual standard probability (p = .84) and tests whether increasing distractor probability near the standard location (nearest 45◦ p = .07, other 3 distractors p = .03) will shift the reaction time curve away from the standard location and if decreased probability near 45◦ (p = .01, other 3 distractors p = .05) shifts it toward the standard. In terms of the computational model we will have to incorporate a way to handle sequences of stimuli. This could be done in different ways. For example we might introduce dedicated new location and intensity variables indexed with the position of the stimulus in the sequence. Another option would be to have a unique “stimulus” variable with a more structured domain, for example comprising or triple (position in stimuli sequence, location, intensity). 7 Conclusions We have presented a constraint-based model of auditory spatial attention. Our model is based on a well established decomposition in top-down and bottom-up components and, to the best of our knowledge, it is the first one focusing on the auditory system. Constraints allow for a high degree of flexibility in terms of hypotheses testing. Our ini- tial results in terms of fitting experimental data are very promising and bring interesting insight on the role and interplay of the two components in the distribution of auditory attention in space. Acknowledgements. This work is supported by NIH under grant number R01-DC015736. References [1] Vera AH, Howes A, McCurdy M, and Lewis RL. A constraint satisfaction approach to predicting skilled interactive cognition. In Proceedings of the 2004 Conference on Human Factors in Computing Systems (CHI 2004), pages 121–128, 2004. [2] Allen Allport. Foundations of Cognitive Science. pages 631–682. MIT Press, Cambridge, MA, USA, 1989. [3] J.R. Anderson. Rule of the Mind. Laurence Erlbaum Assoc, 1993. [4] Stig Arlinger, Thomas Lunner, Bjrn Lyxell, and M. Kathleen Pichora-Fuller. The emergence of cognitive hearing science. Scandinavian Journal of Psychology, 50(5):371–384, October 2009. [5] E. Awh and H. Pashler. Evidence for split attentional foci. Journal of Experimental Psychol- ogy. Human Perception and Performance, 26(2):834–846, April 2000. [6] Alan Baddeley. Working memory. Current Biology, 20(4):R136–R140, February 2010. [7] Duncan P. Brumby, Andrew Howes, and Dario D. Salvucci. A Cognitive Constraint Model of Dual-task Trade-offs in a Highly Dynamic Driving Task. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, CHI ’07, pages 233–242, New York, NY, USA, 2007. ACM. [8] Serge Caparos and Karina J. Linnell. The spatial focus of attention is controlled at per- ceptual and cognitive levels. Journal of Experimental Psychology. Human Perception and Performance, 36(5):1080–1107, October 2010. [9] Patrick Cavanagh and George A. Alvarez. Tracking multiple targets with multifocal attention. Trends in Cognitive Sciences, 9(7):349–354, July 2005. [10] K. R. Cave and N. P. Bichot. Visuospatial attention: beyond a spotlight model. Psychonomic Bulletin & Review, 6(2):204–223, June 1999. [11] Marvin M. Chun, Julie D. Golomb, and Nicholas B. Turk-Browne. A taxonomy of external and internal attention. Annual Review of Psychology, 62:73–101, 2011. [12] N. Cowan. Evolving conceptions of memory storage, selective attention, and their mu- tual constraints within the human information-processing system. Psychological Bulletin, 104(2):163–191, September 1988. [13] Bert De Coensel and Dick Botteldooren. A model of saliency-based auditory attention to environmental sound. In Proceedings of the 20th International Congress on Acoustics (ICA - 2010), pages 1–8, 2010. [14] Susan L. Epstein. For the right reasons: The FORR architecture for learning in a skill domain. Cognitive Science, 18(3):479–511, 1994. [15] C. L. Folk, R. W. Remington, and J. C. Johnston. Involuntary covert orienting is contingent on attentional control settings. Journal of Experimental Psychology. Human Perception and Performance, 18(4):1030–1044, November 1992. [16] Daniel Gall and Thom W. Frühwirth. Exchanging conflict resolution in an adaptable imple- mentation of ACT-R. TPLP, 14(4-5):525–538, 2014. [17] Edward J. Golob and John L. Holmes. Cortical mechanisms of auditory spatial attention in a target detection task. Brain Research, 1384:128–139, April 2011. [18] P. M. Greenwood and R. Parasuraman. Scale of attentional focus in visual search. Percep- tion & Psychophysics, 61(5):837–859, July 1999. [19] Sabine Grimm and Carles Escera. Auditory deviance detection revisited: evidence for a hierarchical novelty system. International Journal of Psychophysiology: Official Journal of the International Organization of Psychophysiology, 85(1):88–92, July 2012. [20] L. Itti and C. Koch. Computational modelling of visual attention. Nature Reviews. Neuro- science, 2(3):194–203, March 2001. [21] William James. The principles of psychology. New York : Holt, 1890. [22] Nilli Lavie. Distracted and confused?: selective attention under load. Trends in Cognitive Sciences, 9(2):75–82, February 2005. [23] Stephan Lewandowsky and Simon Farrell. Computational Modeling in Cognition: Princi- ples and Practice. SAGE Publications, November 2010. [24] Hans Liljenstrom. Neural Stability and Flexibility: A Computational Approach. Neuropsy- chopharmacology, 28(S1):S64–S73, 2003. [25] Mackworth. The breakdown of vigilance during prolonged visual search. The Quarterly Journal of Experimental Psychology, 1:6–21, 1948. [26] G. R. Mangun and S. A. Hillyard. Spatial gradients of visual attention: behavioral and electrophysiological evidence. Electroencephalography and Clinical Neurophysiology, 70(5):417–428, November 1988. [27] T. A. Mondor and R. J. Zatorre. Shifting and focusing auditory spatial attention. Journal of Experimental Psychology. Human Perception and Performance, 21(2):387–409, April 1995. [28] Notger G. Mller, Maas Mollenhauer, Alexander Rsler, and Andreas Kleinschmidt. The attentional field has a Mexican hat distribution. Vision Research, 45(9):1129–1137, April 2005. [29] Damiano Oldoni, Bert De Coensel, Michiel Boes, Michal Rademaker, Bernard De Baets, Timothy Van Renterghem, and Dick Botteldooren. A computational model of auditory at- tention for use in soundscape research. The Journal of the Acoustical Society of America, 134(1):852–861, July 2013. [30] W. B. Pillsbury. Attention. Half-title: Library of philosophy. Ed. by J. H. Muirhead. S. Sonnenschein & Co., ltd. The Macmillan co., London, New York, 1908. [31] S. K. Rader, J. L. Holmes, and E. J. Golob. Auditory event-related potentials during a spatial working memory task. Clinical Neurophysiology: Official Journal of the International Federation of Clinical Neurophysiology, 119(5):1176–1189, May 2008. [32] Gillian Rhodes. Auditory attention and the representation of spatial information. Perception & Psychophysics, 42(1):1–14, January 1987. [33] C. Rorden and J. Driver. Spatial deployment of attention within and across hemifields in an auditory task. Experimental Brain Research, 137(3-4):487–496, April 2001. [34] Francesca Rossi, Peter van Beek, and Toby Walsh, editors. Handbook of Constraint Pro- gramming. Elsevier Science Inc., New York, NY, USA, 2006. [35] E. Schrger and C. Wolff. Attentional orienting and reorienting is indicated by human event- related brain potentials. Neuroreport, 9(15):3355–3358, October 1998. [36] Charles Spence and Jon Driver. Audiovisual links in exogenous covert spatial orienting. Perception & Psychophysics, 59(1):1–22, January 1997. [37] P. L. Wachtel. Conceptions of broad and narrow attention. Psychological Bulletin, 68(6):417–429, December 1967. [38] Stuart N. Wrigley and Guy J. Brown. A computational model of auditory selective attention. IEEE transactions on neural networks / a publication of the IEEE Neural Networks Council, 15(5):1151–1163, September 2004. [39] R. J. Zatorre, T. A. Mondor, and A. C. Evans. Auditory attention to space and frequency activates similar cerebral systems. NeuroImage, 10(5):544–554, November 1999.