<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">I Know What You&apos;re Doing: A Case Study on Case-Based Opponent Modeling and Low-Level Action Prediction</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Thomas</forename><surname>Gabel</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Faculty of Computer Science and Engineering</orgName>
								<orgName type="institution">Frankfurt University of Applied Sciences</orgName>
								<address>
									<postCode>60318</postCode>
									<settlement>Frankfurt am Main</settlement>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<author role="corresp">
							<persName><forename type="first">Eicke</forename><surname>Godehardt</surname></persName>
							<email>tgabel|godehardt@fb2.fra-uas.de</email>
							<affiliation key="aff0">
								<orgName type="department">Faculty of Computer Science and Engineering</orgName>
								<orgName type="institution">Frankfurt University of Applied Sciences</orgName>
								<address>
									<postCode>60318</postCode>
									<settlement>Frankfurt am Main</settlement>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">I Know What You&apos;re Doing: A Case Study on Case-Based Opponent Modeling and Low-Level Action Prediction</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">646E26FF9AB1E16604EF3B3681B02512</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T21:32+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>This paper focuses on an investigation of case-based opponent player modeling in the domain of simulated robotic soccer. While in previous and related work it has frequently been claimed that the prediction of low-level actions of an opponent agent in this application domain is infeasible, we show that -at least in certain settings -an online prediction of the opponent's actions can be made with high accuracy. We also stress why the ability to know the opponent's next low-level move can be of enormous utility to one's own playing strategy.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Opponent Modeling in Robotic Soccer Simulation</head><p>RoboCup <ref type="bibr" target="#b11">[12]</ref> is an international research initiative intending to expedite artificial intelligence and intelligent robotics research by defining a set of standard problems where various technologies can and ought to be combined solving them. Annually, there are championship tournaments in several leagues -ranging from rescue tasks over real soccer-playing robots to simulated ones.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>Recognizing and predicting agent behavior is of crucial importance specifically in adversary domains. The case study presented in this paper is concerned with the prediction of the low-level behavior of agents in the highly dynamic, heterogeneous, and competitive domain of robotic soccer simulation (RoboCup). Case-based reasoning represents one of the potentially useful methodologies for accomplishing the analysis of the behavior of a single or a team of agents. In this sense, the basic idea of our approach is to make a case-based agent observe its opponent and, in an online fashion, i.e. during real game play, build up a case base to be used for predicting the opponent's future actions.</p><p>In Section 2, we introduce the opponent modeling problem, point to related work, and argue why knowing an opponent's next low-level actions can be beneficial. The remainder of the paper then outlines our case-based methodology (Section 3), reviews the experimental results we obtained (Section 4), and summarizes and discusses our findings (Section 5).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1">Robotic Soccer Simulation</head><p>The focus of the paper at hand is laid upon RoboCup's 2D Simulation League, where two teams of simulated soccer-playing agents compete against one another using the Soccer Server <ref type="bibr" target="#b9">[10]</ref>, a real-time soccer simulation system.</p><p>The Soccer Server allows autonomous software agents written in an arbitrary programming language to play soccer in a client/server-based style: The server simulates the playing field, communication, the environment and its dynamics, while the clients -eleven autonomous agents per team -connect to the server and are permitted to send their intended actions (e.g. a parameterized kick or dash command) once per simulation cycle to the server via UDP. Then, the server takes all agents' actions into account, computes the subsequent world state and provides all agents with (partial) information about their environment via appropriate messages over UDP.</p><p>So, decision making must be performed in real-time or, more precisely, in discrete time steps: Every 100ms the agents can execute a low-level action and the world-state will change based on the individual actions of all players. Speaking about low-level actions, we should make clear that the actions themselves are "parameterized basic actions" and the agent can execute only one of them per time step:</p><p>dash(x, α) -lets the agent accelerate along its current body orientation by relative power x ∈ [0, 100] (if it does not accelerate, then its velocity decays) into direction α ∈ (−180, 180] relative to its body orientation turn(α) -makes the agent turn its body by α ∈ (−180, 180] where, however, the Soccer Server reduces α depending on the player's current velocity in order to simulate an inertia moment kick(x, α) -has an effect only, if the ball is within the player's kick range (1.085m around the player) and yields a kick of the ball by relative power x ∈ [0, 100] into direction α ∈ (−180, 180] -There exist a few further actions (like tackling<ref type="foot" target="#foot_0">1</ref> , playing foul, or, for the goal keeper, catching the ball) whose exact description is beyond scope.</p><p>Given this short description of the most important low-level actions that can be employed by the agent, it is clear that these basic actions must be combined cleverly in consecutive time steps in order to create "higher-level actions" like intercepting balls, playing passes, doing dribblings, or marking players. We will call those higher-level actions skills in the remainder of this paper. Robotic Soccer represents an excellent testbed for machine learning, including approaches that involve case-based reasoning. For example, several research groups have dealt with the task of learning parts of a soccer-playing agent's behavior autonomously (for instance <ref type="bibr" target="#b8">[9,</ref><ref type="bibr" target="#b7">8,</ref><ref type="bibr" target="#b2">3]</ref>). In <ref type="bibr" target="#b5">[6]</ref>, as an other example, we specifically addressed the issue of using CBR for the development of a player agent skill for intercepting balls.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2">Related Work on Opponent Modeling</head><p>Opponent modeling is an important factor that can contribute substantially to a player's capabilities in a game, since it enables the prediction of future actions of the opponent. In doing so, it also allows for adapting one's own behavior accordingly. Case-based reasoning has been frequently used as a technique for opponent modeling in multi-agent games <ref type="bibr" target="#b3">[4]</ref>, including the domain of robotic soccer <ref type="bibr" target="#b12">[13,</ref><ref type="bibr" target="#b0">1]</ref>.</p><p>Using CBR, in <ref type="bibr" target="#b12">[13]</ref> the authors make their simulated soccer agents recognize currently executed higher-lever behaviors of the currently ball leading opponent player. These include passing, dribbling, goal-kicking and clearing. These higherlevel behaviors correspond to what we refer to as skills, i.e. action sequences that are executed over a dozen or more time steps. This longer time horizon allows the agent to take appropriate counter measures.</p><p>The authors of <ref type="bibr" target="#b10">[11]</ref> also deal with the case-based recognition of skills (higherlevel behaviors, to be exact the shoot-on-goal skill) executed by an opponent soccer player, focusing on the appropriate adjustment of the similarity measure employed. While we do also think opponent modeling is useful for counteracting adversary agents, we, however, disagree with these authors claiming that "in a complex domain such as RoboCup it is infeasible to predict an agent's behavior in terms of primitive actions". Instead we will show empirically that such a lowlevel action prediction can be achieved during an on-going play using case-based methods. To this end, the work presented in this paper is also related to the work by Floyd et al. <ref type="bibr" target="#b4">[5]</ref> whose goal is to mimic the overall behavior of entire soccer simulation teams, be it for the purpose of analysis or for rapid prototyping when developing one's own team, without putting too much emphasis on whether the imitators yield competitive behavior.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.3">Related Previous Work</head><p>What is the use of knowing exactly whether an opponent is going to execute a kick(40, 30 • ) or a dash(80, 0 • ) low-level action next? This piece of information certainly does not reveal whether this opponent's intention is to play a pass (and to which teammate) in the near future or to dribble along. Clearly, for answering questions like that the approaches listed in the previous section are potentially more useful. But knowing the opposing agent's next low-level actions is extremely useful, when knowing the next state on the field is essential (cf. Figure <ref type="figure">1</ref> for an illustration).</p><p>In <ref type="bibr" target="#b6">[7]</ref>, we considered a soccer simulation defense scenario of crucial importance: We focused on situations where one of our players had to interfere and disturb an opponent ball leading player in order to scotch the opponent team's attack at an early stage and, even better, to eventually conquer the ball initiating a counter attack. We employed a reinforcement learning (RL) methodology that enabled our agents to autonomously acquire such an aggressive duel behavior, and we successfully embedded it into our soccer simulation team's defensive strategy. So, the goal was to learn a so-called "duelling skill" (i.e. a higher-level behavior which in the end yields a sequence of low-level actions) which made our agent conquer the ball from the ball-leading opponent. Fig. <ref type="figure">1</ref>. In the top-left we see the current state of an opponent agent in ball possession. If we assume, this agent does not take any low-level action, then the resulting successor state looks like the one in the bottom left figure: Player and ball have moved according to their recent velocities while the magnitude of the velocity vectors have decayed according to the rules of the simulation. How different the successor state may look, if the opponent, however, does take an action (which is most likely), is shown in the right figures. In example 1 (top) the agent accelerates full power along its current body orientation, while the ball is not affected. In example 2, the player kicks the ball with 50% power into -90 • relative to its current body orientation which yields a resulting ball velocity vector as shown in the bottom right.</p><p>An important feature of the soccer simulation domain is that the model of the environment is known. This means given, for example, the current position and velocity of the ball, it is possible for any agent to calculate the position of the ball in the next time step (because the implementation of the physical simulation by the Soccer Server is open source<ref type="foot" target="#foot_1">2</ref> ). As a second example, when knowing one's own current position, velocity and body angle, and issuing a turn(68 • ) low-level action, the agent can precalculate the position, velocity and body orientation it will have in the next step. Or, finally, when the agent knows the position and velocity of the ball, it can precalculate the ball's position and velocity in the next step, for any kick(x, α) command that it might issue.</p><p>Knowing the model of the environment (formally, the transition function p : S × A × S → R where p(s, a, s ) tells the probability to end up in the next state s when executing action a in the current state s), is extremely advantageous in reinforcement learning, since then model-based instead of model-free learning algorithms can be applied which typically comes along with a pleasant simplification of the learning task.</p><p>So, in soccer simulation the transition function p (model of the environment) is given since the way the Soccer Server simulates a soccer match is known. In the above-mentioned "duelling task", however, the situation is aggravated: Here, we have to consider the influence of an opponent whose next actions cannot be controlled. In <ref type="bibr" target="#b6">[7]</ref>, we stated that the opponent's next (low-level) actions can "hardly be predicted <ref type="bibr">[which]</ref> makes it impossible to accurately anticipate the successor state", knowing which is, as pointed out, extremely useful in RL. In the paper at hand, we will show that predicting the opponent's next low-level action might be easier than expected. As a consequence, in <ref type="bibr" target="#b6">[7]</ref> we had to rely on a rough approximation of p, that merely takes into account that part of the state that can be influenced directly by the learning agent and which ignored the part of the future state which is under direct control of the ball-leading opponent (e.g. the position of the ball in the next state). This corresponded to the unrealistic assumption of an opponent that never takes any action (cf. Figure <ref type="figure">1</ref>, bottom left).</p><p>in future work we can employ a much more accurate version of p based on the case-based prediction of the opponent's low-level actions described in the next section.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Case-Based Prediction of Low-Level Actions</head><p>In what follows, we differentiate between an opponent (OPP) agent whose next low-level actions are to be predicted as well as (our) case-based agent (CBA) that essentially observes the opponent and that is going to build up a case base to be used for the prediction of OPP's actions. When approaching the opponent modeling problem as a case-based reasoning problem, the goal of the case-based agent is to correctly predict the next action of its opponent given a characterization of the current situation. Stated differently, the current state of the system (including the case-based agent itself, its opponent as well as all other relevant objects) represents a new query q. CBA's case base C is made up of cases c = (p, s) whose problem parts p correspond to other, older situations and corresponding solutions s which describe the action OPP has taken in situation p. Next, the case-based agent will search its case base for that case ĉ = (p, ŝ) ∈ C (or for a set of k such cases) whose problem part features the highest similarity to the current problem q and employ its solution ŝ as the current prediction of the opponent's next action.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">Problem Modeling</head><p>In the context of this case study we focus on dribbling opponents, i.e. the opponent has the ball in its kick range and moves along while keeping the ball within its kick range all the time. Stated differently, we focus on situations where OPP behaves according to some "dribble skill" (a higher-level dribbling behavior). Consequently, OPP executes in each time step one of the three actions kick(x, α), dash(x, α), or turn(α). The standard rules of the simulation allow x to be from [0, 100] and α from (−180 • , 180 • ] for kicks and turns. For dashes, α is allowed to take one out of eight values (multiples of 45 • ). In almost all cases occurring during normal play, however, a dribbling player is heading more or less towards his opponent's goal which is why the execution of low-level turn actions represents an exceptional case. Therefore, for the time being, we leave turn actions aside and focus on the correct prediction of dashes and kicks including their parameters x and α.</p><p>Case Structure The state of the dribbling opponent (OPP) can be characterized by the x and y position of the ball within its kick range (pos b,x and pos b,y ) relative to the center of OPP as well as the x and y components of the ball's velocity (vel b,x and vel b,y ; of course, these values are also relative to OPP's body orientation). Moreover, OPP's x and y velocities (vel p,x and vel p,y ) are of relevance, making six features in total. The seventh relevant feature, OPP's current body orientation θ p can be skipped due to the arguments mentioned in the preceding paragraph. Furthermore, the y component of OPP's velocity vector vel p,y is, in general, zero since a dribbling player almost always dribbles along its current body orientation. While this allows us to also skip the sixth feature, we remove a redundancy in the remaining features (and thus arrive at only four of them) by changing to a relative state description that incorporates some background knowledge <ref type="foot" target="#foot_2">3</ref>  where all components characterize the next state as it would arise, if the agent would not take any action (cf. Figure <ref type="figure">1</ref>).</p><p>The solution s of a case c = (p, s) consists of a class label l ("dash" or "kick") as well as two accompanying real-valued attributes for the power x and angle α parameters of the respective action. Thus, the solution is a triple s = (l, x, α).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">Implementing the CBR Cycle</head><p>The case-based agent CBA observes his opponent OPP and, in doing so, builds up its case base. Note that all agents in soccer simulation act on incomplete and uncertain information. Their visual input consists of noisy information about objects in their limited field of vision. However, if the observed opponents are near and constantly focused at, CBA is provided with sufficiently accurate visual state information. In order to fill the contents of the cases' solution parts, however, CBA must apply inverse dynamics of the soccer simulation. If CBA, for example, observes that the velocity vector of the ball has been changed at time t + 1 as in the bottom right part of Figure <ref type="figure">1</ref>, then it can conclude that OPP has executed a kick(50, −90 • ) action at time t and can use that information to complete the case it created at time step t.</p><p>With ongoing observation of dribbling opponent players, CBA's case base C grows and becomes more and more competent. Therefore, after |C| exceeds some threshold, CBA can utilize its case base and query it to find a prediction of the action that OPP is going to take in the current time step.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Retrieval and Similarity Measures</head><p>We model the problem similarity using the local-global principle <ref type="bibr" target="#b1">[2]</ref> with identical local similarity measures for all problem attributes, sim i (q i , c i ) = ( qi−ci maxi−mini ) 2 , where min i and max i denote the minimum and maximum value of the domain of the ith feature. The global similarity is formed as a weighted average according to</p><formula xml:id="formula_0">Sim(q, c) = n i=1 w i • sim i (q i , c i ) n i=1 w i</formula><p>where attributes pos bnx and pos bny are weighted twice as much as vel pnx and vel pny .</p><p>We perform standard k-nearest neighbor retrieval using a value of k = 3 in our experiments. When predicting the class of the solution, i.e. the type of the low-level action (dash or kick), we apply a majority voting, and for the prediction of the action parameters (x and α) we calculate the average over all cases among the k nearest neighbors whose class label matches the majority class.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Experimental Results</head><p>To evaluate our approach we selected a set of contemporary soccer simulation team binaries (top teams from recent years) and made one of their agents (OPP) dribble for up to 2000 simulated time steps<ref type="foot" target="#foot_3">4</ref> . Our case-based agent CBA was allowed meanwhile to observe OPP and build up its case base. We evaluated CBA's performance in predicting OPP's low-level actions for increasing case base sizes.</p><p>Figure <ref type="figure" target="#fig_1">2</ref> visualizes the learning progress against an opponent agent from team WrightEagle. As can be seen, compelling accuracies can be achieved for both, the correctness of the type of the action (dash or kick) as well as for the belonging action parameters. Interestingly, even the relative power / angle of kicks can be predicted quite reliably with a remaining absolute error of less than ten percent / ten degrees. Figure <ref type="figure" target="#fig_2">3</ref> focuses on different opponent agents and highlights the fact that a substantial improvement in action type prediction accuracy can be obtained with as little as 100 collected cases. Baseline to all these classification experiments is the error of the "trivial" classifier (black) that predicts each action type to be of the majority class. The right part of Figure <ref type="figure" target="#fig_2">3</ref> presents the recall of both, dash and kick actions. Apparently, dashes are somewhat easier to predict than kicks where, however, the recall of the latter is still above 65% for each of the opponent agents considered.  In Figure <ref type="figure">4</ref>, we present aggregate numbers (averages over all opponents) that emphasize how accurately the parameters of an action were predicted, given that the type of the action could be identified correctly. To this end, dash angles α are disregarded since more than 99.2% of all dash actions performed used α = 0, i.e. yielded a dash forward. Here, we compare (a) an "early" case base with only 10 cases, (b) an intermediate one <ref type="foot" target="#foot_4">5</ref> with |C| = 100 as well as (c) one that has resulted from 2000 simulated time steps and contains circa 1500 cases. Interestingly, even in (a) comparatively low errors can be obtained. In (b) and (c), however, the resulting average absolute prediction errors become really competitive (±2.9 for dash powers x with x ∈ [0, 100], ±6.3 for kick powers x with x ∈ [0, 100], and ±19.7 • for kick angles α with α ∈ [0 Fig. <ref type="figure">4</ref>. Exactness of the prediction of the action parameters for different stages during ongoing learning (10, 100, and ≈1500 cases in the case base). Left: Average error of the predicted angle of a kick action. Right: Average error of the predicted relative power of a kick action and dash action (averages over agents from all opponent teams considered).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Discussion and Conclusion</head><p>Clearly, dribbling opponents are very likely to behave differently when they are disturbed, tackled, or attacked by a nearby opponent. Therefore, the approach presented needs to be extended to "duelling situations" as they frequently arise in real matches. For example, in scenarios like that the dribbler will presumably not just dribble straight ahead, but also frequently execute turn actions (e.g. in order to dribble around its disturber). This represents an aggravation of the action type prediction problem since then three instead of two classes of actions must be considered (dask, kick, turn).</p><p>While the case study presented focused solely on non-attacked dribbling opponents, this approach can easily be transferred to related or similar situations where knowing the opponent's next move is crucial, too. This includes, but is not limited to the behavior of an opponent striker when trying to perform a shoot onto the goal (which typically requires a couple of time steps), the behavior of the shooter as well as the goal keeper during penalty shoot-outs, or the positioning behavior of the opponent goalie (anticipating which can be essential for the striker).</p><p>As a next step, we plan to combine the presented case-based prediction of lowlevel actions with the reinforcement learning-based acquisition of agent behaviors as outlined in Section 2.3. This involves, first, solving the aggravated problem of correctly recognizing three different classes of low-level actions mentioned at the beginning of this section and, second, a proper utilization of the thereby obtained improved model when learning a higher-level duelling skill using RL. Another interesting direction for future work is the idea to let CBA start off with some opponent model in form of a case-base acquired offline (against, for example, an older version of the team to be faced) and, using appropriate techniques for case base maintenance, to successively replace old experience by new experience gained online during the current match.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head></head><label></label><figDesc>from the simulation. Hence, the problem part p of a case c = (p, s) is a four-tuple p = (pos bnx , pos bny , vel bnx , vel bny ) with pos bnx = pos b,x + 0.94 • vel b,x − 0.4 • vel p,x pos bny = pos b,y + 0.94 • vel b,y − 0.4 • vel p,y vel pnx = 0.94 • vel b,x − 0.4 • vel p,x vel pny = 0.94 • vel b,y − 0.4 • vel p,y</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Fig. 2 .</head><label>2</label><figDesc>Fig. 2. Progress of CBA's competence in predicting the next low-level actions of a dribbling opponent agent from team WrightEagle. A case base of about 1500 cases was created during the course of 2000 simulated time steps.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Fig. 3 .</head><label>3</label><figDesc>Fig.3. Left: Case-based prediction of the type (dash or kick) of the next low-level action for opponents from different teams. Right: Recall, i.e. share of dashes that were correctly predicted as dashes and kicks that were correctly predicted as kicks.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head></head><label></label><figDesc>• , 360 • ]).</figDesc><table><row><cell>32,1</cell><cell>29,3 32,3</cell><cell>19,7 23,6 35,3</cell><cell>|CB|=10 |CB|=100 |CB|~1500</cell><cell>Average Error (within [0,100])</cell><cell>14 6 8 10 12 4</cell><cell>8,3</cell><cell>3,7</cell><cell>2,9</cell><cell>12,3</cell><cell>9,0</cell><cell>6,3</cell></row><row><cell></cell><cell cols="2">Average Error in Predicted Kick Angle (within [0°,360°]),</cell><cell></cell><cell></cell><cell>2</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell></cell><cell></cell><cell>Standard Deviations in Gray</cell><cell></cell><cell></cell><cell>0</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell cols="3">Average Error in Dash Power</cell><cell cols="3">Average Error in Kick Power</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">To tackle for the ball with a low-level action tackle(α) means to straddle for the ball and thus changing its velocity, even if it is not in the player's immediate kick range; such an action succeeds only with limited probability which decreases the farther the ball is away from the agent.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_1">In practice, the Soccer Server adds some noise to all low-level actions executed, but this is of minor importance to our concerns.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_2">Knowledge about how the Soccer Server decays objects.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_3">Duration of a regular match is 6000 time steps.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_4">A case base of a size of about 100 to 500 cases can easily be created within the first half of a match for most players.</note>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Using a Two-Layered Case-Based Reasoning for Prediction in Soccer Coach</title>
		<author>
			<persName><forename type="first">M</forename><surname>Ahmadi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Keighobadi-Lamjiri</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Nevisi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Habibi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Badie</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the International Conference of Machine Learning; Models, Technologies and Applications (MLMTA&apos;03</title>
				<meeting>the International Conference of Machine Learning; Models, Technologies and Applications (MLMTA&apos;03</meeting>
		<imprint>
			<publisher>CSREA Press</publisher>
			<date type="published" when="2003">2003</date>
			<biblScope unit="page" from="181" to="185" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Utility-Oriented Matching: A New Research Direction for Case-Based Reasoning</title>
		<author>
			<persName><forename type="first">R</forename><surname>Bergmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Richter</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Schmitt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Stahl</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Vollrath</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 9th German Workshop on Case-Based Reasoning</title>
				<meeting>the 9th German Workshop on Case-Based Reasoning</meeting>
		<imprint>
			<date type="published" when="2001">2001</date>
			<biblScope unit="page" from="264" to="274" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Reinforcement Learning for the Soccer Dribbling Task</title>
		<author>
			<persName><forename type="first">A</forename><surname>Carvalho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Cheriton</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of IEEE Conference on Computational Intelligence and Games (CIG)</title>
				<meeting>IEEE Conference on Computational Intelligence and Games (CIG)<address><addrLine>Seoul, South Korea</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2011">2011</date>
			<biblScope unit="page" from="95" to="101" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Improving Modeling of Other Agents Using Stereotypes and Compactification of Observations</title>
		<author>
			<persName><forename type="first">J</forename><surname>Denzinger</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Hamdan</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of Third International Joint Conference on Autonomous Agents and Multiagent Systems (AAMAS)</title>
				<meeting>Third International Joint Conference on Autonomous Agents and Multiagent Systems (AAMAS)<address><addrLine>New York, USA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2004">2004</date>
			<biblScope unit="page" from="1414" to="1415" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">A Case-Based Reasoning Approach to Imitating RoboCup Players</title>
		<author>
			<persName><forename type="first">M</forename><surname>Floyd</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Esfandiari</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Lam</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 21st International Florida Artificial Intelligence Research Society Conference</title>
				<meeting>the 21st International Florida Artificial Intelligence Research Society Conference<address><addrLine>Coconut Grove, USA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2008">2008</date>
			<biblScope unit="page" from="251" to="256" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">CBR for State Value Function Approximation in Reinforcement Learning</title>
		<author>
			<persName><forename type="first">T</forename><surname>Gabel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Riedmiller</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 6th International Conference on Case-Based Reasoning (ICCBR</title>
				<meeting>the 6th International Conference on Case-Based Reasoning (ICCBR<address><addrLine>Chicago, USA</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2005">2005. 2005</date>
			<biblScope unit="page" from="206" to="221" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">A Case Study on Improving Defense Behavior in Soccer Simulation 2D: The NeuroHassle Approach</title>
		<author>
			<persName><forename type="first">T</forename><surname>Gabel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Riedmiller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Trost</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">RoboCup 2008: Robot Soccer World Cup XII, LNCS</title>
				<editor>
			<persName><forename type="first">L</forename><surname>Iocchi</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">H</forename><surname>Matsubara</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Weitzenfeld</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">C</forename><surname>Zhou</surname></persName>
		</editor>
		<meeting><address><addrLine>Suzhou, China</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2008">2008</date>
			<biblScope unit="page" from="61" to="72" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Half Field Offense in RoboCup Soccer: A Multiagent Reinforcement Learning Case Study</title>
		<author>
			<persName><forename type="first">S</forename><surname>Kalyanakrishnan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Stone</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">RoboCup-2006: Robot Soccer World Cup X</title>
				<meeting><address><addrLine>Berlin</addrLine></address></meeting>
		<imprint>
			<publisher>Springer Verlag</publisher>
			<date type="published" when="2007">2007</date>
			<biblScope unit="page" from="72" to="85" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Progress in Learning 3 vs. 2 Keepaway</title>
		<author>
			<persName><forename type="first">G</forename><surname>Kuhlmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Stone</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">RoboCup-2003: Robot Soccer World Cup VII</title>
				<meeting><address><addrLine>Berlin</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2004">2004</date>
			<biblScope unit="page" from="694" to="702" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Soccer Server: A Tool for Research on Multi-Agent Systems</title>
		<author>
			<persName><forename type="first">I</forename><surname>Noda</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Matsubara</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Hiraki</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Frank</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Applied Artificial Intelligence</title>
		<imprint>
			<biblScope unit="volume">12</biblScope>
			<biblScope unit="issue">2-3</biblScope>
			<biblScope unit="page" from="233" to="250" />
			<date type="published" when="1998">1998</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Similarity-Based Opponent Modelling Using Imperfect Domain Theories</title>
		<author>
			<persName><forename type="first">T</forename><surname>Steffens</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the IEEE Symposium on Computational Intelligence and Games (CIG05)</title>
				<meeting>the IEEE Symposium on Computational Intelligence and Games (CIG05)</meeting>
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">RoboCup 2001: The Fifth Robotic Soccer World Championships</title>
		<author>
			<persName><forename type="first">M</forename><surname>Veloso</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Balch</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Stone</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">AI Magazine</title>
		<imprint>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="issue">23</biblScope>
			<biblScope unit="page" from="55" to="68" />
			<date type="published" when="2002">2002</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Recognizing and Predicting Agent Behavior with Case-Based Reasoning</title>
		<author>
			<persName><forename type="first">J</forename><surname>Wendler</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Bach</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">RoboCup 2003: Robot Soccer World Cup VII</title>
				<editor>
			<persName><forename type="first">D</forename><surname>Polani</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Bonarini</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">B</forename><surname>Browning</surname></persName>
		</editor>
		<meeting><address><addrLine>Padova, Italy</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2004">2004</date>
			<biblScope unit="page" from="729" to="728" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
