<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Quantifying Uncertainty in Machine Theory of Mind Across Time</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Shanshan</forename><surname>Zhang</surname></persName>
							<email>shanshan.zhang@helsinki.fi</email>
							<affiliation key="aff0">
								<orgName type="institution">University of Helsinki</orgName>
								<address>
									<addrLine>Pietari Kalmin katu 5</addrLine>
									<postCode>00560</postCode>
									<settlement>Helsinki</settlement>
									<country key="FI">Finland</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Chuyang</forename><surname>Wu</surname></persName>
							<email>chuyang.wu@helsinki.fi</email>
							<affiliation key="aff0">
								<orgName type="institution">University of Helsinki</orgName>
								<address>
									<addrLine>Pietari Kalmin katu 5</addrLine>
									<postCode>00560</postCode>
									<settlement>Helsinki</settlement>
									<country key="FI">Finland</country>
								</address>
							</affiliation>
							<affiliation key="aff1">
								<orgName type="institution">University of Jyväskylä</orgName>
								<address>
									<addrLine>Seminaarinkatu 15, PL 35</addrLine>
									<postCode>40014</postCode>
									<settlement>Jyväskylä</settlement>
									<country key="FI">Finland</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Jussi</forename><forename type="middle">P P</forename><surname>Jokinen</surname></persName>
							<email>p.jokinen@jyu.fi</email>
							<affiliation key="aff1">
								<orgName type="institution">University of Jyväskylä</orgName>
								<address>
									<addrLine>Seminaarinkatu 15, PL 35</addrLine>
									<postCode>40014</postCode>
									<settlement>Jyväskylä</settlement>
									<country key="FI">Finland</country>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff2">
								<address>
									<addrLine>10.-11.6</addrLine>
									<postCode>.2024</postCode>
									<settlement>Vaasa</settlement>
									<country key="FI">Finland</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Quantifying Uncertainty in Machine Theory of Mind Across Time</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">88CC403DD6F78BAB4D4ECDFBFBEE73F5</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T17:26+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Human-Computer Interaction</term>
					<term>Machine Theory of Mind</term>
					<term>Mentalizing</term>
					<term>Uncertainty Quantification</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>As intelligent interactive technologies advance, ensuring alignment with user preferences is critical. Machine theory of mind enables systems to infer latent mental states from observed behaviors, similarly to humans. Currently, there is no formal mechanism for integrating multiple observations over time and quantifying the uncertainty of inferences as the function of accumulated evidence in a provably human-like way. This paper addresses the issue through Bayesian inference, proposing a model that maintains a posterior belief about mental states as a probability distribution, updated with observational data. The advantage of Bayesian statistics lies in the possibility of evaluating the certainty of these inferences. We validate the model's human-like mental inference capabilities through an experiment.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Theory of mind, the innate human capacity to deduce others' latent mental states from observable behavior <ref type="bibr" target="#b0">[1,</ref><ref type="bibr" target="#b1">2]</ref>, underpins social collaboration <ref type="bibr" target="#b2">[3,</ref><ref type="bibr" target="#b3">4]</ref>. As artificial intelligence (AI) advances, aligning intelligent machines with users' preferences becomes imperative <ref type="bibr" target="#b4">[5]</ref>. Achieving alignment between human and machine objectives is facilitated when machines adopt reasoning processes that can be understood by humans <ref type="bibr" target="#b5">[6]</ref>, suggesting the importance of machines emulating human mental inference. A machine theory of mind seeks to provide machines with the ability to infer mental states in a human-like manner.</p><p>Mental inference facilitates collaboration by informing the agent and impacting its actions. The idea is that if an intelligent machine has knowledge of the user's goals, it can better make decisions to help the user. However, there is also an inherent risk in making decisions based on inferences: because all inferences contain uncertainty <ref type="bibr" target="#b6">[7,</ref><ref type="bibr" target="#b7">8]</ref>, the intelligent agent should have a way of considering the amount of uncertainty when taking actions. There needs to be a way to quantify the amount of uncertainty, so that the agent can robustly consider this when choosing what actions to take. In this paper, we formalize a computational model that infers preferences of observed agents. Observations from multiple time steps are integrated, and the uncertainty associated with inferences is quantified in a posterior distribution.</p><p>The problem that our paper tackles is illustrated in Figure <ref type="figure" target="#fig_0">1</ref>. The three panels depict an evolving inference by an observer of Janice's drink preference under varying conditions in three consecutive days. Initially, Janice selects tea, but the positioning of coffee on a high shelf introduces ambiguity regarding her preference -does she favor tea, or does she simply wish to avoid climbing the kitchen ladder? This uncertainty prevents a clear inference of her preference. In the second panel, Janice uses a stool to reach the now higher-placed tea jar, while the coffee remains even further out of reach, potentially accessible with taller kitchen stairs.</p><p>The scenario hints at a preference for tea, yet the possibility that Janice may have an aversion to heights carries a degree of uncertainty, nudging the likelihood slightly in favor of tea.</p><p>The final panel of Figure <ref type="figure" target="#fig_0">1</ref> offers a decisive moment: both coffee and tea jars are easily accessible, and Janice opts for coffee. Given the equal effort required to reach both, her choice of coffee indicates a genuine preference for coffee, revealing that her earlier decisions were influenced by a reluctance to climb too high rather than a preference for tea. Consequently, our inference shifts significantly towards coffee with increased certainty. In this paper, we hypothesize that humans are able to carry out these sorts of inferences and meta-cognitively assess how certain they are in inferred preferences. Moreover, we formalize a computational model of this process.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Background Review</head><p>Theory of mind, or mentalizing, enables humans to infer others' mental states <ref type="bibr" target="#b8">[9,</ref><ref type="bibr" target="#b9">10,</ref><ref type="bibr" target="#b10">11]</ref>. It facilitates social interaction <ref type="bibr" target="#b2">[3,</ref><ref type="bibr" target="#b3">4]</ref> such as communication <ref type="bibr" target="#b11">[12,</ref><ref type="bibr" target="#b12">13]</ref> and collaboration <ref type="bibr" target="#b0">[1,</ref><ref type="bibr" target="#b1">2]</ref>. Likewise, a machine that is able to carry out mentalization can better account user variability, improving the quality of interaction <ref type="bibr" target="#b13">[14,</ref><ref type="bibr" target="#b14">15,</ref><ref type="bibr" target="#b15">16]</ref>. Experiments have demonstrated that machines capable of mentalization achieve superior performance in communication <ref type="bibr" target="#b16">[17,</ref><ref type="bibr" target="#b17">18]</ref> and team cooperation tasks <ref type="bibr" target="#b18">[19]</ref>.</p><p>Models of mentalizing target the inference of mental states such as preferences, costs <ref type="bibr" target="#b19">[20]</ref>, knowledge <ref type="bibr" target="#b20">[21]</ref>, and beliefs <ref type="bibr" target="#b8">[9]</ref>. These models incorporate psychological hypotheses concerning of observed actors as computational frameworks, enabling the simulation of predicted behavior. Parameters within the model reflect various mental states, including goals, guiding the behavior prediction for actors under specific objectives in a given context <ref type="bibr" target="#b21">[22]</ref>. Assuming the psychological underpinnings are accurate, these models can predict an actor's behavior based on their goals. Inverse modeling techniques are then employed to deduce the parameters most likely to account for the observed behavior <ref type="bibr" target="#b22">[23,</ref><ref type="bibr" target="#b23">24]</ref>.</p><p>How to create a psychologically plausible model that can be parametrized with mental states and that then simulates behavior? One emerging popular approach is called computational rationality <ref type="bibr" target="#b24">[25,</ref><ref type="bibr" target="#b25">26]</ref>. It posits that intelligent agents, such as humans, choose actions that maximize expected utility. The agent must optimize its behavior with respect to the constraints environment. In addition, the approach is sensitive to the fact that intelligent agents have internal cognitive bounds as well, such as limited knowledge and information processing capacity. The approach is suitable for computational modeling of theory of mind, because it helps to prune the space of possible explanations by assuming that the observed behavior is produced by a computational rational agent. When the bounds of the environment and the cognition are known and modeled correctly, the model can then be applied for reliable parameter inference <ref type="bibr" target="#b26">[27]</ref>.</p><p>Inferences, including those related to mentalizing, are often made under conditions of limited data, inherently involving uncertainty <ref type="bibr" target="#b27">[28,</ref><ref type="bibr" target="#b6">7]</ref>. The similarity in actions among individuals with diverse preferences in specific contexts implies that observations alone may not suffice for conclusive inferences. The complexity of social settings further amplifies this uncertainty, highlighting the importance of incorporating it into models of social collaboration <ref type="bibr" target="#b28">[29]</ref>. Thus, agents capable of mentalizing should not only emulate human-like inference of mental states but also assess the uncertainty of these inferences.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Method</head><p>Following the standard modeling pipeline in computational rationality <ref type="bibr" target="#b25">[26]</ref>, we formalize the task environment as a Markov Decision Process (MDP). It is represented as a tuple &lt; 𝑆, 𝐴, 𝑇 , 𝑅 &gt;, consisting state space 𝑆, action space 𝐴, transition probabilities 𝑇 and reward function 𝑅. A state 𝑠 ∈ 𝑆 encoding current information of the environment, transfers to next state 𝑠 ′ ∈ 𝑆 by performing an action 𝑎 ∈ 𝐴 according to transition probability 𝑇 (𝑠, 𝑎, 𝑠 ′ ) = 𝑃(𝑠 ′ |𝑠, 𝑎), and gains the reward 𝑟 = 𝑅(𝑠, 𝑎). Reinforcement learning (RL) solves the optimization problem of how to choosing the action 𝑎 through policy 𝜋(𝑎) = 𝑃(𝑎|𝑠) that maximizes the expected reward by interacting with the environment and learning from experience. The learning process can be expressed as the function</p><formula xml:id="formula_0">𝑉 𝜋 * (𝑠) = max 𝑎 [𝑅(𝑠, 𝑎) + 𝛾 ∑ 𝑠 ′ ∈𝑆 𝑇 (𝑠, 𝑎, 𝑠 ′ )𝑉 𝜋 * (𝑠 ′ )],</formula><p>where 𝑉 𝜋 * (𝑠) is the value of a state 𝑠 ∈ 𝑆 under an optimal policy 𝜋 * , discounting future rewards using 𝛾 ∈ [0, 1]. This optimality assumption ties in with computational rationality. Importantly, it is possible to implement bounds in the MDP formalism, forcing a bounded optimal behavior to emerge.</p><p>The bounded optimal agent described via an MDP can be parametrized. For instance, a parameter can govern its preferences, that is, the state rewards. This permits mentalizing: given observed data, what parameters best produce predicted data that fits the observations? To this end, we utilize Bayesian inference, described by Bayes' rule:</p><formula xml:id="formula_1">𝑃(𝜃|𝑥) = 𝑃(𝑥|𝜃)𝑃(𝜃) 𝑃(𝑥) ,</formula><p>where 𝜃 represents the latent factors to be inferred, and 𝑥 represents observed data. The inference uses a prior 𝑃(𝜃) and a likelihood 𝑃(𝑥|𝜃) to calculate posterior probability 𝑃(𝜃|𝑥), normalized with marginal likelihood 𝑃(𝑥). However, the intractability of the likelihood 𝑃(𝑥|𝜃) prevents us from deriving the posterior directly. This can be overcome with approximation and likelihood free inference methods <ref type="bibr" target="#b29">[30]</ref>, such as Bayesian Optimization for Likelihood-Free Inference (BOLFI) <ref type="bibr" target="#b30">[31]</ref>. Figure <ref type="figure" target="#fig_1">2</ref> illustrates the information flow in our model. Prior knowledge and observation data serve as inputs of an inference module, which parameterizes a RL agent. The agent then learns a bounded optimal policy within a simulator modeling the observed real-world task. Through multiple samplings, plausibility for various parameter values is evaluated, forming a posterior distribution that serves as the prior for subsequent inference with new observation data. This framework facilitates the temporal integration of inferences and allows for uncertainty analysis within the posterior probability distribution. All model details are available at the model's code repository (https://version.helsinki.fi/shanz/quantifyinguncertainty-in-mtom.git). </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Evaluation</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.">Participants</head><p>We recruited 𝑁 = 10 participants via the Prolific online platform. The number of participants was small, but because our experiment setup was well defined, we expected them to have a high agreement with each other. This was the case, meaning that a larger number of participants would likely not have changed the results. Their mean age was 35.6, and age range 23-56. They were required to be fluent in English, and be on a PC (no mobile devices were allowed).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.">Materials</head><p>The experiment consisted of eight distinct tasks, each including five stimulus images. One image shows a trajectory of a robot on a grid from a birds-eye perspective. The robot is moving from its starting position to either a blue or red circle, representing charging stations. There may also be walls, and the robot must navigate around them. Each picture is different, and there were a total of 8 ⋅ 5 = 40 stimuli. An example task is shown in Figure <ref type="figure" target="#fig_2">3</ref>. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3.">Experiment Procedure</head><p>Participants were tasked with discerning the preferred charging station of a specific task's robot, understanding that while the robot could charge at either, it had a latent preference for one. Instructed that the robot also aimed to conserve energy, possibly choosing a less favored station if it were closer, participants rated the likelihood of the robot's preference for each station on a scale from 1 (very unlikely) to 5 (very likely). After making their likelihood assessment for the stations, they were presented with the next stimulus, with instructions to refine their inferences based on all previously shown images of the present task. Only one image was shown at any single time. Upon the task changing after five stimuli, participants were reminded that a new robot with different preferences was introduced.</p><p>For our model, we represented the tasks within a grid world that the RL agent needed to navigate. It incurred a minor negative penalty for movement and obtained positive rewards from both charging stations, determined by two specific parameters. The objective was to infer these parameters based on the observed data. We measured the discrepancy between observed and generated trajectories using Jaccard similarity. Essentially, our inference engine recreated the world as depicted in the stimulus, then ran the RL agent across varying parameters, comparing the generated trajectory against the observed one to form a posterior distribution for the two preferences. Preference likelihood ratings for the model were derived by computing the mean of the posterior distribution for preferences associated with both the blue and red charging stations.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.4.">Results</head><p>The preference ratings of each response were first standardized so that they sum up to 1. Then, a mean rating for each stimulus in each task was computed. The model's ratings were likewise standardized to sum up to 1, allowing comparison between human and model inferences. This comparison is shown in Figure <ref type="figure" target="#fig_3">4</ref>. For calculating model fit, we selected only the inferences of the other color, because their values are inversions of each other after standadization. The model achieves a good fit, 𝑅 2 = 0.78, 𝑅𝑀𝑆𝐸 = 0.1. The most salient discrepancy between the model and human inferences is that the model is more careful in its estimates. Importantly, these results were obtained without any parameter tuning, meaning the model was not fit to the human data, but emerged similar data due to strong psychological assumptions about theory of mind.</p><p>The results exhibit the expected patterns of inference. Initially, participants faced uncertainty due to the limited evidence available. As they were exposed to additional stimuli, their inferences regarding the robot's preferences became more definite: one station's likelihood ratings increased, while the other's decreased. Task 1 serves as an example of this (Figure <ref type="figure" target="#fig_2">3</ref>): the participants' inference that the robot prefers the red station gets stronger with each stimulus image shown. However, in tasks 3, 4, 6, 7, and 8, early stimuli suggested a certain preference, but subsequent stimuli revealed a stronger preference for the alternate station. This is similar to our motivating example in Figure <ref type="figure" target="#fig_0">1</ref>. In these instances, the inferred preference for the more favored station shifted as the task progressed. Task 6 is an example of this (Figure <ref type="figure" target="#fig_4">5</ref>): the participants are shown that the robot selects the red station, but it is always closer than the blue one, so there is uncertainty. Finally, in stimulus 5, it is revealed that the robot in fact prefers the blue station.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.5.">Discussion</head><p>Human-AI alignment necessitates that both humans and intelligent machines accurately interpret each other's inten-  tions and actions <ref type="bibr" target="#b4">[5]</ref>. This paper introduces a human-like theory of mind model capable of temporal observation integration, while being sensitive to uncertainty inherent in mentalizing. We validated the model's human-like inference capabilities through a grid world task focused on preference determination between two goals. The work carried here is theoretical in nature, and future studies should focus on more complex scenarios. While computational rationality has effectively modeled complex behaviors, such as multitasking while driving <ref type="bibr" target="#b31">[32]</ref> and touchscreen typing <ref type="bibr" target="#b32">[33]</ref>, the exploration of long-term parameter inference in such contexts remains to be done.</p><p>Exploring decision-making under uncertainty is a large research topic. In our experiments, both humans and the model engaged in inferences and explicitly evaluated uncertainty, but they were not required to act on these inferences. A scenario where the model assists the observed actor will introduce the question of how to integrate uncertainty into decision-making. Taking the example of Janice from Figure <ref type="figure" target="#fig_0">1</ref>, if adjusting the positions of the coffee and tea jars could aid her, the decision to do so necessitates careful consideration of potential consequences, ensuring the action truly benefits rather than hinders her. The manner in which a decision-making algorithm accounts for uncertainty during collaborative efforts is impacts the helpfulness of interventions and carries a risk of unintended obstruction.</p><p>All code, materials, and data are published online (https://version.helsinki.fi/shanz/quantifying-uncertaintyin-mtom.git) to facilitate open science.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Inferences of preferences based on observed behavior contain uncertainty, especially when there are confounding factors such as effort. As more evidence accumulates, certainty increases.</figDesc><graphic coords="2,105.84,65.61,383.59,219.90" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: The overall structure of the model. It consists of simulation of external world and inference module, which can be repeated as new observation comes.</figDesc><graphic coords="3,77.31,65.61,203.07,110.67" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: The five stimuli shown sequentially to the participants, Task 1. Stimulus numbers are added here, and were not present in the experiment.</figDesc><graphic coords="3,77.31,489.15,203.07,137.55" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Figure 4 :</head><label>4</label><figDesc>Figure 4: Comparison of model and human inferences across eight tasks. As more evidence accumulates, the inferences become more certain. Values close to 0.5 indicate high uncertainty, and values close to either 0 or 1 high certainty.</figDesc><graphic coords="4,94.57,65.61,406.16,152.24" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_4"><head>Figure 5 :</head><label>5</label><figDesc>Figure 5: In Task 6, the participants only learned the true preference in the final image.</figDesc><graphic coords="4,77.31,266.19,203.07,136.96" type="bitmap" /></figure>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>This research has been supported by the Academy of Finland (grant 330347).</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Theory of mind and peer cooperation in two play contexts</title>
		<author>
			<persName><forename type="first">E</forename><surname>Etel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Slaughter</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Applied Developmental Psychology</title>
		<imprint>
			<biblScope unit="volume">60</biblScope>
			<biblScope unit="page" from="87" to="95" />
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Adult theory of mind, cooperation, machiavellianism: The effect of mindreading on social relations</title>
		<author>
			<persName><forename type="first">T</forename><surname>Paal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Bereczkei</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Personality and individual differences</title>
		<imprint>
			<biblScope unit="volume">43</biblScope>
			<biblScope unit="page" from="541" to="551" />
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">The social shapes test: A new measure of social intelligence, mentalizing, and theory of mind</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">I</forename><surname>Brown</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Ratajska</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">L</forename><surname>Hughes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">B</forename><surname>Fishman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Huerta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">F</forename><surname>Chabris</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Personality and Individual Differences</title>
		<imprint>
			<biblScope unit="volume">143</biblScope>
			<biblScope unit="page" from="107" to="117" />
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title/>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">F</forename><surname>Kihlstrom</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Cantor</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Social intelligence</title>
		<imprint>
			<date type="published" when="2000">2000</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<monogr>
		<title level="m" type="main">Human compatible: Artificial intelligence and the problem of control</title>
		<author>
			<persName><forename type="first">S</forename><surname>Russell</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2019">2019</date>
			<publisher>Penguin</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Building machines that learn and think like people</title>
		<author>
			<persName><forename type="first">B</forename><forename type="middle">M</forename><surname>Lake</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">D</forename><surname>Ullman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">B</forename><surname>Tenenbaum</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">J</forename><surname>Gershman</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Behavioral and brain sciences</title>
		<imprint>
			<biblScope unit="volume">40</biblScope>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Reasoning about mental states under uncertainty</title>
		<author>
			<persName><forename type="first">I</forename><surname>Cho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Kamkar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Hosseini-Kamkar</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">PloS one</title>
		<imprint>
			<biblScope unit="volume">17</biblScope>
			<biblScope unit="page">e0277356</biblScope>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Resolving uncertainty in a social world</title>
		<author>
			<persName><forename type="first">O</forename><surname>Feldmanhall</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Shenhav</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Nature human behaviour</title>
		<imprint>
			<biblScope unit="volume">3</biblScope>
			<biblScope unit="page" from="426" to="435" />
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Rational quantitative attribution of beliefs, desires and percepts in human mentalizing</title>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">L</forename><surname>Baker</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Jara-Ettinger</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Saxe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">B</forename><surname>Tenenbaum</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Nature Human Behaviour</title>
		<imprint>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="1" to="10" />
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Ten-month-old infants infer the value of goals from the costs of actions</title>
		<author>
			<persName><forename type="first">S</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">D</forename><surname>Ullman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">B</forename><surname>Tenenbaum</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">S</forename><surname>Spelke</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Science</title>
		<imprint>
			<biblScope unit="volume">358</biblScope>
			<biblScope unit="page" from="1038" to="1041" />
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Development of the social brain from age three to twelve years</title>
		<author>
			<persName><forename type="first">H</forename><surname>Richardson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Lisandrelli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Riobueno-Naylor</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Saxe</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Nature communications</title>
		<imprint>
			<biblScope unit="volume">9</biblScope>
			<biblScope unit="page" from="1" to="12" />
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Introducing masc: a movie for the assessment of social cognition</title>
		<author>
			<persName><forename type="first">I</forename><surname>Dziobek</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Fleck</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Kalbe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Rogers</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Hassenstab</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Brand</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Kessler</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">K</forename><surname>Woike</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><forename type="middle">T</forename><surname>Wolf</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Convit</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of autism and developmental disorders</title>
		<imprint>
			<biblScope unit="volume">36</biblScope>
			<biblScope unit="page" from="623" to="636" />
			<date type="published" when="2006">2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">It is not all about you: Communicative cooperation is determined by your partner&apos;s theory of mind abilities as well as your own</title>
		<author>
			<persName><forename type="first">R</forename><surname>Markiewicz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Rahman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Apperly</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Mazaheri</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Segaert</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Experimental Psychology: Learning, Memory, and Cognition</title>
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Modeling agents with a theory of mind</title>
		<author>
			<persName><forename type="first">M</forename><surname>Harbers</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Van Den</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J.-J</forename><surname>Bosch</surname></persName>
		</author>
		<author>
			<persName><surname>Meyer</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IEEE/WIC/ACM International Joint Conference on Web Intelligence and Intelligent Agent Technology</title>
				<imprint>
			<publisher>IEEE</publisher>
			<date type="published" when="2009">2009. 2009</date>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="page" from="217" to="224" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">An implemented theory of mind to improve human-robot shared plans execution</title>
		<author>
			<persName><forename type="first">S</forename><surname>Devin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Alami</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">11th ACM/IEEE International Conference on Human-Robot Interaction (HRI), IEEE</title>
				<imprint>
			<date type="published" when="2016">2016. 2016</date>
			<biblScope unit="page" from="319" to="326" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Towards a simple robotic theory of mind</title>
		<author>
			<persName><forename type="first">K.-J</forename><surname>Kim</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Lipson</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 9th workshop on performance metrics for intelligent systems</title>
				<meeting>the 9th workshop on performance metrics for intelligent systems</meeting>
		<imprint>
			<date type="published" when="2009">2009</date>
			<biblScope unit="page" from="131" to="138" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Reflexively mindblind: Using theory of mind to interpret behavior requires effortful attention</title>
		<author>
			<persName><forename type="first">S</forename><surname>Lin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Keysar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Epley</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Experimental Social Psychology</title>
		<imprint>
			<biblScope unit="volume">46</biblScope>
			<biblScope unit="page" from="551" to="556" />
			<date type="published" when="2010">2010</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Towards mutual theory of mind in human-ai interaction: How language reflects what students perceive about a virtual teaching assistant</title>
		<author>
			<persName><forename type="first">Q</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Saha</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Gregori</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Joyner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Goel</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2021 CHI conference on human factors in computing systems</title>
				<meeting>the 2021 CHI conference on human factors in computing systems</meeting>
		<imprint>
			<date type="published" when="2021">2021</date>
			<biblScope unit="page" from="1" to="14" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<monogr>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">M</forename><surname>Hiatt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">M</forename><surname>Harrison</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">G</forename><surname>Trafton</surname></persName>
		</author>
		<title level="m">Accommodating human variability in human-robot teams through theory of mind</title>
				<imprint>
			<date type="published" when="2011">2011</date>
		</imprint>
	</monogr>
	<note>Twenty-second international joint conference on artificial intelligence</note>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">The naive utility calculus as a unified, quantitative framework for action understanding</title>
		<author>
			<persName><forename type="first">J</forename><surname>Jara-Ettinger</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">E</forename><surname>Schulz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">B</forename><surname>Tenenbaum</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Cognitive Psychology</title>
		<imprint>
			<biblScope unit="volume">123</biblScope>
			<biblScope unit="page">101334</biblScope>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">Learning from others: The consequences of psychological reasoning for human learning</title>
		<author>
			<persName><forename type="first">P</forename><surname>Shafto</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><forename type="middle">D</forename><surname>Goodman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">C</forename><surname>Frank</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Perspectives on Psychological Science</title>
		<imprint>
			<biblScope unit="volume">7</biblScope>
			<biblScope unit="page" from="341" to="351" />
			<date type="published" when="2012">2012</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">Bayesian parameter inference for cognitive simulators</title>
		<author>
			<persName><forename type="first">Remes</forename><surname>Jokinen</surname></persName>
		</author>
		<author>
			<persName><surname>Kujala</surname></persName>
		</author>
		<author>
			<persName><surname>Corander</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Bayesian methods for interaction design</title>
				<editor>
			<persName><forename type="first">J</forename><surname>Williamson</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Oulasvirta</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Kristensson</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Banovic</surname></persName>
		</editor>
		<imprint>
			<publisher>Cambridge University Press</publisher>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">Action understanding as inverse planning</title>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">L</forename><surname>Baker</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Saxe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">B</forename><surname>Tenenbaum</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Cognition</title>
		<imprint>
			<biblScope unit="volume">113</biblScope>
			<biblScope unit="page" from="329" to="349" />
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">Parameter inference for computational cognitive models with approximate bayesian computation</title>
		<author>
			<persName><forename type="first">A</forename><surname>Kangasrääsiö</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">P</forename><surname>Jokinen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Oulasvirta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Howes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Kaski</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Cognitive science</title>
		<imprint>
			<biblScope unit="volume">43</biblScope>
			<biblScope unit="page">e12738</biblScope>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<analytic>
		<title level="a" type="main">Computational rationality: Linking mechanism and behavior through bounded utility maximization</title>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">L</forename><surname>Lewis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Howes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Singh</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Topics in cognitive science</title>
		<imprint>
			<biblScope unit="volume">6</biblScope>
			<biblScope unit="page" from="279" to="311" />
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<analytic>
		<title level="a" type="main">Computational rationality as a theory of interaction</title>
		<author>
			<persName><forename type="first">A</forename><surname>Oulasvirta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">P</forename><surname>Jokinen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Howes</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2022 CHI Conference on Human Factors in Computing Systems</title>
				<meeting>the 2022 CHI Conference on Human Factors in Computing Systems</meeting>
		<imprint>
			<date type="published" when="2022">2022</date>
			<biblScope unit="page" from="1" to="14" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<analytic>
		<title level="a" type="main">Towards machines that understand people</title>
		<author>
			<persName><forename type="first">A</forename><surname>Howes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">P</forename><surname>Jokinen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Oulasvirta</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">AI Magazine</title>
		<imprint>
			<biblScope unit="volume">44</biblScope>
			<biblScope unit="page" from="312" to="327" />
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b27">
	<analytic>
		<title level="a" type="main">Making predictions in a changing world-inference, uncertainty, and learning</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">X</forename><surname>O'reilly</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Frontiers in neuroscience</title>
		<imprint>
			<biblScope unit="volume">7</biblScope>
			<biblScope unit="page">105</biblScope>
			<date type="published" when="2013">2013</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b28">
	<analytic>
		<title level="a" type="main">The computational challenge of social learning</title>
		<author>
			<persName><forename type="first">O</forename><surname>Feldmanhall</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">R</forename><surname>Nassar</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Trends in Cognitive Sciences</title>
		<imprint>
			<biblScope unit="volume">25</biblScope>
			<biblScope unit="page" from="1045" to="1057" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b29">
	<analytic>
		<title level="a" type="main">Bayesian optimization for likelihood-free inference of simulator-based statistical models</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">U</forename><surname>Gutmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Corander</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Machine Learning Research</title>
		<imprint>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b30">
	<analytic>
		<title level="a" type="main">Elfi: Engine for likelihood-free inference</title>
		<author>
			<persName><forename type="first">J</forename><surname>Lintusaari</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Vuollekoski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Kangasrääsiö</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Skytén</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Järvenpää</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Marttinen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">U</forename><surname>Gutmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Vehtari</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Corander</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Kaski</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Machine Learning Research</title>
		<imprint>
			<biblScope unit="volume">19</biblScope>
			<biblScope unit="page" from="1" to="7" />
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b31">
	<analytic>
		<title level="a" type="main">Multitasking in driving as optimal adaptation under uncertainty</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">P</forename><surname>Jokinen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Kujala</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Oulasvirta</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Human factors</title>
		<imprint>
			<biblScope unit="volume">63</biblScope>
			<biblScope unit="page" from="1324" to="1341" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b32">
	<analytic>
		<title level="a" type="main">Touchscreen typing as optimal supervisory control</title>
		<author>
			<persName><forename type="first">J</forename><surname>Jokinen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Acharya</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Uzair</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Jiang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Oulasvirta</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems</title>
				<meeting>the 2021 CHI Conference on Human Factors in Computing Systems</meeting>
		<imprint>
			<date type="published" when="2021">2021</date>
			<biblScope unit="page" from="1" to="14" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
