<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">A Semantic Representation of Pedestrian Crossing Behavior</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">He</forename><surname>Tan</surname></persName>
							<affiliation key="aff0">
								<orgName type="department" key="dep1">Department of Computing</orgName>
								<orgName type="department" key="dep2">School of Engineering</orgName>
								<orgName type="institution">Jönköping University</orgName>
								<address>
									<country key="SE">Sweden</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Florian</forename><surname>Westphal</surname></persName>
							<affiliation key="aff0">
								<orgName type="department" key="dep1">Department of Computing</orgName>
								<orgName type="department" key="dep2">School of Engineering</orgName>
								<orgName type="institution">Jönköping University</orgName>
								<address>
									<country key="SE">Sweden</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">A Semantic Representation of Pedestrian Crossing Behavior</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">FA90465E4A1BC644AE36B81BBBC40D1A</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T19:29+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Ontology</term>
					<term>Knowledge Graph Construction from trajectory data</term>
					<term>Visual Question Answering</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>In this paper, we focus on the crucial task of understanding and modeling pedestrian behavior, which is essential for numerous applications. We introduce a semantic representation of pedestrian crossing behavior. The representation is to capture the sub-events within a behavior and the spatial-temporal evolution of interactions between pedestrians and other objects involved in crossing events. We demonstrate its practical application by utilizing it to analyze pedestrian crossing behavior from road user movement data (i.e. trajectories). By constructing a knowledge graph from detailed road user dynamics data using this representation, we enable queries that address safety concerns related to pedestrian crossing behavior, aiding traffic engineers in their work on urban traffic infrastructure design.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Vulnerable road users, such as children, the elderly, and disabled individuals, are integral to the dynamics of city traffic. They play essential roles in establishing a sustainable, active, and inclusive mobility environment. Therefore, understanding and modeling pedestrian behavior is fundamental for many applications, including traffic flow analysis, traffic safety improvement, urban planning, and intelligent driving systems. Pedestrian crossing behavior is one of the main aspects of pedestrian behavior that has been studied in numerous research studies <ref type="bibr" target="#b0">[1]</ref>. It involves the actions and movements of pedestrians while crossing streets or roadways, often guided by traffic signals, road markings, and traffic conditions.</p><p>Traditionally, stochastic, linear regression, and discrete choice models are used to build an understanding of how pedestrians make crossing decisions considering various factors related to people, roadway, traffic, traffic controls and traffic rules <ref type="bibr" target="#b0">[1,</ref><ref type="bibr" target="#b1">2]</ref>. Parameters of the models are estimated from survey and/or questionnaire data or manually screened video recordings. More recently, agent-based modeling has been used to model road users as intelligent agents attempting to make rational decisions in uncertain and complex situations. However, most work has focused on modeling vehicle behaviors. Very limited studies dedicated to develop models for other road users such as pedestrians <ref type="bibr" target="#b2">[3]</ref>. Often, these studies focus on pedestrian-vehicle conflicts and model pedestrians' collision avoidance mechanism <ref type="bibr" target="#b3">[4]</ref>. The parameters of the models are typically calibrated by detecting and tracking road users from video data or using results from literature.</p><p>At the same time, researchers across diverse disciplines, such as computer vision, artificial intelligence (AI), cognitive science, and neuroscience, have conducted numerous studies on understanding human activities. Depending on their complexity, human activities can be classified into different levels: gestures, actions, interactions, and group activities <ref type="bibr" target="#b4">[5]</ref>. This paper specifically focuses on understanding pedestrian crossing behavior on the level of human-object interactions. Human activity is a spatial-temporal evolution of interactions <ref type="bibr" target="#b5">[6]</ref>. Here, we present a semantic representation of pedestrian crossing behavior, capturing sub-events of a behavior and their temporal and spatial structures. Previous studies within computer vision have suggested that a structured spatial-temporal representation can lead to more accurate activity understanding and improve the performance of various computer vision tasks, including image captioning and visual question answering <ref type="bibr" target="#b6">[7]</ref>. Studies (e.g., <ref type="bibr" target="#b7">[8,</ref><ref type="bibr" target="#b8">9]</ref>) in the AI area indicate that such a representation can cope with less training data by incorporating prior knowledge and help to understand human activities.</p><p>In this paper we present the structured spatial-temporal representation of pedestrian crossing behavior, and describe its application to gain an understanding of pedestrian crossing behavior from recorded road user dynamics data. Utilizing the representation, a knowledge graph is constructed from road user dynamics data. The queries over the knowledge graph can answer safety related questions on pedestrian crossing behavior for traffic engineers and help with their work on urban traffic infrastructure design.</p><p>The remainder of this paper is organized as follows: In Section 2, we introduce the methods that have been employed to semantically represent human activities, particularly pedestrian behaviors. In Section 3, we present our approach to semantically representing crossing behaviors. Section 4 outlines the utilization of semantic representation within the context of traffic data analysis for traffic engineers. Finally, the paper concludes the work in Section 5.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Related Work</head><p>Pedestrian behavior has been widely analyzed in various research works using a plethora of methods. Nevertheless, understanding pedestrian behavior remains challenging due to the inherent complexity of human activities. Despite the diverse analysis methods used, the significance of semantic representation in understanding pedestrian behavior has often been overlooked. Only a limited number of studies have explored the semantic representation for pedestrian behavior.</p><p>Chai et al. <ref type="bibr" target="#b9">[10]</ref> utilized fuzzy logic to model the cognitions and behavioral patterns of pedestrians, in order to understand the effect of age and gender when pedestrians are crossing a signalized crosswalk and jaywalking. Gharebaghi et al. <ref type="bibr" target="#b10">[11]</ref> developed a mobility ontology for people with motor disabilities (PWMD). Specifically, it considers the interactions between people and both the social and physical environment. The ontology was used to support the development of assistive technologies for the mobility of PWMD. Fang et al. <ref type="bibr" target="#b11">[12]</ref> developed an ontology defining various kinds of road users, including pedestrians, and describing their relationships. The concepts from the ontology are used to define the rules for describing the interactions between road users and to support rule-based reasoning for predicting road users' behavior.</p><p>In this paper, we present a semantic representation of pedestrian crossing behavior. The representation describes the dynamic evolution of interactions between pedestrians and objects within the physical environment over time, capturing interactions in both spatial and temporal dimensions.</p><p>In 1970, Hägerstrand <ref type="bibr" target="#b12">[13]</ref> introduced the concept of a time-space path in understanding human activities. This theory has laid the groundwork for trajectories that have been shown to be useful in representing people's movements. Inspired by Hägerstrand's work, Orellana and Renso <ref type="bibr" target="#b13">[14]</ref> developed an interaction ontology. The ontology conceptualizes the characteristics of pedestrian movement behaviour. It has focused on identifying various movement patterns from time-space paths, and the different categories of interactions, spatial and temporal contexts, behavior, and the high-level relations between these concepts. Logic-based reasoning is used to categorize pedestrian movement behavior based on its movement patterns, interactions, and contexts.</p><p>Meanwhile, in cognitive science and neuroscience, it has been recognized that segmentation is a fundamental component of perception, playing a critical role in understanding activities. People tend to perceive ongoing continuous activity as series of discrete events (or called segments) <ref type="bibr" target="#b14">[15,</ref><ref type="bibr" target="#b15">16,</ref><ref type="bibr" target="#b16">17]</ref>. The relationships between segments are encoded in partonomic hierarchies <ref type="bibr" target="#b17">[18]</ref>. Coarse segmentation is often related to objects' locations and their goals, and the causal relations between their actions. Fine segmentation is closely linked to changes in the interactions between objects <ref type="bibr" target="#b18">[19]</ref>. Building on these findings in cognitive science and neuroscience, Ji et al. <ref type="bibr" target="#b6">[7]</ref> proposed a spatial-temporal scene graph to represent human activity and to improve the performance of action recognition and few-shot action recognition using neural networks. Mlodzian et al. <ref type="bibr" target="#b8">[9]</ref> presented an ontology that was tailored for representing entities and their spatial and temporal relations in traffic scenes in the nuScenes dataset <ref type="foot" target="#foot_0">1</ref> . A knowledge graph was constructed from the nuScenes dataset using the ontology and provided as a benchmark dataset for developing advanced trajectory prediction models.</p><p>In this paper, drawing from these insights in cognitive science, neuroscience and computer vision, we propose a structured spatial-temporal representation for pedestrian crossing behavior and present its application to gain an understanding of pedestrian crossing behavior from road user movement data.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Semantic Representation</head><p>In this section we present the semantic representation for pedestrian crossing behavior. A pedestrian crossing behavior can be seen as a dynamic evolution of interactions between pedestrians and objects within the physical environment over time. Every crossing behavior can be broken down into segments, each representing a distinct phase of the behavior. These segments capture the changes of the interactions between pedestrians and objects in both physical and temporal dimensions, and together represent the pedestrian crossing behavior. For example, Fig. <ref type="figure" target="#fig_1">1</ref> shows a crossing event, which is extracted from road user behavior measurement performed at a zebra-free crossing in Lindholmen, Gothenburg in Sweden. Each triple follows the format ((id, object_1), spatial_relation, (id, object_2)), where the object 1 is a moving object such as pedestrian, cyclist, and vehicle, and the object 2 can be a moving object or a static object such as crossing, area and sidewalk, and id is the unique identifier for each object.   related to regions in an image in computer vision, the term frame is used instead. In computer vision, a video can be divided into a sequence of frames. Each frame represents a single still image in the video sequence. The blue arrow represents subclass relations between concepts. Currently, the ontology includes only a limited number of categories for both moving and static objects. However, additional categories will be integrated as the ontology continues to undergo further development.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Usage of the Representation</head><p>In this section, we describe an application of the semantic representation of pedestrian crossing behavior mentioned beforehand in Section 3. The application aims to provide information support for traffic engineers during traffic infrastructure planning and development, with a particular focus on pedestrian safety. In the application, pedestrian crossing behaviors are described using the semantic representation, and a knowledge graph is constructed for these behaviors. Subsequently, a number of queries serve as question-answering tools to provide information for traffic engineers. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Crossing Behavior Dataset</head><p>The crossing behavior dataset is prepared from the traffic measurement aforementioned in Section 3. Fig <ref type="figure" target="#fig_5">4</ref> shows an example frame extracted from the dataset. The road user positions and trajectories are displayed in a camera view, overlaid on the anonymized video frame. The measurement is performed by Viscando AB<ref type="foot" target="#foot_2">3</ref> using the 3D&amp;AI based infrastructure sensor OTUS3D. The total period of the measurment is 11 hours and 5 minutes. The data contains trajectories of all road users recorded 10 times per second. Trajectories contain the unique track ID for each object, the UTC time stamp, position (i.e. X-coordinate and Y-coordinate), velocity (i.e. object speed in the direction of motion (km/h)) and object type. Currently, the object types include pedestrian, cyclist, light vehicle and heavy vehicle. Vision data are processed in the embedded computational unit and removed within 20 ms from being captured. Thus, the dataset is stored fully anonymously, ensuring compliance with the General Data Protection Regulation (GDPR) of the European Union<ref type="foot" target="#foot_3">4</ref> , because personal information is neither stored in the sensors nor transmitted. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Knowledge Graph Construction</head><p>In this section we describe the construction of the knowledge graph that describes the pedestrian crossing behaviors recorded in the aforementioned dataset. Since the application is to support the traffic infrastructure planning and development prioritizing pedestrian safety, the construction has focused on the crossing events involving pedestrians/cyclists but also vehicle(s). The spatial relationship between objects was calculated based on the physical distance between them. The current spatial relationships include the ones between moving objects, i.e. close_to and far_away, and the ones between a moving object and a static object, i.e., left_close_to, right_close_to, left_far_away, right_close_to, on, out_of_area. If the x-coordinate of one object is smaller than that of another, the former is positioned to the left of the latter; otherwise, it is positioned to the right.</p><p>When the information was extracted from the aforementioned dataset, the ontology described in Section 3 was populated, and the knowledge graph is set up. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Question Answering</head><p>In this section, we present the SPARQL queries to retrieve answers from the knowledge graph or to inquire the information from it to formulate responses to a few example questions that traffic engineers might pose.</p><p>First, two prefixes are predefined for the following SPARQL queries, i.e., tsdata (http://www. example.com/ontology/traffic_scene_kg# and ts (http://www.example.com/ontology/traffic_ scene_ontology.owl#).</p><p>Example 1: describe a crossing behavior. The query will return an RDF dataset describing a specific crossing behavior. Fig <ref type="figure" target="#fig_7">5</ref> shows a visualization of such an RDF dataset. It was generated by using the Stardog Studio visualization tool <ref type="foot" target="#foot_4">5</ref> . Such an RDF dataset can also be converted into text, allowing traffic engineers to easily access and understand the information <ref type="bibr" target="#b19">[20]</ref>. Example 2: find and describe the crossing behaviors within a specified time period. The query will return an RDF dataset containing the crossing behaviors within the specified time period. For each behavior, using the query given in Example 1, it can be described in text, allowing traffic engineers to access and understand the information. Example 4: find the crossing events where pedestrians/cyclists are close to vehicles and their speed is too fast. Such behaviors are considered unsafe. The query is an extension of the one given in Example 3, with the addition of the following triple patterns and filter.</p><p>?i ts:hasObject1Info ?obj1info . ?i ts:hasObject2Info ?obj2info. ?obj1info ts:speed ?s1. ?obj2info ts:speed ?s2 FILTER (?s1 &gt;= highest_safe_speed || ?s2 &gt;= highest_safe_speed )</p><p>Example 5: find the crossing behaviors where pedestrians take a shortcut to the crossing, specifically by crossing diagonally across the street. Such a behavior is considered unsafe. This query is separated into two steps. The first step is retrieving the crossing events and the frames where pedestrians are involved. In the second step, the y-coordinates of the pedestrians during the crossing are retrieved. If the changes of the y-coordinates exceed a certain threshold, the pedestrians are considered as being taking a shortcut to the crossing. As an example, the following query shows how to retrieve the y-coordinates of the pedestrian involved in the crossing event presented in Section 3. The queries over the knowledge graph are not limited to the ones listed in this paper. More complex queries can be constructed when traffic engineers require more intricate information. For instance, cyclists swinging out at a crossing are considered as unsafe behavior. Such behavior can be identified by using a number of queries in a simple program.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Conclusions</head><p>In this paper we have introduced a structured spatial-temporal representation of pedestrian crossing behavior and demonstrated its application in understanding such behavior from recorded road user dynamics data. By leveraging this representation, we construct a knowledge graph from the road user dynamics data. Queries made over this knowledge graph can address safety-related inquiries regarding pedestrian crossing behavior for traffic engineers, supporting them in urban traffic infrastructure design work.</p><p>In future work, we aim to enhance the ontology by incorporating more granular categories of road users and other spatial relations between objects. Additionally, we plan to develop a tool that enables traffic engineers to pose text-based questions and receive text-based answers, thereby enhancing their workflow support. This way of interacting with the road user dynamics data could be implemented with the help of large language models (LLMs) and retrieval-augmented generation (RAG) <ref type="bibr" target="#b20">[21]</ref>. In such a system, the user's question would be translated into a query against the knowledge graph and the returned information would be transformed into natural language text by the LLM.</p><p>Apart from querying the constructed knowledge graph to gain insights into the behavior of different traffic participants, the proposed semantic representation could also serve as base for trajectory prediction approaches. With increased interest in the development of self-driving cars, predicting the behavior of other traffic participants has come more into focus <ref type="bibr" target="#b21">[22]</ref>. For this task, it is important to understand the spatial relationships between different actors. Hence, different approaches have been investigated to integrate these relationships into trajectory prediction, including simple graph structures <ref type="bibr" target="#b22">[23]</ref>, heterogeneous graphs <ref type="bibr" target="#b23">[24]</ref>, and knowledge graphs <ref type="bibr" target="#b8">[9]</ref>. While clearly belonging to the latter category, our representation focuses particularly on static objects, such as road infrastructure elements, to capture their impact on trajectories of traffic participants. Presumably, this will not only improve trajectory predictions, but also help traffic engineers to understand the impact different road infrastructure elements will have on traffic. Therefore, another direction of the future work is to investigate the incorporation of the constructed knowledge graphs with graph neural networks for trajectory prediction.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head></head><label></label><figDesc>Fig 1-a displays the trajectories of the pedestrian and other moving objects involved in the event. The red trajectory represents a pedestrian, the blue trajectory represents a cyclist, and the cyan trajectory represents a light vehicle. Fig 1-b1 to b8 show a sequence of distinct segments that capture the changes in interactions between pedestrians and objects over time during the event. These interactions are expressed in a set of triples, as shown in Fig 2.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: An example of pedestrian crossing behavior.</figDesc><graphic coords="4,108.88,204.38,375.03,382.87" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Fig 3</head><label>3</label><figDesc>Fig 3 illustrates the current version of the ontology designed to represent the spatial-temporal evolution of crossing behavior. This ontology is accessible on GitHub 2 . Since segment is often</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: The interactions and their changes in the crossing behavior.</figDesc><graphic coords="5,89.29,84.19,416.69,274.93" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_4"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: The ontology representing temporal and spatial structures of the interactions in pedestrian crossing behavior.</figDesc><graphic coords="6,89.29,84.19,416.70,287.45" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_5"><head>Figure 4 :</head><label>4</label><figDesc>Figure 4: An example frame extracted from the dataset.</figDesc><graphic coords="7,192.22,84.19,208.35,154.61" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_6"><head></head><label></label><figDesc>Fig 5 shows the fragment of the knowledge graph that represents the pedestrian crossing behavior presented in Section 3.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_7"><head>Figure 5 :</head><label>5</label><figDesc>Figure 5: The fragment of the knowledge graph that represents a pedestrian crossing behavior.</figDesc><graphic coords="8,108.88,84.19,375.02,289.53" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head></head><label></label><figDesc>Example 3: find the crossing events where pedestrians/cyclists are close to vehicles and return the frames when this happens.</figDesc><table><row><cell>SELECT DISTINCT ?b ?f</cell></row><row><cell>WHERE {</cell></row><row><cell>?b rdf:type ts:Behavior .</cell></row><row><cell>?b ts:hasFrame ?f .</cell></row><row><cell>?f ts:containsInteraction ?i .</cell></row><row><cell>?i ts:hasSpatialRelationship ts:close_to .</cell></row><row><cell>{ ?i ts:hasObject1 ?obj1 .</cell></row><row><cell>{?obj1 rdf:type ts:Pedestrian}</cell></row><row><cell>UNION {?obj1 rdf:type ts:Bicyclist}.</cell></row><row><cell>?i ts:hasObject2 ?obj2 .</cell></row><row><cell>{?obj2 rdf:type ts:HeavyVehicle}</cell></row><row><cell>UNION {?obj2 rdf:type ts:LightVehicle}}</cell></row><row><cell>UNION</cell></row><row><cell>{?i ts:hasObject2 ?obj2 .</cell></row><row><cell>{?obj2 rdf:type ts:Pedestrian}</cell></row><row><cell>UNION {?obj2 rdf:type ts:Bicyclist}.</cell></row><row><cell>?i ts:hasObject1 ?obj1 .</cell></row><row><cell>{?obj1 rdf:type ts:HeavyVehicle}</cell></row><row><cell>UNION {?obj1 rdf:type ts:LightVehicle}}.</cell></row><row><cell>}</cell></row><row><cell>ORDER BY ?b</cell></row><row><cell>SELECT DISTINCT ?b</cell></row><row><cell>WHERE {</cell></row><row><cell>?b a ts:Behavior .</cell></row><row><cell>?b ts:hasFrame ?f.</cell></row><row><cell>?f ts:absoluteTime ?t.</cell></row><row><cell>FILTER (?t &gt;= "2019-05-17 08:00:00"^^xsd:dateTime</cell></row><row><cell>&amp;&amp; ?t &lt;= "2019-05-17 08:20:00"^^xsd:dateTime)</cell></row><row><cell>}</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">https://www.nuscenes.org/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_1">https://github.com/tanhe-git/crossing_behavior/blob/main/traffic_scene_ontology.owl</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_2"> www.viscando.com   </note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_3">https://gdpr-info.eu/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_4">https://cloud.stardog.com/</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>This work has been conducted in the project "Data and AI for decision Making suppOrt in traffic iNfrastructure Development (DAIMOND)" , which is funded by Vinnova (the Sweden's innovation agency) and AI Sweden (the Swedish national center for applied AI). The authors would like to thank the traffic department in Jönköping municipality for providing traffic safety related use cases and Viscando AB for providing traffic measurement dataset and expertise in traffic measurements and analysis.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">A critical assessment of pedestrian behaviour models</title>
		<author>
			<persName><forename type="first">E</forename><surname>Papadimitriou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Yannis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Golias</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Transportation research part F: traffic psychology and behaviour</title>
		<imprint>
			<biblScope unit="volume">12</biblScope>
			<biblScope unit="page" from="242" to="255" />
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Theoretical framework for modeling pedestrians&apos; crossing behavior along a trip</title>
		<author>
			<persName><forename type="first">E</forename><surname>Papadimitriou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Yannis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Golias</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of transportation engineering</title>
		<imprint>
			<biblScope unit="volume">136</biblScope>
			<biblScope unit="page" from="914" to="924" />
			<date type="published" when="2010">2010</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Modeling pedestrian behavior in pedestrian-vehicle near misses: A continuous Gaussian Process Inverse Reinforcement Learning (GP-IRL) approach</title>
		<author>
			<persName><forename type="first">P</forename><surname>Nasernejad</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Sayed</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Alsaleh</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Accident Analysis &amp; Prevention</title>
		<imprint>
			<biblScope unit="volume">161</biblScope>
			<biblScope unit="page">106355</biblScope>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Intend-wait-cross: Towards modeling realistic pedestrian crossing behavior</title>
		<author>
			<persName><forename type="first">A</forename><surname>Rasouli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Kotseruba</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IEEE Intelligent Vehicles Symposium (IV)</title>
				<imprint>
			<publisher>IEEE</publisher>
			<date type="published" when="2022">2022. 2022</date>
			<biblScope unit="page" from="83" to="90" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Human activity analysis: A review</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">K</forename><surname>Aggarwal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">S</forename><surname>Ryoo</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Acm Computing Surveys (Csur)</title>
		<imprint>
			<biblScope unit="volume">43</biblScope>
			<biblScope unit="page" from="1" to="43" />
			<date type="published" when="2011">2011</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Visual genome: Connecting language and vision using crowdsourced dense image annotations</title>
		<author>
			<persName><forename type="first">R</forename><surname>Krishna</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Zhu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Groth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Johnson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Hata</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Kravitz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Kalantidis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L.-J</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">A</forename><surname>Shamma</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">International journal of computer vision</title>
		<imprint>
			<biblScope unit="volume">123</biblScope>
			<biblScope unit="page" from="32" to="73" />
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Action genome: Actions as compositions of spatio-temporal scene graphs</title>
		<author>
			<persName><forename type="first">J</forename><surname>Ji</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Krishna</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Fei-Fei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">C</forename><surname>Niebles</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition</title>
				<meeting>the IEEE/CVF Conference on Computer Vision and Pattern Recognition</meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="10236" to="10247" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Transferring skills to humanoid robots by extracting semantic representations from observations of human activities</title>
		<author>
			<persName><forename type="first">K</forename><surname>Ramirez-Amaro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Beetz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Cheng</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Artificial Intelligence</title>
		<imprint>
			<biblScope unit="volume">247</biblScope>
			<biblScope unit="page" from="95" to="118" />
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">nuScenes Knowledge Graph -A Comprehensive Semantic Representation of Traffic Scenes for Trajectory Prediction</title>
		<author>
			<persName><forename type="first">L</forename><surname>Mlodzian</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Sun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Berkemeyer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Monka</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Dietze</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Halilaj</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Luettin</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV) Workshops</title>
				<meeting>the IEEE/CVF International Conference on Computer Vision (ICCV) Workshops</meeting>
		<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="42" to="52" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Fuzzy logic-based observation and evaluation of pedestrians&apos; behavioral patterns by age and gender</title>
		<author>
			<persName><forename type="first">C</forename><surname>Chai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Shi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><forename type="middle">D</forename><surname>Wong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">J</forename><surname>Er</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">T M</forename><surname>Gwee</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Transportation research part F: traffic psychology and behaviour</title>
		<imprint>
			<biblScope unit="volume">40</biblScope>
			<biblScope unit="page" from="104" to="118" />
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Integration of the social environment in a mobility ontology for people with motor disabilities</title>
		<author>
			<persName><forename type="first">A</forename><surname>Gharebaghi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M.-A</forename><surname>Mostafavi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Edwards</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Fougeyrollas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Gamache</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Grenier</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Disability and Rehabilitation: Assistive Technology</title>
		<imprint>
			<biblScope unit="volume">13</biblScope>
			<biblScope unit="page" from="540" to="551" />
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Ontology-based reasoning approach for long-term behavior prediction of road users</title>
		<author>
			<persName><forename type="first">F</forename><surname>Fang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Yamaguchi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Khiat</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IEEE Intelligent Transportation Systems Conference (ITSC), IEEE</title>
				<imprint>
			<date type="published" when="2019">2019. 2019</date>
			<biblScope unit="page" from="2068" to="2073" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">What about people in Regional Science?</title>
		<author>
			<persName><forename type="first">T</forename><surname>Hägerstrand</surname></persName>
		</author>
		<idno type="DOI">10.1007/bf01936872</idno>
		<ptr target="http://dx.doi.org/10.1007/bf01936872.doi:10.1007/bf01936872" />
	</analytic>
	<monogr>
		<title level="j">Papers of the Regional Science Association</title>
		<imprint>
			<biblScope unit="volume">24</biblScope>
			<biblScope unit="page" from="6" to="21" />
			<date type="published" when="1970">1970</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Developing an interactions ontology for characterising pedestrian movement behaviour</title>
		<author>
			<persName><forename type="first">D</forename><surname>Orellana</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Renso</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Movement-aware applications for sustainable mobility: Technologies and approaches</title>
				<imprint>
			<publisher>IGI Global</publisher>
			<date type="published" when="2010">2010</date>
			<biblScope unit="page" from="62" to="86" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Attribution and the unit of perception of ongoing behavior</title>
		<author>
			<persName><forename type="first">D</forename><surname>Newtson</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of personality and social psychology</title>
		<imprint>
			<biblScope unit="volume">28</biblScope>
			<biblScope unit="page">28</biblScope>
			<date type="published" when="1973">1973</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Planning, neuropsychology, and artificial intelligence: crossfertilization</title>
		<author>
			<persName><forename type="first">L</forename><surname>Spector</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Grafman</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Handbook of neuropsychology</title>
		<imprint>
			<biblScope unit="volume">9</biblScope>
			<biblScope unit="page" from="377" to="392" />
			<date type="published" when="1994">1994</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Discovering event structure in continuous narrative perception and memory</title>
		<author>
			<persName><forename type="first">C</forename><surname>Baldassano</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Zadbood</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">W</forename><surname>Pillow</surname></persName>
		</author>
		<author>
			<persName><forename type="first">U</forename><surname>Hasson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">A</forename><surname>Norman</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Neuron</title>
		<imprint>
			<biblScope unit="volume">95</biblScope>
			<biblScope unit="page" from="709" to="721" />
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Perceiving, remembering, and communicating structure in events</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">M</forename><surname>Zacks</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Tversky</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Iyer</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of experimental psychology: General</title>
		<imprint>
			<biblScope unit="volume">130</biblScope>
			<biblScope unit="page">29</biblScope>
			<date type="published" when="2001">2001</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Perceiving narrated events</title>
		<author>
			<persName><forename type="first">N</forename><forename type="middle">K</forename><surname>Speer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">M</forename><surname>Zacks</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">R</forename><surname>Reynolds</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Annual Meeting of the Cognitive Science Society</title>
				<meeting>the Annual Meeting of the Cognitive Science Society</meeting>
		<imprint>
			<date type="published" when="2004">2004</date>
			<biblScope unit="volume">26</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">The WebNLG challenge: Generating text from RDF data</title>
		<author>
			<persName><forename type="first">C</forename><surname>Gardent</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Shimorina</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Narayan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Perez-Beltrachini</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 10th International Conference on Natural Language Generation</title>
				<meeting>the 10th International Conference on Natural Language Generation</meeting>
		<imprint>
			<date type="published" when="2017">2017</date>
			<biblScope unit="page" from="124" to="133" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<monogr>
		<author>
			<persName><forename type="first">P</forename><surname>Zhao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Yu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Geng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Fu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Cui</surname></persName>
		</author>
		<idno>arXiv e-prints</idno>
		<ptr target="http://arxiv.org/abs/2402.19473v1.arXiv:2402.19473v1" />
		<title level="m">Retrievalaugmented generation for ai-generated content: A survey</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">A Survey on Trajectory-Prediction Methods for Autonomous Driving</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Du</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Zhou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Chen</surname></persName>
		</author>
		<idno type="DOI">10.1109/tiv.2022.3167103</idno>
		<ptr target="http://dx.doi.org/10.1109/tiv.2022.3167103.doi:10.1109/tiv.2022.3167103" />
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Intelligent Vehicles</title>
		<imprint>
			<biblScope unit="volume">7</biblScope>
			<biblScope unit="page" from="652" to="674" />
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">VectorNet: Encoding HD Maps and Agent Dynamics From Vectorized Representation</title>
		<author>
			<persName><forename type="first">J</forename><surname>Gao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Sun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Zhao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Shen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Anguelov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Schmid</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)</title>
				<meeting>the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)</meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">Holistic Graph-based Motion Prediction</title>
		<author>
			<persName><forename type="first">D</forename><surname>Grimm</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Schörner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Dreßler</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J.-M</forename><surname>Zöllner</surname></persName>
		</author>
		<idno type="DOI">10.1109/icra48891.2023.10161468</idno>
		<ptr target="http://dx.doi.org/10.1109/icra48891.2023.10161468.doi:10.1109/icra48891.2023.10161468" />
	</analytic>
	<monogr>
		<title level="m">IEEE International Conference on Robotics and Automation (ICRA)</title>
				<imprint>
			<publisher>IEEE</publisher>
			<date type="published" when="2023">2023. 2023</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
