<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Streaming and Learning the Personal Context</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Fausto</forename><surname>Giunchiglia</surname></persName>
							<email>fausto.giunchiglia@unitn.it</email>
							<affiliation key="aff0">
								<orgName type="institution">University of Trento</orgName>
								<address>
									<settlement>Trento</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Marcelo</forename><surname>Rodas Britez</surname></persName>
							<email>marcelo.rodasbritez@unitn.it</email>
							<affiliation key="aff0">
								<orgName type="institution">University of Trento</orgName>
								<address>
									<settlement>Trento</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Andrea</forename><surname>Bontempelli</surname></persName>
							<email>andrea.bontempelli@unitn.it</email>
							<affiliation key="aff0">
								<orgName type="institution">University of Trento</orgName>
								<address>
									<settlement>Trento</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Xiaoyue</forename><surname>Li</surname></persName>
							<email>xiaoyue.li@unitn.it</email>
							<affiliation key="aff0">
								<orgName type="institution">University of Trento</orgName>
								<address>
									<settlement>Trento</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Streaming and Learning the Personal Context</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">E309865108AC112544E1D2878D52B8D5</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T01:31+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>The representation of the personal context is complex and essential to improve the help machines can give to humans for making sense of the world, and the help humans can give to machines to improve their efficiency. We aim to design a novel model representation of the personal context and design a learning process for better integration with machine learning. We aim to implement these elements into a modern system architecture focus in real-life environments. Also, we show how our proposal can improve in specifically related work papers. Finally, we are moving forward with a better personal context representation with an improved model, the implementation of the learning process, and the architectural design of these components.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>Every person makes sense of their personal context differently because of their different sets of personal characteristics (intelligence) and behaviour (life choices). However, the machine's understanding of the personal context is radically different from the user's understanding. This limitation is due to the limited definition of the personal context, and the lack of tools to make sense of the personal context. For instance, while the person you are with now can be linked to a name, for people it has more meaning than just a name, e.g., friend and colleague. Additionally, these meanings are not fixed, they may change at any time, and every person can assign additional meaning using different criteria. Thus, effective context recognition requires a complex and dynamic representation of the personal context and the collaboration of the people to fill the cognitive gap of machines.</p><p>The addition of human collaboration into the context recognition learning of machines is an important part of supervised machine learning <ref type="bibr" target="#b9">[Vapnik, 2013]</ref>. These interactions bring new challenges to the implementation of machine learning algorithms. For instance, humans can be defined as the expert of the supervised algorithms, interacting in an offline fashion by annotating sensor data <ref type="bibr" target="#b10">[Webb, 2003]</ref>, or the interaction can be directly online, as active learning <ref type="bibr">[Settles, 2009;</ref><ref type="bibr" target="#b6">Hoque and Stankovic, 2012;</ref><ref type="bibr" target="#b6">Hossain et al., 2017]</ref>. The human collaboration is important when we are moving into reallife scenarios <ref type="bibr" target="#b7">[Kwapisz et al., 2011]</ref>.</p><p>Other challenges of this collaborative approach are the possibility of overwhelming the humans and the possible differences between the assignment of meaning between people <ref type="bibr" target="#b2">[Chang et al., 2017]</ref>, thus, making the annotation a personal activity. Then, having humans as the input of the information opens the possibility of human error in the collaboration process <ref type="bibr">[Tourangeau et al., 2000]</ref>, and this issue is well known in social sciences and psychology, because of response biases in answering self-reports <ref type="bibr" target="#b10">[West and Sinibaldi, 2013]</ref>, and more importantly, these biases are not well-understood <ref type="bibr" target="#b4">[Freedman et al., 2013]</ref>.</p><p>We propose a novel context model based on the work from <ref type="bibr" target="#b6">[Giunchiglia et al., 2018]</ref>. That work focused on ensuring the reliability of annotations, whereas our focus is on improving personal context representations to get closer to work in reallife scenarios. So, we propose to add a more precise representation of personal context that can also work with machine learning algorithms. We formalize a context model based on ontology and use it with the streaming data to have a knowledge representation of context data. This formalization allows moving towards a generic definition of context that can work with existing multi-label machine learning approaches, using a conversion algorithm. Eventually, the last piece of the puzzle will be the design and development of the Streaming System to manage technically the dynamic context data, and it will be organized in the system architecture with modular components for independent development and easy deployment in current cloud environments.</p><p>Some examples of our model improvement can be seen compared with our main related work <ref type="bibr" target="#b6">[Giunchiglia et al., 2018;</ref><ref type="bibr" target="#b0">Bontempelli et al., 2020;</ref><ref type="bibr" target="#b10">Zeni et al., 2019]</ref>. All of them can take benefit from our novel personal context representation and can use our conversions algorithms to explicitly implement the transformations needed for the machine learning algorithms.</p><p>The paper is structured as follows. Section 2 introduces context modeling. Section 3 illustrates our representation of personal context, while we provide the formal representation in Section 4. Then, we show the learning process to transition from our formal representation to machine learning representation in Section 5 and how our formal representation can be converted to a Direct Acyclic Graph (DAG). Finally, Section 6 describes works related to ours, and Section 7 concludes our paper.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">The context in time</head><p>When we talk about the context, we concentrate on the context of a person called observer. The observer's context is the representation of a partial view of the world. We describe this context into three main dimensions: the viewpoint, the part-whole relation, and the endurant-perdurant.</p><p>Firstly, we have the viewpoint dimension divided as outside viewpoint and observer viewpoint. The outside viewpoint is the view of an ideal observer who can describe everything from a certain point of view. We distinguish this view from the world's static and dynamic properties. In the static property, there are whatever does not change in time, e.g., mountains, buildings, streets. In the dynamic property, there are not only the moving people, but also the moving animals, and facilities in their manifestations, like trains. Then, the observer viewpoint describes how the observer perceives what is around her or him. In this view, we also have the property of being static or dynamic, but it is relative to the movements of the observer.</p><p>Secondly, context is a part-whole relation. In our everyday life, when we do things, we are always embedded in the world. From an ontological view, we are part of the whole world. Thus, we call reference context as the element of the outside context with a volume and extension that is large enough to contain all our movements and changes. For instance, the reference context is the city of Trento when the user walks around, or the users' home when they are at home. In turn, our body as a whole has parts (e.g., arms, mind, legs) that are with us all the time, and they define the internal context of the user. The internal context identifies the elements of the user's body at different levels of abstraction. We usually distinguish between physical parts, such as arms, body, fingers, and mental parts, such as mind, memory, emotions. The context as a part-whole relation is divided into reference context and internal context, and both contexts with different dimensions play a role in our life.</p><p>Thirdly, context as endurant and perdurant refers to the bigger relation to changes in time or space. Events and actions are perdurants and elements, like me, are endurants.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1">The spatio-temporal context</head><p>The context, as viewpoints, defines the reference point from which we construct the context and the context, as partwhole, defines which parts we should consider. So, next, we need to define how we keep track of the context from a quantitative point of view, with a set of quantitative and qualitative measures. Therefore, based on these measures, we introduce the spatio-temporal context.</p><p>The spatio-temporal context consists of the temporal and spatial reference context. The former includes dates, times and all the additional notations like weekdays and seasons. The latter contains the world coordinate system. There are various reasons why context should be represented as a spatio-temporal context. First, this is a common representation when we think of the world. Second, any device today can easily retrieve the time and time zones, and the space coordinates (e.g., via GPS). Third, time can be used to measure the changes in all the elements of the world, all evolving at different speeds, thus temporal context allows us to use them together based time. Finally, a lot of data about the spatial reference context and its sub-contexts are available from external sources (e.g., Google Maps, OpenStreetMaps).</p><p>The spatio-temporal context, also called objective context, at time t is defined as: The number and type of persons and objects change over time. Hence, we will have a sequence of time-tagged states, namely O = {o 1 , . . . , o n }. We call the sequence O as the streaming context. In the streaming context, within the given reference location, it is easy to compute spatial relations (e.g., near, right, left, in front, far relative to the location) of the different elements among themselves. For instance, the system can compute that the smartphone is in the home building and the smartphone is near the computer.</p><formula xml:id="formula_0">o t = (D t , T t : L t ,</formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2">The objective and subjective context</head><p>The spatio-temporal context is also called objective context, since all the relations are computed concerning what is objectively measured, in terms of spatial relations. However, notice that different observers will have different views of the world. For instance, the school building has the function of studyplace from the point of view of a student and is the teacher's workplace. The word function here is used with the precise meaning defined in <ref type="bibr" target="#b4">[Giunchiglia and Fumagalli, 2017;</ref><ref type="bibr" target="#b6">Giunchiglia et al., 2018]</ref>. Hence, "the function of an object formalizes the behavior that an object is expected to have" <ref type="bibr" target="#b4">[Giunchiglia and Fumagalli, 2017]</ref>. For instance, objects are trains and buildings. The expected behavior may be the purpose of the object (e.g., fridge) or the role of a person (e.g., friend).</p><p>The subjective context includes both the objective context elements and the function of persons and objects as seen by the user. Thus, the subjective context at time t is defined as:</p><formula xml:id="formula_1">s t =(D t , T t : L t , me, coord t (me), F t (P 1 ), . . . , F t (P k ), F t (O 1 ), . . . , F t (O m )),</formula><p>where F n (P k ) and F n (O m ) are the functions with respect to a person P k and an object O m , respectively. The number and type of persons and objects, and their functions change over time. The sequence of subjective contexts over time is defined as the subjective streaming context S = {s 1 , . . . , s n }.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.3">The endurant and perdurant context</head><p>In the endurant context <ref type="bibr">[Giunchiglia et al., 2017]</ref>, its parts are endurants, essentially objects where the spatial extension of their actions is contained by the space defined by the spatial (reference) context. The actions "represents how objects change in time" <ref type="bibr" target="#b4">[Giunchiglia and Fumagalli, 2017]</ref>. For instance, running in a park performed by a runner. We also need to represent actions, in particular, the actions that are executed by the endurant me and also by any other elements of the outside dynamic context. Actions can be seen in two ways: (i) actions modeled as processes, namely as sequences of single micro-steps, each of length close to zero; (ii) actions modeled as events, which are often also called perdurants, namely as complete movements which last for a certain duration. Actions as events have key properties, similar to endurants. An event and an action can associate with a set of component sub-events and sub-actions.</p><p>Considering the mentioned concepts, the fundamentally different role of space and time should become clear. Whereas the parts of the space context are only used to limit the space where things happen, the parts of time have the main goal to detail how actions get executed. Things get complicated by adding objects, people, and functions as shown in the data representations shown in Table <ref type="table" target="#tab_1">1</ref>.</p><p>It is worth noting that each function of a person is associated with a limited set of actions, and the type of action that a person can perform can be considerably limited by knowing his or her function.</p><p>The actions apply to me and persons, whereas functions apply to me, other objects and persons. Notice also, that the stored location L is limited by the most specific location that we can compute. This is because the bigger locations are assumed to be static and stored in the system. For events, instead, we store the smallest possible most general event as well as those component actions which are done during a certain period. Thus, for instance, the action/event meeting can have sub-actions such as talking, walking, listening, typing. Table <ref type="table" target="#tab_2">2</ref> reports the streaming context matrix of Example 1 and shows how the context changes over time.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">The current context</head><p>The streaming context describes the contexts of an observer according to time. To represent each context occurrence, we define a set of notions to build a figure of the current context. We mainly divide the current context into four types of context according to how things compose in space and time. From top to bottom we have the following cases:</p><p>• 1L1E (One Location One Event), such as a lecture holds in a classroom;</p><p>• 1LME (One Location Multiple Events), such as a sequence of meetings hold in an office, or eat breakfast, lunch and dinner at home;</p><p>• 1EMC (One Event Multiple Locations), such as a travel goes from many different places;</p><p>• MEMC (Multiple Events Multiple Locations), it is the most complex case, which mixes with the former three cases.</p><p>An 1EMC example is shown in Figure <ref type="figure" target="#fig_1">1</ref>, which describes the following travel scenario around observer me.</p><p>Example 1 In the travel scenario, me is Xiaoyue, she has travel named Travel 1 in the Trentino of Italy from 12:00 to 13:00 of 2th, June 2021. From 12:00 to 12:30 on this day, she takes train 1 from Rovereto to Trento, she sits on seat 1 by herself. From 12:30 to 12:55 on the same day, she walks on Roads 2 from Trento Train Station to Xiaoyue's Home, together with her friend named Haonan. In addition, Xiaoyue talks to Haonan, and Haonan listens to Xiaoyue when they are walking. This scenario involves one event travel and multiple locations.</p><p>In general, in the Figure <ref type="figure" target="#fig_1">1</ref> In the rest of this section, we list the attributes of Location, Event, Object, Person and Action, each kind of attributes is represented as a box in the figure. The Location has the following attributes:</p><p>• Spatial properties: Coordinates (x i , y i , z i ), Volume (∆x i , ∆y i , ∆y i ) , and InLocation(L i ) shows the super Location for the top level Location;</p><p>• Visual properties, namely some properties of Location that can be observed visually;</p><p>• Location' functions: F unctionOf (U ) with U ∈ {P 1 , . . . , P k , O 1 , . . . , O M }, which shows location's functions for persons and objects;</p><p>• An extra box: including the rest part of the context that happens in the Location.</p><p>The Event orders by time and is represented by a box with round corners. Events include actual events and virtual events. An Event can have Sub-Events, the Event and Sub-Event have the following attributes:</p><p>• Super Event for the top level event: InEvent(E i ), which shows the super event E i for the current Event;</p><p>• Temporal properties: Begin Time -End Time (Date i , T ime i -Date j , T ime j ), which shows the date and time of the start and end of the event;</p><p>• An extra box: including the rest part of the context that happens in the Event.</p><p>The Object appears in Event and has the following attributes:</p><p>• Spatial properties: Coordinates (x i , y i , z i ), In/Far/... (P i /O m /...);</p><p>• Visual properties, namely some properties of Object that can be observed visually; </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">The current context as a Knowledge Graph</head><p>We use the Entity Type Graph (ETG) and the Entity Graph (EG) in ontology to represent the context. ETG is a knowledge graph where nodes are entity types, which are further decorated with data properties. The object properties are presented in the graph representing the relations among the entity types.</p><p>In Figure <ref type="figure" target="#fig_3">2</ref>, white box nodes are entity types that include data properties with data types, the green box nodes enumerates the values of data property. The object properties are connecting all entity types, represented by diamond nodes with arrows. The inheritance relation in the ETG is represented by an arrow from the super-class to the sub-class, and the subclass inherits all the data properties and object properties of its super-class. The EG populates the entity types and properties defined in the ETG with specific values. It is a data graph where nodes are entities that are connected by object property values representing the relations. Each entity further includes data property values. The streaming context can be viewed as a stream of EGs, in which each EG describes the context at a different time.</p><p>We design an EG example as Figure <ref type="figure" target="#fig_4">3</ref> according to the scenario in Example 1. The graph represents the context around "Me", this contains entities shown by nodes, e.g., "Smartphone", "Talking", "Walk". Also, we can see object property values.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Learning Context</head><p>AI applications like smart personal assistants provide a service to the users based on their context. The context information is usually not available to the machine, and hence it has to infer the location or the activity of the user from sensor data (e.g., GPS, accelerometer, nearby Bluetooth devices). In our scenario, Xiaoyue is carrying a smartphone that generates a stream of sensor readings, and she annotates the data by answering questions about her context, e.g., "Where are you" and "What are you doing?". The sensor data are aggregated in time windows generating a stream of instances (e.g., the average number of nearby Bluetooth devices in the last 30 minutes). On each incoming instance, the machine decides whether to query the user to acquire the labels. The machine learning technique defines the query strategy, and in the simplest case, the labels are acquired on every instance.  belong to i-th concept in H and 0 otherwise. The machine is trained on a stream of examples z t = (x t , y t ) drawn from a ground-truth distribution P (X, Y) that is always consistent with a ground-truth hierarchy H, i.e., if there is an edge from class c i to class c j , then y i = 1 implies y j = 1 and conversely y j = 0 implies y i = 0. The goal of this hierarchical classification tasks is to learn a classifier that recognize well the context on future sensor readings.</p><p>The ETG and EG introduced in Section 4 can be used as prior knowledge about the structure of the hierarchy. They encode the available information about the user and the world. Algorithm 1 shows the conversion from ETG and EG to a DAG H. The first step is to convert each entity type (etype) in the ETG as a node in H (lines 3 -5). Second, each entity in EG also becomes a node that is added as a child of the node referring to the etype of the entity (lines 6 -10). The hierarchy encodes the information about the current user, so the Me etype and the corresponding entity (e.g., Xiaoyue entity in Figure <ref type="figure" target="#fig_4">3</ref>) are not considered.</p><p>The properties of the ETG are grouped in properties that are context depends and properties that are static. The value of the former changes every time the users change their context and, in Figure <ref type="figure" target="#fig_3">2</ref> are Q = {near, use, interact, in, do, happenIn, during, participate}. For instance, if Xiaoyue travel from the city of Trento to Rovereto, the in property will change accordingly. In contrast, the fact that Trento is partOf Italy can be assumed to remain valid even if user's context is changed. This distinction is necessary since the value of context-dependent properties are derived from the output of the context recognition task (e.g., the machine recognizes that the user is in the city of Trento and thus updates the in property accordingly). The object properties that are not contextual are converted as follow. Given a object property p ∈ {isA, partOf, has} that links the etype A to B, then the node referring to etype A becomes a child of the node of etype B. For the other object properties, a new node referring to the property is added as child of the codomain etype node (lines 13 -19). For every object property value i of the prop- erty p linking the entities a and b, a new node c i is added as child of c p (i.e., the node referring to the property P ), and as parent of c b (i.e., the node pointing to the entity b) (lines 20 -25). Finally, all nodes that does not have a parent are connect to the root node and the transitive reduction is applied (lines 27 -30). Figure <ref type="figure" target="#fig_5">4</ref> shows an extract of the DAG resulting from applying Algorithm 1 on EG and ETG presented in Section 4.</p><p>Every node in H has a unique identifier that is used to reference back to the ETG and EG. The node name can be translated into a human-readable text that is used to interact with the user. This aspect is left as future work. The concept hierarchy available at the beginning can evolve over time and has to be continually updated. This aspect has been defined as knowledge drift and is addressed in <ref type="bibr" target="#b1">[Bontempelli et al., 2021]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6">Related Work</head><p>Considering our novel context modelling of the personal context, the context learning, and the architecture presented in this paper, we can show as use cases the papers that individually compartmentalizes examples for these improvements. Current works on context recognition focus on learning the relationship between the input data (sensor readings) and the target concepts (context). The structure of the context is implicitly learned by the implemented machine learning algorithm during the training phase. The new parts are described in Section 3 and compared with related work we can outline the following: 1) the connection between the learned context model; 2) The extension of new dimensions and classification of context, i.e. Internal State, Functions, Actions. The Algorithm 1 bridges the gap between the two modellings allowing context recognition to leverage both machine learning solutions and knowledge graph representation (ETG and EG). For instance, the ETG and EG can generate the questions needed by the machine to interact with the users. For instance, the personal context recognition model shown in <ref type="bibr" target="#b6">[Giunchiglia et al., 2018]</ref>, showed to be a good approach to increase the accuracy of the context recognition algorithms. Our work described in Section 3 enhance the representation of the personal context with the hope to perform better in real-life scenarios. In practice, existing approaches for context recognition using batch or streaming sensor data <ref type="bibr" target="#b8">[Vaizman et al., 2018;</ref><ref type="bibr" target="#b0">Bontempelli et al., 2020;</ref><ref type="bibr" target="#b10">Zeni et al., 2019]</ref> do not leverage on an explicit context modeling. The context representation is implicit in the la-bels used to train the machine learning model. The context formalization introduced in this work can be used to structure the output of these machine learning models according to our representation. Moreover, it can help machine learning approaches to interact with users. For instance, the machine can ask if Haonan is a friend of Xiaoyue since they are walking together. Approaches that use active learning strategies (e.g., <ref type="bibr">[Settles, 2009;</ref><ref type="bibr" target="#b6">Hoque and Stankovic, 2012;</ref><ref type="bibr" target="#b6">Hossain et al., 2017]</ref>) can benefit of our representation.</p><p>Also, existing frameworks for creating context-aware mobile applications, such as Ferreira et al., do not consider the modelling of the context.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7">Conclusion and Future Work</head><p>In this paper, we moved forward with a better representation of personal context in real-life environments. We proposed an improved representation of the personal context, adding the internal state, functions, and actions. The learning aspect of our work is the formal definition of an algorithm to transform the streaming input data to ML algorithms. We will put all these components in the system architecture.</p><p>In comparison with the work on personal context recognition for human-machine collaboration <ref type="bibr" target="#b6">[Giunchiglia et al., 2018]</ref>, we have shown an enhancement related to the model representation of the personal context.</p><p>Additionally, we have shown how our novel personal context representation can also be leveraged by machine learning algorithms to include prior knowledge about the structure of their output and can be used to drive the interaction with the user. Future work will focus on evaluating the impact of our formalization on an existing approach for fixing mislabeled data when learning the users' contexts <ref type="bibr" target="#b10">[Zeni et al., 2019]</ref>.</p><p>The next step is to propose and implement a modern design of the services related to iLog <ref type="bibr" target="#b10">[Zeni et al., 2014]</ref> by a centralised streaming system and linking the personal context data collections with other distributed services of machine learning. This implementation will allow us to test and evaluate our novel context model in near real-life scenarios.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head></head><label></label><figDesc>, all elements are divided into Perdurants and Endurants. Perdurants are the Event and Action, and Endurants include Person, Object, and Location. An Event happens in a Location, a Person and Object appear in an Event and Action is in a Person. Most elements' inclusion relationships IN can be represented by the positions of those elements' internal boxes. For the top-level Location and Event, we can add an extra attribute box for them respectively as InLocation(Location) and InEvent(Event), to represent their belongs.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: One Event Multiple Locations: a travel around me. Representation of the Example 1.</figDesc><graphic coords="5,53.88,53.84,504.23,523.22" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head></head><label></label><figDesc>The user's context recognition is a supervised learning problem in which an instance x is associated to multiple concepts y (aka classes in machine learning). The concepts are organized in a ground-truth hierarchy H = (C, I), which is a direct acyclic graph (DAG) where nodes C = {1, . . . , c} are the concepts and edges I ⊂ C × C are is-a relations, i.e., I = {(c i , c j ) | c i , c j ∈ C and c i is a child of c j } [Silla and Freitas, 2011]. The labels of the instances are indicator vectors y ∈ {0, 1} c , where the i-th elements is 1 if x</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: An ETG representing partially the personal context in our travel example.</figDesc><graphic coords="6,53.88,53.83,504.25,396.24" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_4"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: An EG representing partially the scenario about travelling of Example 1.</figDesc><graphic coords="7,66.49,53.84,479.02,280.81" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_5"><head>Figure 4 :I</head><label>4</label><figDesc>Figure 4: A partial representation of the DAG obtained from ETG and EG examples in Section 4. Orange nodes are derived from EG. Blue nodes are derived from ETG. Red nodes are actions and green nodes are the functions.</figDesc><graphic coords="7,315.00,424.62,246.12,175.58" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head></head><label></label><figDesc>me, coord t (me), P 1 t : coord t (P 1 ), . . . , P k t : coord t (P k ), O 1 t : coord t (O 1 ), . . . , O m t : coord t (O m )) where D n , T n stands for date and time respectively, L n is the location, namely the smallest possible spatial reference context that we can compute. Here me is the observer, P i are persons and O m are objects. The function coord(. . . ) computes the spatial coordinates of me, objects, and persons.</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 1 :</head><label>1</label><figDesc>, T 1 : super(L 1 ), super(E 1 ), L 1 , E 1 , me, coord 1 (me), A me 1 , F 1 (P 1 ) :A P 1 1 , . . . , F 1 (P k ) : A P k 1 , F 1 (O 1 ), . . . , F 1 (O m ) , D 2 , T 2 : super(L 2 ), super(E 2 ), L 2 , E 2 , me, coord 2 (me), A me 2 , F 2 (P 1 ) : A P 1 2 , . . . , F 2 (P k ) : A P k 2 , F 2 (O 1 ), . . . , F 2 (O m ) , . . . , D n , T n : super(L n ), super(E n ), L n , E n , me, coord n (me), A me n , F n (P 1 ) : A P 1 n , . . . , F n (P k ) : A P k n , F n (O 1 ), . . . , F n (O m ) }The personal streaming context, where En is an event, super(Ln) and super(En) are the super-classes of Ln or En, respectively. The set of actions performed by me or by the persons based on their functions are denoted with A k n = {a1, . . . , ai}, with k ∈ {me, P 1 , . . . , P k }.</figDesc><table><row><cell>Dn 02/06/2021 12:15 Trentino Tn super(Ln) super(En) Travel 1 02/06/2021 12:30 Trentino Travel 1</cell><cell>Ln Train 1 Take Train x41, y41, z41 Sitting En coordn(me) A me n Roads 2 Walk x43, y43, z43 Walking, Talking</cell><cell>Fn(P 1 ) : A P 1 n NaN FriendOf( Xiaoyue, Haonan): Walking, Listening</cell><cell>Fn(O 1 ) RestToolOf( Xiaoyue, Seat 1) NaN</cell></row></table><note>• Object's functions: F unctionOf (U ) with U ∈ {P 1 , . . . , P k , O 1 , . . . , O M }, which shows Object's functions for persons and other objects. The Person appears in Event and has the following attributes: { D 1</note></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 2 :</head><label>2</label><figDesc>A streaming context matrix representing the travel scenario of Example 1 from the point of view of Xiaoyue, i.e., the observer me. P 1 is Haonan and O 1 is the object "Seat 1". Each column is a property, and every row stands for the current context in a specific timestamp. • An extra box: including the Actions of Person. The Action is similar with Event, it orders by time and is represented by a box with round corners. An Action can has many Sub-Action, the Action and Sub-Action have following attributes, each attribute has a box for itself. • Temporal properties: Begin Time -End Time (Date i , T ime i -Date j , T ime j ), which shows the date and time of the start and end of the Action; • Visual properties, namely some properties of Action that can be observed visually; • Means of the Action: Means(O m /...); • Sub-action: Action i; • Action's functions: F unctionOf (U ) with U ∈ {P 1 , . . . , P k , O 1 , . . . , O M }, which shows Action's functions for persons and objects.</figDesc><table /><note>• Spatial properties: Coordinates (x i , y i , z i ), In/Far/...(P i /O m /...); • Visual properties, namely some properties of Person that can be observed visually; • Person's functions: F unctionOf (U ) with U ∈ {P 1 , . . . , P k , O 1 , . . . , O M }, which shows Person's functions for other persons and objects; • Internal states: Physical states (InP ain()), Mental states (InM ood(), InStress());</note></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" xml:id="foot_0">Copyright c 2021 for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" xml:id="foot_1">Twelfth International Workshop Modelling and Reasoning in Context (MRC) @IJCAI 2021</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" xml:id="foot_2">Copyright c 2021 for this paper by its authors. Use permitted under Creative Commons Attribution 4.0 International (CC BY 4.0).</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgements</head><p>The research conducted by Fausto and Xiaoyue has received funding from the European Union's Horizon 2020 FET Proactive project "WeNet -The Internet of us", grant agreement No 823783.</p><p>The research conducted by Marcelo and Andrea has received funding from the "DELPhi -DiscovEring Life Patterns" project funded by the MIUR Progetti di Ricerca di Rilevante Interesse Nazionale (PRIN) 2017 -DD n. 1062 del 31.05.2019.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Learning in the wild with incremental skeptical gaussian processes</title>
		<author>
			<persName><surname>Bontempelli</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence, IJCAI-20</title>
				<meeting>the Twenty-Ninth International Joint Conference on Artificial Intelligence, IJCAI-20</meeting>
		<imprint>
			<date type="published" when="2020">2020. 2020</date>
			<biblScope unit="page">7</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<monogr>
		<author>
			<persName><surname>Bontempelli</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2103.14874</idno>
		<title level="m">Humanin-the-loop handling of knowledge drift</title>
				<imprint>
			<date type="published" when="2021">2021. 2021</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">An investigation of using mobile and situated crowdsourcing to collect annotated travel activity data in real-word settings</title>
		<author>
			<persName><forename type="first">Chang</forename></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">ternational Journal of Human-Computer Studies</title>
		<imprint>
			<biblScope unit="volume">102</biblScope>
			<biblScope unit="page" from="81" to="102" />
			<date type="published" when="2017">2017. 2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Aware: mobile context instrumentation framework</title>
		<author>
			<persName><surname>Ferreira</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Frontiers in ICT</title>
		<imprint>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="issue">6</biblScope>
			<date type="published" when="2015">2015. 2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Interviewer and respondent interactions and quality assessments in a time diary study</title>
		<author>
			<persName><surname>Freedman</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Teleologies: Objects, actions and functions. In International conference on conceptual modeling</title>
				<editor>
			<persName><forename type="first">Fumagalli</forename><surname>Giunchiglia</surname></persName>
		</editor>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2013">2013. 2013. 2017. 2017</date>
			<biblScope unit="volume">10</biblScope>
			<biblScope unit="page" from="520" to="534" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Personal context modelling and annotation</title>
		<author>
			<persName><surname>Giunchiglia</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IEEE International Conference on Pervasive Computing and Communications Workshops (Per-Com Workshops)</title>
				<imprint>
			<publisher>IEEE</publisher>
			<date type="published" when="2017">2017. 2017. 2017</date>
			<biblScope unit="page" from="117" to="122" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Activity recognition in smart homes using active learning in the presence of overlapped activities</title>
		<author>
			<persName><surname>Giunchiglia</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">2012 6th International Conference on Pervasive Computing Technologies for Healthcare (PervasiveHealth) and Workshops</title>
				<imprint>
			<publisher>IEEE</publisher>
			<date type="published" when="2012">2018. 2018. 2018. 2012. 2012. 2017. 2017</date>
			<biblScope unit="volume">38</biblScope>
			<biblScope unit="page" from="312" to="330" />
		</imprint>
	</monogr>
	<note>Active learning enabled activity recognition</note>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Activity recognition using cell phone accelerometers</title>
		<author>
			<persName><surname>Kwapisz</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">A survey of hierarchical classification across different application domains</title>
				<editor>
			<persName><forename type="first">Roger</forename><surname>Tourangeau</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Lance</forename><forename type="middle">J</forename><surname>Rips</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Kenneth</forename><surname>Rasinski</surname></persName>
		</editor>
		<meeting><address><addrLine>Settles</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2000">2011. 2011. 2009. 2011. 2000. 2000</date>
			<biblScope unit="volume">12</biblScope>
			<biblScope unit="page" from="74" to="82" />
		</imprint>
	</monogr>
	<note>The psychology of survey response</note>
</biblStruct>

<biblStruct xml:id="b8">
	<monogr>
		<title level="m" type="main">Context recognition in-the-wild: Unified model for multi-modal sensors and multi-label classification</title>
		<author>
			<persName><surname>Vaizman</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2018-01">2018. January 2018</date>
			<biblScope unit="volume">1</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<title level="m" type="main">The nature of statistical learning theory</title>
		<author>
			<persName><forename type="first">Vladimir</forename><surname>Vapnik</surname></persName>
		</author>
		<author>
			<persName><surname>Vapnik</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2013">2013. 2013</date>
			<publisher>Springer science &amp; business media</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Fixing mislabeling by human annotators leveraging conflict resolution and prior knowledge</title>
		<author>
			<persName><forename type="first">Andrew</forename><forename type="middle">R</forename><surname>Webb</surname></persName>
		</author>
		<author>
			<persName><forename type="first">;</forename><surname>Webb</surname></persName>
		</author>
		<author>
			<persName><surname>West</surname></persName>
		</author>
		<author>
			<persName><forename type="first">;</forename><surname>Sinibaldi</surname></persName>
		</author>
		<author>
			<persName><surname>Zeni</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2014 ACM international joint conference on pervasive and ubiquitous Computing: Adjunct publication</title>
				<editor>
			<persName><forename type="first">Brady</forename><forename type="middle">T</forename><surname>West</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Jennifer</forename><surname>Sinibaldi</surname></persName>
		</editor>
		<meeting>the 2014 ACM international joint conference on pervasive and ubiquitous Computing: Adjunct publication</meeting>
		<imprint>
			<publisher>John Wiley &amp; Sons</publisher>
			<date type="published" when="2003">2003. 2003. 2013. 2013. 2014. 2014. 2019. 2019</date>
			<biblScope unit="volume">3</biblScope>
			<biblScope unit="page" from="1" to="23" />
		</imprint>
	</monogr>
	<note>Proceedings of the ACM on Interactive</note>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
