=Paper=
{{Paper
|id=Vol-2142/paper9
|storemode=property
|title=Autonomous adaptation of software agents in the support of human activities
|pdfUrl=https://ceur-ws.org/Vol-2142/paper9.pdf
|volume=Vol-2142
|authors=Esteban Guerrero,Ming-Hsin Lu,Hsiu-Ping Yueh,Helena Lindgren
|dblpUrl=https://dblp.org/rec/conf/ijcai/GuerreroLYL18
}}
==Autonomous adaptation of software agents in the support of human activities==
Autonomous adaptation of software agents in the support of human activities Esteban Guerrero1 , Ming-Hsin Lu2 , Hsiu-Ping Yueh2 , and Helena Lindgren1 1 Computing Science Department, Umeå University, Sweden {esteban, helena}@cs.umu.se 2 Human Performance and Technology Lab., National Taiwan University, Taiwan {f01630002,yueh}@ntu.edu.tw Abstract. This paper is aimed at formalizing the interplay among a person to be assisted, an assistive agent-based software, and a caregiver. We propose principles that agents should follow in such interplay, this principles may have impact in different agent-based assistive technology. We propose a mechanism to integrate individual’s information into the final decision-making process of an agent. Moreover, we endow agents with mechanisms for evaluating the distance between independent and supported activity execution, the so called zone of proximal development (ZPD) in four scenarios: I) independent activity execution by a person; II) ZP DH activity performance of a person supported by another person (e.g. a therapist); III) the ZP DS representing a potential activities when a person is supported by a software agent; and IV) the ZP DH+S when a person is supported by a caregiver and a software agent. Formal argu- mentation theory is used as foundation. Our formal models were tested using a prototype using augmented reality as assistive software. A pilot study with older adults and health-care personnel was performed and formal and empirical results are presented. Keywords: Argumentation theory · Rational agents · Assistive technol- ogy · Human activity · Activity theory. 1 Introduction This paper is aimed at investigating assistive technology using argumentation- based agents and the interplay with individuals that require physical assistance and their caregivers. We present formal and empirical results on how intelligent software adapts to support activities of individuals including, those who need assistance and care givers. The focus of the paper is on the provision of human-like characteristics to software agents in order to provide adaptable support, namely common-sense and reflection on action decision. The proposed agent model is oriented to reason about human activities, i.e., identify and interpret activities, and support indi- viduals during the execution of physical activities. To this end, representations of complex activities from Activity Theory [15] were utilized to characterize the 2 Guerrero et al. knowledge of software agents about human activities and model their decision- making process. Formal argumentation theory is used to provide non-monotonic reasoning to the agents. Moreover, we present a novel information model oriented at how an agent3 may reflect on their actions. In human learning literature, re- flection enables a person to correct distortions in her/his beliefs and errors in problem-solving [17]. We contribute with a first step on how rational software agents may reflect during the support of human activities. Finally, as core of our research, we propose a model of adaptation of a support level for agents, based on a computation version of the so-called zone of proximal development (ZPD) [23]. Our model of adaptation is formally presented and empirically tested. The research questions (RQ) addressed in this paper are the following: – RQ1: how an agent may infer potential activities that an individual needs and performs with and without its assistance? – RQ2: in a smart environment scenarios, where individuals require support from others to execute an activity, how an agent-based software may adapt autonomously to team-up with humans to enhance such support? – RQ3: how rational agents can “reflect” on decision to make when a human is in the loop? This paper is structured as follows: Section 2 presents definitions about how human activities are structured, we also introduce some definitions about formal argumentation theory and argument-based reasoning. In Section 3 our main contributions are presented trying to solve RQ1-3. Using a medical scenario, we developed a prototype to test empirically the level of assistance, some results are presented in Section 4. Conclusions and a discussion is presented in Section 5. 2 Background In this paper, the human perspective is investigated using Activity Theory [15], which is a social sciences framework oriented to understand human complex ac- tivities. On the other hand, formal argumentation theory is used to characterize precisely the internal reasoning of agent software. 2.1 Activity theory In this paper, activity theory is used with two purposes: 1) for structuring the knowledge of an agent following a hierarchical model; and 2) to understand the potential level of activity achievement of a person. Activity as structure. An activity consists of a set of actions. At the lowest level, an action consists of a set of operations. Actions are oriented to goals and are executed by the actor at a conscious level, in contrast with operations which 3 Hereinafter we will identify a rational software agent as just an agent. Assistive agents and human activities 3 do not have a goal of their own and which are executed at the lowest level as automated, unconscious processes. An activity model (AT) (see Definition 1) corresponds to information of a person that an agent uses to reason about an activity. Definition 1 (Activity model). Let P be a logic program capturing the behav- ior rules of an activity. LP denotes the set of atoms which appear in a program P . An AT model is a tuple of the form hAx, Go, Opi in which: – Ax = {ax1 , . . . , axj }(j > 0) is a set of atoms such that Ax ⊆ LP . Ax denotes the set of actions in an AT model. – Go = {g1 , . . . , gk }(k > 0) is a set of atoms such that Go ⊆ LP . Go denotes the set of goals in an AT model. – Op = {o1 , . . . , ol }(l > 0) is a set of atoms such that Op ⊆ LP . Op denotes the set of goals in an AT model. In artificial intelligence literature, this hierarchical structure has been used as framework to represent knowledge of software agents, e.g. in [8,10,11,18]. An activity framework corresponds to the goals, observations and actions of an agent oriented to assist a human during the execution of an activity, which in turn is represented by the AT model. To capture the knowledge of an agent about its environment, we use extended logic programs [7] . An extended normal program P is a finite set of extended normal clauses. By LP , we denote the set of atoms which appear in a program P. ELP use both strong negation ¬ and not, representing common-sense knowledge through logic programs. Definition 2 (Activity framework). An activity framework ActF is a tuple of the form hP, HA , G, O, ATi in which: – P is a logic program. LP denotes the set of atoms which appear in P . – HA = {h1 , . . . , hi } is a set of atoms such that HA ⊆ LP . HA denotes the set of hypothetical actions which an agent can perform in a world. – G = {g1 , . . . , gj } is a set of atoms such that G ⊆ LP . G denotes a set of goals of an agent. – O = {o1 , . . . , ok } is a set of atoms such that O ⊆ LP . O denotes a set of world observations of an agent. – AT is an activity model of the form: hAx, Go, Opi, following Definition 1. ActF according to Definition 2 defines the space of knowledge of assistive agents without considering external assistance, for example from other assistive agents (human or software) actions. In this knowledge space, an argument-based process (see Figure 1) can be performed to obtain sets (or sets of sets) of ex- plainable structures support-conclusion for what is the best assistive action to take. Potential level of activity achievement. Vygotsky in [23] proposed to mea- sure the level of development not through the level of current performance, but through the difference (“the distance”) between two performance indicators: 1) 4 Guerrero et al. an indicator of independent problem solving, and 2) an indicator of problem solv- ing in a situation in which the individual is provided with support from other people [14]. This indicator was coined as a zone of proximal development ZPD and it has been used extensively in social sciences (see [1,4,13,21]) to understand changes of individuals during assisted learning processes. ICF qualifiers. The notion of qualifier to specify the extent of the func- tioning or disability of an individual was introduced by the International Clas- sification of Functioning, Disability and Health (ICF)4 [19] proposing two main quantifiers: performance and capacity. In general, a qualifier specifies informa- tion about functioning status: the magnitude, the location and the nature of any activity-related problem [20]. 2.2 Formal argumentation theory Argumentation-based systems, have become influential in artificial intelligence particularly in multi-agent systems design (see [3] for a systematic review). Formal argumentation process Activity Acceptability Argument-based Knowledge Conflict Justified fragments of activity Conclusions base analysis conclusions construction fragments Activity Argument Extensions fragments framework STEP 1 STEP 2 STEP 3 STEP 4 Fig. 1. Inference of an argument-based conclusion using a formal argumentation pro- cess Reasoning about human activities may be performed using a bottom-up man- ner, building fragments of an activity explaining the current situation of a person. This process compresses the STEP 1 in Figure 1. Definition 3 (Hypothetical fragments). Let ActF = hP, HA , G, O, ATi be an activity framework. A hypothetical fragment of an activity is of the form 0 0 0 HF = hS, O , h, gi such that: 1) S ⊆ P, O ⊆ O, h ∈ HA , g ∈ G; 2) S ∪O ∪{h} 0 0 is consistent; 3) g ∈ T such that ASP (S ∪ O ∪ {h}) = hT, F i; and 4) S and O are minimal w.r.t. set inclusion. ASP (S) is a function that returns an answer-set solution of an ELP program, i.e., it provides a common-sense reasoning process given a program as input. In short, an hypothetical fragment is a consistent manner to explain (inter- connected) parts of an activity. Some of these fragments may be contradictory given inconsistent information in the AT model or/and defeasible information captured by an agent (STEP 2 in Figure 1). 4 http://www.who.int/classifications/icf/en/ Assistive agents and human activities 5 Definition 4 (Contradictory relationships among fragments). Let ActF = hP, HA , G, O, Actsi be an activity framework. Let HF1 = 0 0 hS1 , O1 , a1 , g1 i, HF2 = hS2 , O2 , a2 , g2 i be two fragments such that HF1 , HF2 ∈ HF. ASP (Supp(HF1 )) = hT1 , F1 i and ASP (Supp(HF2 )) = hT2 , F2 i denote the semantic evaluation of the support, then HF1 attacks HF2 if one of the following conditions hold: 1) α ∈ T1 and ¬α ∈ T2 .; 2) α ∈ T1 and α ∈ F2 . An argumentation framework is a pair hArgs, atti in which Args is a finite set of arguments and att ⊆ Args × Args. In [10] an argumentation-based ac- tivity framework for reasoning about activities was proposed, by considering argumentation as inference method: Definition 5 (Activity argumentation framework). Let ActF be an activ- ity framework of the form hP, HA , G, O, Actsi; let HF be the set of fragments w.r.t. ActF and AttHF or simply Att the set of all the attacks among HF. An activity argumentation framework AAF with respect to ActF is of the form: AAF = hActF, HF, Atti Dung in his seminal work [6] introduced a set of patterns of selection of arguments called argumentation semantics. An argumentation semantics SEM is a formal method to identify conflict outcomes from argumentation frameworks (AF). Definition 6. Let AAF = hActF, HF, Atti be an activity argumentation frame- work AAF with respect to ActF = hP, HA , G, O, Actsi An admissible set of frag- ments S ⊆ HF is stable extension if and only if S attacks each argument which does not belong to S. preferred extension if and only if S is a maximal (w.r.t. inclusion) admissible set of AAF. complete extension if and only if each argu- ment, which is acceptable with respect to S, belongs to S. grounded extension if and only if it is a minimal (w.r.t. inclusion) complete extension. ideal extension if and only if it is contained in every preferred set of AAF. The sets of arguments suggested by an argumentation semantics are called extensions. Let SEM () be a function returning a set of extensions, given an AF such as an AAF. We denote SEM (AAF ) = {Ext1 , . . . , Extk } as the set of k extensions generated by an argumentation semantics w.r.t. an activity argumen- tation framework AAF . Definition 7. 1) An fragment HFA ∈ HF is acceptable w.r.t. a set S of frag- ments iff for each fragment HFB ∈ HF: if HFB attacks HFA , then HFB is attacked by S. 2) conflict-free set of fragments S in an activity is admissible iff each fragment in S is acceptable w.r.t. S. Using these notions of fragment admissibility, different argumentation se- mantics can draw given an activity argumentation framework (STEP 3 Figure 1): 6 Guerrero et al. Definition 8. Let AAF = hActF, HF, Atti be an activity argumentation frame- work following Definition 5. An admissible set of fragments S ⊆ HF is: 1) stable if and only if S attacks each fragment which does not belong to S; 2) preferred if and only if S is a maximal ( w.r.t. inclusion) admissible set of AAF ; 3) complete if and only if each fragment, which is acceptable with respect to S, belongs to S; and 4) the grounded extension of AAF if and only if S is the minimal ( w.r.t. inclusion) complete extension of AAF . Conclusions of an argument-based reasoning about an activity (see STEP 4 in Figure 1) may be obtained using a skeptical perspective, i.e., accepting only irrefutable conclusions as follows: Definition 9 (Justified conclusions). Let P be an extended logic program, AFP = hArgP , At(ArgP )i be the resulting argumentation framework from P and SEMArg be an argumentation semantics. If SEMArg (AFP ) = {E T1 , . . . , En }(n ≥ 1), then Concs(Ei ) = {Conc(A) | A ∈ Ei }(1 ≤ i ≤ n). Output = i=1...n Concs(Ei ). Where Ei are sets of fragments called extensions. The set of all the extensions generated by SEMArg (AFP ) are denoted as E 3 Reflection on decisions about human activity In this section, we report formal results to understand how autonomous agents may change the level of assistance in different scenarios. Two main formal results are presented in this section: 1) a mechanism for agent’s decision-making based on the individual’s information analyzing consequences of hypothetical actions a mechanism that we called reflection; and 2) a formalism to determine the potential of activity performance in four different cases: independence, supported by another person, supported by a software agent and supported by a team person-agent. 3.1 Reflection on decisions about human activity Conclusions of an argument-based process (Definition 9) about an activity, may contain sets of goal-based conclusions sets, indicating that the agent has different available decision alternatives which are consistent. We propose adding a mech- anism for selecting an appropriate decision but considering those agent’s actions that maximize humans’ goals (Go) in an activity model (AT model Definition 1). An AT model captures all the information necessary to define a human activity. We condensed this process in Algorithm 1. In short, Algorithm 1 takes as input the AT model and the set of extensions from a previous common-sense reasoning output. In lines 8-15 of Algorithm 1 a qualifier is calculated (line 12) over sets of sets of fragments (the so-called extensions in argumentation theory, see Appendix 2). This qualifier calculation is based on computing a similarity function between the current achievement of human goals in AT (OGo ) w.r.t. a set of goal reference (Ref Go line 12). The Assistive agents and human activities 7 Algorithm 1: Goal-based action reflection input : E, AT output : h ∈ HA 1 H ←− ∅ // list of agent’s decisions 2 Go ←− ∅ // list of human’s goals 3 Ref ←− ∅ // list of human’s reference goals 4 numExt = |E| // number of extensions 5 numArg = |A| // number of arguments per extension 6 α ←− 0 // numeric value of a qualifier (0 ≤ α ≤ 4) 7 decisionLat < α, h >= // lattice of decisions 8 for i ← 0 to numExt do 9 for j ←− 0 to numArg do 10 h ←−Act (hfj ) 11 O ←−Obs (hfj ) 12 α ←−Q(OGo , RefGo ) // Qualifier function considering observations and a reference value w.r.t. person goals Go 13 decisionLat ←− (α, h)hfj // decision tuple is qualifier and an agent’s decision w.r.t the current fragment 14 end 15 end 16 return max(α, h) Q function depicted in line 12, follows the qualifier idea presented in previous approaches [10,11], returning a numerical value (0 ≤ α ≤ 4). The importance of Algorithm 1 lies on the mechanism for associating a human activity quantification with the internal action decision of an agent. Proposition 1 and Proposition 2 present two special cases of agent’s behavior when Algorithm 1 is used5 . One is the possibility to have a conclusion with no action, and the second expresses an inconclusive behavior given that stable semantics may return ∅ as output. Proposition 1. An agent calculating a goal-based action reflection Algorithm 1 using a skeptic semantics, grounded or ideal, may result in a conclusive empty decision.6 Proposition 2. An agent calculating a goal-based action reflection Algorithm 1 using the credulous semantics: stable, may result in a inconclusive decision.7 5 We refrain of describing fully the proofs of these propositions due the lack of paper space 6 Proof sketch: output of grounded and ideal may include {∅}. See [5] 7 Proof sketch: output of stable semantics may include ∅. See [5] 8 Guerrero et al. 3.2 Zone of proximal development using formal argumentation In this section, based on the common-sense reasoning of activities using argu- mentation theory, we propose a theory to calculate the following four scenarios in assistive agent-based technology: I. Independent activity execution This scenario describes an observer agent which takes a decision which is purposefully do nothing to support a per- son, or the decision is empty. More formally, the type of fragments (Definition 0 3) generated by the agents are with the form HF = hS, O , h∗ , gi such that ∗ h ∈ HA = {∅, do N othing}. In this setting, all the extensions generated by SEM (AFP ) = E during a period of time will create an activity structure. In other words, the cumulative effect of generating fragments, re-construct an ac- tivity in a bottom-up manner. Moreover, Algorithm 1 returns only values of α, i.e. the current value of a qualifier when the agent does not take any support action. This context defines the baseline of activity execution independence of a person. II. ZP DH : activities supported by another person Similarly to previous scenario, the role of the software agent is being an observer. However, built frag- ments have the form HF = hS, O∗ , h∗ , gi such that h∗ ∈ HA = {∅, do N othing} 0 00 and O∗ = O ∪ O , where O∗ is the set of joint observations from the agent’s 0 00 perspective about the individual supported (O ) and the supporter O . We have 0 00 0 00 00 that O ⊆ O , and O , O 6= ∅. In this scenario, O is considered a reference set of observations (Ref lines 3 and 12 in Algorithm 1). Algorithm 1 will re- turn a value of α which measures in what extent an individual follows the guide provided by another person. When multiple extensions are collected during the period of time that the individual is supported, then a different set of activities than individual activity execution may be re-generated in a bottom-up manner. III. ZP DS : activities supported by an agent In this scenario, an assistive agent takes a decision oriented to uphold human interests. This is a straightfor- ward scenario where h ∈ HA 6= {∅, do N othing}. IV. ZP DH+S : human-agent supporting In this scenario, the main challenge for the agent perspective is detect: 1) actions that an assistant person executes, and 2) observations of both, the person assisted and the person who attends. This is similar to ZP DH but with fragments built from HA 6= {∅, do N othing}. In this case, the level of ZP DH+S is given by Algorithm 1, and the set of extensions E with aligned goals between agent and human assistant. Proposition 3. Let OGo be a set of observations about human goals (Go) and actions (Ax) framed on an activity, captured by an agent using an activity model AT. Let G and HA be agent’s goals and its hypothetical actions. In order to provide non-conflicting assistance two properties have to be fulfilled: – PROP1: OGo ∩ G 6= ∅ – PROP2: OAx ∩ HA 6= ∅ Assistive agents and human activities 9 PROP1 and PROP2 provides coherence among human-agents actions and goals. This two properties may define a first attempt to establish consistency principles of agent-based assistance. This is a future work in our research. 4 Empirical results In this section, we present results of an empirical evaluation of different values of ZPD. The pilot test was illustrative, aimed at exploring ZPD values in a real setting, then make a qualitatively comparison with our formal approach. 4.1 Prototype and pilot evaluation The scenario selected to test our approach was framed on supporting an older adult in the activity: medication management using a smart medicines cabinet. In a smart environment8 developed at the user, interaction and knowledge man- agement research group9 , we setup the smart cabinet 10 (see Figure 2). projector smart medicines cabinet Kinect sensor 2 Text Recognition II Google API Kinect sensor 1 Gesture Argument-based Goal-based action Augmented Recognition Reasoning Reflection reality III IV Kinect sensor 3 I Medicine database projection (doses) V Local machine Fig. 2. Smart medicines cabinet using argument-based reasoning and an augmented reality projection. I) Gesture recognition using three Kinect cameras, one for client body capture, another for assistant personal gesture recognition, last one (Kinect sensor 2) on the top of the cabinet to recognize text from medicines boxes; II) Google API for text recognition; III) common-sense reasoning; IV) goal-based action reflection to consider human side; V) database containing doses and timing of pill intake. 8 360 degrees view of the lab: https://goo.gl/maps/rq3YiF1c5An 9 Computing Science department Umeå University- Sweden 10 Due the lack of space, we briefly describe the smart cabinet prototype which is connected to our agent-based platform 10 Guerrero et al. Architecture summary: Our prototype consists of five main parts: 1) ges- tures recognition: obtaining observations from individuals using Kinect cameras; 2) text recognition using another Kinect camera with Google API text recogni- tion (https://cloud.google.com/vision); 3) argument-based reasoning: the main agent-based mechanism of common sense reasoning; 4) goal-based action reflection generating an augmented reality feedback: a module to generate sup- port indications as projections in the smart environment; and 5) a database of medicine doses to obtain appropriate messages11 . We use three 3D cameras to capture: 1) observations of an individual that needs help in a physical activ- ity; 2) observations of the smart environment, including a supporting person; and 3) information of the handle gestures of medicine manipulation. A central computer was connected to the cameras, processing the information in real-time analyzing gestures of individuals as observations for the agent. The agent plat- form (JaCaMo) was used to build the agent. An argumentation process was used using an argumentation library previously developed (see [9]). An agent try to update/trigger its plan every frame time that a pre-defined gesture of the 3D camera. Those pre-defined gestures were defined in Stage 1 and Stage 2 with users and experts. Pilot evaluation setting summary: This pilot study recruited five par- ticipants, see Table X. All of the participants had technology experience using computers . The procedure comprised three stages: 1) baseline interview (sub- Type of user Age Sex Participation TA-1 67 M Stage 1 - Stage 3 TA-2 71 F Stage 1 - Stage 3 TA-3 57 F Stage 1 - - S-1 50 F - Stage 2 - S-2 25 F - - Stage 3 Table 1. Pilot setting jects: TA-1, TA-2, TA-3 + S-1, S-2); 2) interview with a nurse (S-2); and 3) prototype evaluation (subjects: TA-1, TA-2 + S-2). For a lack of space we only describe the third stage in which participants interact the smart platform. In the third stage, TA-1 participated the evaluation in his home and the other two participates evaluated the system. They were asked to read the instruc- tion message from the augmented reality projection and then, distribute three medications with different prescriptions by using the system. The Assessment of Autonomy in Internet-Mediated Activity protocol (AAIMA) [16] was used to evaluate ZPD. A comparison between our agent ZPD and AAIMA results were obtained. In Table 2 results of ZPD-S and ZPD-H were obtained. 5 Discussion and conclusions Formal argumentation can be seen as a process consisting of the following 11 Sources and documentation of the prototype can be found in https://github.com/ esteban-g Assistive agents and human activities 11 Table 2. AAIMA Protocol for assessing medication management Our main contribution in this paper is in general, a formal understanding of the interplay among an assistive agent-based software, a person to be assisted and a caregiver. Moreover, as far as we know, this is a first attempt to formalize the behavior of rational agents using formal argumentation theory, in four scenarios: I) independent activity execution, which resembles the so-called zone of current development ZCD [12]; II) ZP DH which is a set of potential activities that a person can execute with the support of another person (e.g. a therapist); III) the ZP DS representing a potential activities when a person is supported by a software agent; and IV) the ZP DH+S which is the set of activities that a person may be able to perform when is supported by another person and a software agent. ). We propose two properties (Proposition 3) that software agents should follow if their goals are linked to human goals. The relevance and impact of these properties not only covers agents based on formal argumentation theory, but other approaches, such as those based on the Belief Desire Intention model [2]. We propose an algorithm to integrate individual’s information (the AT model Definition 1) into the final decision-making process of an agent. This mechanism captured in Algorithm 1, resembles a process of “reflection” which in humans is a re-consideration of actions and goals given some other parameters. In fact, our reflection mechanism maybe seen as an action-filtering process with the human- in-the-loop12 . We also analyze different outputs of Algorithm 1 considering two groups of argumentation semantics (Propositions 1 and 2). 12 A concept to integrate human information in cyber-physical systems [22] 12 Guerrero et al. We evaluate our approach in a three stages pilot study using a scenario of medication management as a complex activity. In this regard, we conducted an experiment with older adults and practitioners to evaluate such activity. We developed a prototype platform using augmented reality projecting assistive messages about medication when a person required some support. For lack of space, we did not fully report in this paper, the full functioning of the platform neither the process of co-design and expert feedback. References 1. Aljaafreh, A.L., Lantolf, J.P.: Negative feedback as regulation and second lan- guage learning in the zone of proximal development. The Modern Language Journal 78(4), 465–483 (1994) 2. Bratman, M.: Intention, plans, and practical reason. Harvard University Press (1987) 3. Carrera, Á., Iglesias, C.A.: A systematic review of argumentation techniques for multi-agent systems research. Artificial Intelligence Review 44(4), 509–535 (2015) 4. Chaiklin, S.: The zone of proximal development in vygotskys analysis of learning and instruction. Vygotskys educational theory in cultural context 1, 39–64 (2003) 5. Dix, J.: A classification theory of semantics of normal logic programs: II. weak properties. Fundam. Inform. 22(3), 257–288 (1995) 6. Dung, P.M.: On the acceptability of arguments and its fundamental role in non- monotonic reasoning, logic programming and n-person games. Artificial Intelligence 77(2), 321–357 (1995) 7. Gelfond, M., Lifschitz, V.: Classical negation in logic programs and disjunctive databases. New generation computing 9(3-4), 365–385 (1991) 8. Guerrero, E., Nieves, J.C., Lindgren, H.: Ali: An assisted living system for persons with mild cognitive impairment. In: Computer-Based Medical Systems (CBMS), 2013 IEEE 26th International Symposium on. pp. 526–527. IEEE (2013) 9. Guerrero, E., Nieves, J.C., Lindgren, H.: Semantic-based construction of argu- ments: An answer set programming approach. International Journal of Approxi- mate Reasoning 64, 54 – 74 (2015) 10. Guerrero, E., Nieves, J.C., Sandlund, M., Lindgren, H.: Activity qualifiers in an argumentation framework as instruments for agents when evaluating human activ- ity. In: Advances in Practical Applications of Scalable Multi-agent Systems. The PAAMS Collection, pp. 133–144. Springer (2016) 11. Guerrero, E., Nieves, J.C., Sandlund, M., Lindgren, H.: Activity qualifiers using an argument-based construction. Knowledge and Information Systems 54(3), 633–658 (2018) 12. Harland, T.: Vygotsky’s zone of proximal development and problem-based learning: Linking a theoretical concept with practice through action research. Teaching in higher education 8(2), 263–272 (2003) 13. Hedegaard, M.: The zone of proximal development as basis for instruction. In: An introduction to Vygotsky, pp. 183–207. Routledge (2002) 14. Kaptelinin, V., Nardi, B.A.: Acting with Technology: Activity Theory and Inter- action Design. Acting with Technology, MIT Press (2006) 15. Leontyev, A.N.: Activity and consciousness. Moscow: Personality (1974) Assistive agents and human activities 13 16. Lindgren, H.: Personalisation of internet-mediated activity support systems in the rehabilitation of older adults–a pilot study. proc Personalisation for e-Health pp. 20–27 (2009) 17. Mezirow, J., et al.: How critical reflection triggers transformative learning. Foster- ing critical reflection in adulthood 1, 20 (1990) 18. Nieves, J.C., Guerrero, E., Lindgren, H.: Reasoning about human activities: an ar- gumentative approach. In: 12th Scandinavian Conference on Artificial Intelligence (SCAI 2013) (2013) 19. Organisation mondiale de la santé and World Health Organization: International Classification of Functioning, Disability and Health: ICF. Nonserial Publication, World Health Organization (2001), http://www.who.int/classifications/icf 20. Organization, W.H.: How to use the ICF: A practical manual for using the Inter- national Classification of Functioning, Disability and Health (ICF). Geneva:WHO (2013), http://www.who.int/classifications/drafticfpracticalmanual2.pdf 21. Salomon, G., Globerson, T., Guterman, E.: The computer as a zone of proximal development: Internalizing reading-related metacognitions from a reading partner. Journal of educational psychology 81(4), 620 (1989) 22. Schirner, G., Erdogmus, D., Chowdhury, K., Padir, T.: The future of human-in- the-loop cyber-physical systems. Computer 46(1), 36–45 (2013) 23. Vygotsky, L.S.: Mind in society: The development of higher psychological processes. Harvard university press (1980)