<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Explainability and self-disclosure for robot ethical introspection</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Valeria Seidita</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Antonio Chella</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Dipartimento di Ingegneria, Universitá degli Studi di Palermo</institution>
          ,
          <country country="IT">Italy</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>ICAR-CNR National Research Council</institution>
          ,
          <addr-line>Palermo</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Human-robot or human-AI interaction systems require a high degree of autonomy, proactivity, and adaptvity. The decisions that intelligent systems must make are highly dependent on the application context and trust is an essential element in task assignment. Explainability and ethical introspection capabilities are important in building trust and understanding in artificial processes. In this paper, we present our ongoing work aimed at equipping robots with ethical introspection capabilities when interacting with humans by designing and implementing explainable and self-disclosure capabilities. Using a computational model of ethical introspection that incorporates theories of psychology, ethics, and AI, we built robots that examine and reflect on their actions to evaluate and validate them. We use the Belief-Desire-Intention (BDI) agent paradigm and related programming languages along with the speech act mechanism to improve and extend the robot's ethical values to better guide its decision-making process and the impact it has on humans.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Ethics</kwd>
        <kwd>Self-Disclosure</kwd>
        <kwd>Ethical Introspection</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>Robots that emulate human behavior and assist individuals in their daily lives have long been a
dream. A pioneer of this concept was Isaac Asimov, whose Three Laws of Robotics encapsulated
the notion of robots interacting with humans. Asimov’s three laws provide a valuable foundation;
however, they are not suficient when addressing contexts where robots must exhibit autonomy,
proactivity, and self-adaptation.</p>
      <p>
        The scenarios under consideration pertain to human-robot team interaction (HRTI) [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ][
        <xref ref-type="bibr" rid="ref2">2</xref>
        ][
        <xref ref-type="bibr" rid="ref3">3</xref>
        ].
Within HRTI, the focal point is the collaboration between teams of humans and robots to
achieve common goals. Irrespective of the nature of the task, whether it is purely social, such as
providing companionship to an elderly individual, or high-risk, as in military settings, humans
and robots must exchange information regarding the objectives, mission, their respective
limitations, capabilities, and the work environment.
      </p>
      <p>Typically, drawing upon their knowledge of the aforementioned elements, each team member
selects an action to fulfill their mission. Nevertheless, this process is not strictly individualistic.
The choice of action is also contingent upon the presence and actions of other team members,
their knowledge, and competencies. Each team member must possess the ability to comprehend
and anticipate the actions of their peers, as well as make decisions regarding which actions
in the plan they can execute. Ultimately, each team member must decide which actions to
undertake personally and which to delegate.</p>
      <p>Several factors come into play in this decision-making process, with trust in one another
being a pivotal element. One of the key factors that engenders or enhances trust in a fellow
team member is the ability to expound on the rationale behind their actions. In a complex
environment characterized by high dynamism, leading to uncertainty and decision-making
challenges, such as in healthcare or military contexts, the selection of the optimal action is
contingent upon the ability to evaluate the outcomes of actions in relation to predefined goals
and conditions.</p>
      <p>
        In some of our previous work, we have explored strategies for enhancing a robot’s
decisionmaking abilities through the concept of ‘anticipation’ [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ][
        <xref ref-type="bibr" rid="ref5">5</xref>
        ][6]. We have developed a tool that
allows the robot to transparently present its decision-making process. The robot selects an
action given at design time and simulates the outcome before execution. If the simulated result
aligns with the post-conditions of the goal, the mission is deemed successful; otherwise, the
robot must opt for an alternative action. Concurrently, the robot provides an explanation to its
human companion regarding its actions.
      </p>
      <p>In any complex scenario involving the autonomy of a robot in the interactive domain,
considerations concerning ‘ethics’ also come to the forefront. For instance, the potential invasion or
violation of privacy when employing robots to assist patients necessitates that robots adhere
to two fundamental principles: (i) making decisions aligned with the ethical standards of the
society in which they operate, and (ii) articulating the rationale behind their actions to build
human trust and influence human decisions. For example, in the context of a robot assisting
a medical doctor, the robot can suggest an action to the doctor, justifying it from an ethical
perspective, thereby stimulating thoughtful consideration. This process supports not only
eficient decision-making but also provokes reflection in the human user.</p>
      <p>
        Our endeavor revolves around the exploration of the roles of introspection and self-disclosure
in ethical deliberation and social influence. In this paper, we present an initial hypothesis on
how to formulate a computational model of ethical introspection in robots, building upon prior
work on anticipation and trust [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ][7][8].
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. Enabling Ethical Reasoning in Artificial Agents: a Model</title>
      <p>Ethical introspection is an activity that focuses on mental states and mental processes as they
occur or have just occurred. Our concept of endowing a robot with human-driven ethical
introspection is to integrate it with the robot’s self-awareness. Self-awareness implies a critical
examination of one’s own actions, intending to comprehend the ethical ramifications of those
actions. In our approach, we draw inspiration from psychological, philosophical, and
neuroscientific studies on ethical introspection, with Sullins serving as primary references [ 9][10][11].
Sullins proposed a theoretical framework based on the concept of ‘artificial phronesis,’ which
pertains to human practical reasoning abilities and the virtue of rational thinking. ‘Phronesis’
represents philosophical wisdom in practical ethics and serves as the foundation for constructing
machines capable of emulating human moral and practical reasoning.</p>
      <p>Sullins asserts that ethical decision-making cannot always be reduced to a simple ’if situation
(x), then action (y)’ formula. Ethical decisions are predicated on the agent’s habits, practices,
and a comprehensive analysis of all elements of an ethical problem and potential consequences
of actions. The framework proposed by Sullins involves two interacting agents: one for problem
classification and one for case analysis. The former examines the situation to extract critical
features, such as context and urgency. The latter collaborates in searching for analogous ethical
cases within specific repositories and online sources to enhance the analysis.</p>
      <p>Our objective is to fuse computational models of introspection, self-awareness, justification,
and anticipation with mechanisms for ethical reasoning. In Figure 1, we present an integration of
Sullins’ model with our work spanning several years. In structuring the decision-making process,
we also take internal states into account. To this end, we have conducted experiments and
opted to employ BDI (Belief-Desire-Intention) agent technology [12][13], utilizing the JaCaMo
implementation framework [14][15]. BDI agents aptly embody practical thinking, encompassing
the process of agents determining how to translate their intentions into actionable steps within
their environment.</p>
      <p>Practical reasoning encompasses actions such as planning, resource allocation, and action
sequence management, all of which consider agent constraints, environmental conditions, and
potential consequences. This aligns well with the underlying concept of ‘artificial phronesis.’
Ethical reasoning can involve multiple agents, including humans. In Figure 1, we depict three
artificial agents: the problem classification agent (PC), the case analysis agent (CA), and the
planner agent (P). Upon encountering an ethical quandary, the first two agents engage in
continuous interaction at the outset of the reasoning process. The initial step involves inspecting
the environment and situation to formulate the problem and context. The CA leverages this
context to seek analogous situations, potentially engaging with humans. A novel scenario is
then created and analyzed, potentially prompting a reevaluation of the problem. Eventually,
the CA and PC concur on the context and contact the planner, who selects a plan and executes
actions, which may also be simulated. These actions impact the environment, initiating a new
situation and reinitiating the cycle. Actions receive positive or negative feedback, signifying
their contribution to or deviation from the desired goal. This cycle embodies a
sensing-decisionaction process, with key points focused on ethical introspection through self-disclosure.</p>
      <p>Figure 2 illustrates the reasoning cycle of each agent of Figure 1, highlighting the points at
which we introduce extensions to, in order: (i) generate a self-model, (ii) explain actions, and (iii)
facilitate ethical introspection. Our goal is to adapt the reasoning cycle of the agents featured
in Figure 1, as depicted in Figure 2. To create an agent with ethical introspection capabilities,
we prioritize the need for explanations, even before it needs to construct a self-model.</p>
      <p>From a technical standpoint, actions can be explained as a function of the couple &lt;belief,
capabilities&gt;. The agent decomposes the plan for mission execution into actions, closely
associated with its knowledge of those actions and its ability to execute them. Consequently,
the agent maintains a self-model and justifies action outcomes. We propose that, for each
action, a rehearsal function can be defined, closely tied to self. The rehearsal function facilitates
starten(svitiruoantmioenn)t APnCalyse Situation
context</p>
      <p>Agreement Select Plan</p>
      <p>P</p>
      <p>Simulate</p>
      <p>Act
end
repository of
problems
humans
new context
(situation)
Search
CA</p>
      <p>new formalized problem
feedback with self, allowing the agent to reassess and adjust its knowledge base, beliefs, and
decision-making process. This adjustment is based on observations of the efects of actions and
new generated beliefs. When actions encompass ethical norms, the refined beliefs result from
ethical introspection.</p>
      <p>The practice of rehearsing has been implemented in previous works through the mechanism
of speech acts [16]. A speech act [17][18] embodies the communicative action of the agent
in which it expresses its actions, thus enabling self-revelation capabilities within an agent or
agent system. The efect on the thinking process is indirect, indeed there is a change in the
agents’ knowledge as the result of speech acts. Speech acts along with agent technology allow
to autonomously produce self-disclosure and then explainability.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Conclusion</title>
      <p>This paper presents our ongoing research focused on the implementation of an ethical
introspection model within robots endowed with self-awareness. Our research endeavors encompass
computational models of introspection, self-awareness, justification, and anticipation,
integrated with processes conducive to ethical deliberation. The principal objective underlying this
research has been the facilitation of ethics within the decision-making framework of artificial
agents. Initial progress was achieved through the integration of Sullins’ model with our prior
advancements, utilizing a Belief-Desire-Intention (BDI) agent-based paradigm. This model
inherently captures the agent’s practical reasoning process, placing significant emphasis on
aspects such as planning, resource allocation, and the management of action sequences.</p>
      <p>Furthermore, we have introduced an extended facet to the model, integrating ethical practice
through the medium of speech acts. This augmentation empowers the agent to articulate
and elucidate its actions. This communicative process catalyzes ethical evaluation and
selfawareness, with the overarching aim of augmenting the decision-making capabilities and ethical
conduct of these agents.</p>
      <p>In the future, our research trajectory will involve further refinement of the model, drawing
inspiration from in-depth interdisciplinary studies spanning the realms of psychology, philosophy,
and neuroscience, all of which delve into the domains of ethics and introspection.</p>
      <p>Collectively, our work lays the foundation for an in-depth exploration of ethics and
introspection within the domain of autonomous robots, holding substantial potential for application in
diverse domains, including healthcare and assistive robotics, among others. The incorporation
of ethical models into robots represents a pivotal stride towards the responsible utilization of
technology in complex and dynamic contexts, thereby enhancing the synergy between machines
and human agents.</p>
    </sec>
    <sec id="sec-4">
      <title>Acknowledgments</title>
      <p>International Exchanges 2022 Cost Share (Italy only) IEC\R2\222031 - Joint Research Program
on Assistive Robots using Theory of Mind
[6] V. Seidita, C. Diliberto, P. Zanardi, A. Chella, F. Lanza, Inside the robot’s mind during
human-robot interaction, in: 7th International Workshop on Artificial Intelligence and
Cognition, AIC 2019, volume 2483, CEUR-WS, 2019, pp. 54–67.
[7] A. Pipitone, A. Geraci, A. D’ Amico, V. Seidita, A. Chella, Robot’ s inner speech efects on
human trust and anthropomorphism, International Journal of Social Robotics (2023) 1–13.
[8] A. Chella, A. Pipitone, F. Lanza, V. Seidita, Toward virtuous machines: When ethics meets
robotics, in: Ethics in Research: Principles and Practical Considerations, Springer, 2023,
pp. 81–91.
[9] J. P. Sullins, The role of consciousness and artificial phron e¯sis in ai ethical reasoning., in:</p>
      <p>AAAI spring symposium: towards conscious AI systems, 2019.
[10] J. P. Sullins, Artificial phronesis, Science, Technology, and Virtues: Contemporary</p>
      <p>Perspectives (2021) 136.
[11] A. Chella, A. Pipitone, J. P. Sullins, Competent moral reasoning in robot applications:
Inner dialog as a step towards artificial phronesis., in: Trolley Crash: Approaching Key
Metrics for Ethical AI Practitioners, Researchers, and Policy Makers, 2024.
[12] A. S. Rao, M. P. Georgef, et al., BDI agents: from theory to practice., in: Icmas, volume 95,
1995, pp. 312–319.
[13] L. De Silva, F. R. Meneguzzi, B. Logan, BDI agent architectures: A survey, in: Proceedings
of the 29th International Joint Conference on Artificial Intelligence (IJCAI), 2020, Japão.,
2020.
[14] O. Boissier, R. H. Bordini, J. F. Hübner, A. Ricci, A. Santi, Multi-agent oriented programming
with jacamo, Science of Computer Programming 78 (2013) 747–761.
[15] R. H. Bordini, J. F. Hübner, M. Wooldridge, Programming multi-agent systems in
AgentSpeak using Jason, John Wiley &amp; Sons, 2007.
[16] V. Seidita, A. M. P. Sabella, F. Lanza, A. Chella, Agent talks about itself: an implementation
using jason, cartago and speech acts, Intelligenza Artificiale 17 (2023) 7–18.
[17] J. Searle, J. R. Searle, Speech acts: An essay in the philosophy of language, volume 626,</p>
      <p>Cambridge university press, 1969.
[18] J. L. Austin, How to do things with words, volume 88, Oxford university press, 1975.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>L. Mingyue</given-names>
            <surname>Ma</surname>
          </string-name>
          , T. Fong,
          <string-name>
            <given-names>M. J.</given-names>
            <surname>Micire</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y. K.</given-names>
            <surname>Kim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Feigh</surname>
          </string-name>
          ,
          <article-title>Human-robot teaming: Concepts and components for design</article-title>
          ,
          <source>in: Field and Service Robotics: Results of the 11th International Conference</source>
          , Springer,
          <year>2018</year>
          , pp.
          <fpage>649</fpage>
          -
          <lpage>663</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>T.</given-names>
            <surname>Chakraborti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Kambhampati</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Scheutz</surname>
          </string-name>
          ,
          <string-name>
            <surname>Y. Zhang,</surname>
          </string-name>
          <article-title>AI challenges in human-robot cognitive teaming</article-title>
          ,
          <source>arXiv preprint arXiv:1707.04775</source>
          (
          <year>2017</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>A.</given-names>
            <surname>Chella</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Lanza</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Seidita</surname>
          </string-name>
          ,
          <article-title>A cognitive architecture for human-robot teaming interaction</article-title>
          ,
          <source>in: Proceedings of the 6th International Workshop on Artificial Intelligence and Cognition</source>
          , Palermo,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>C.</given-names>
            <surname>Castelfranchi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Chella</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Falcone</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Lanza</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Seidita</surname>
          </string-name>
          , et al.,
          <article-title>Endowing robots with self-modeling abilities for trustful human-robot interactions</article-title>
          ,
          <source>in: CEUR WORKSHOP PROCEEDINGS</source>
          , volume
          <volume>2404</volume>
          ,
          <string-name>
            <surname>CEUR-WS</surname>
          </string-name>
          ,
          <year>2019</year>
          , pp.
          <fpage>22</fpage>
          -
          <lpage>28</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>A.</given-names>
            <surname>Chella</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Lanza</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Seidita</surname>
          </string-name>
          ,
          <article-title>Decision Process in Human-Agent Interaction: Extending Jason Reasoning Cycle</article-title>
          ,
          <source>in: Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)</source>
          , volume
          <volume>11375</volume>
          LNAI, Springer, Cham,
          <year>2019</year>
          , pp.
          <fpage>320</fpage>
          -
          <lpage>339</lpage>
          . doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>030</fpage>
          -25693-7_
          <fpage>17</fpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>