<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>The IJCAI-23 Joint Workshop on Artificial Intelligence Safety and Safe Reinforcement Learning (AISafety-SafeRL2023)</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Gabriel Pedroza</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff8">8</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Xin Cynthia Chen</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff8">8</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>José Hernández-Orallo</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff5">5</xref>
          <xref ref-type="aff" rid="aff8">8</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Xiaowei Huang</string-name>
          <email>xiaowei.huang@liverpool.ac.uk</email>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff6">6</xref>
          <xref ref-type="aff" rid="aff8">8</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Andreas</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff8">8</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Theodorou</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff4">4</xref>
          <xref ref-type="aff" rid="aff8">8</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Nikolaos Matragkas</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff8">8</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Huascar Espinoza</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff8">8</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Richard Mallah</string-name>
          <email>richard@futureoflife.org</email>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff8">8</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>John McDermid</string-name>
          <email>john.mcdermid@york.ac.uk</email>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff7">7</xref>
          <xref ref-type="aff" rid="aff8">8</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Mauricio Castillo-Effen</string-name>
          <email>mauricio.castillo-effen@lmco.com</email>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
          <xref ref-type="aff" rid="aff8">8</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>David Bossens</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff8">8</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Bettina Koenighofer</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff8">8</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Sebastian</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff8">8</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Anqi Liu</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff8">8</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>ANSYS</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff8">8</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>France gabriel.pedroza@ansys.com</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff8">8</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>ETH Zurich</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff8">8</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Switzerland xin.chen@inf.ethz.ch</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff8">8</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>CEA LIST</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff8">8</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>France n.matragkas@hull.ac.uk</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff8">8</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Bettina Koenighofer</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff8">8</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>TU Graz bettina.koenighofer@iaik.tugraz.at</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff8">8</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>, held at the 32nd International Joint Conference on Artificial</institution>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Future of Life Institute</institution>
          ,
          <country country="US">USA</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Learning</institution>
          ,
          <addr-line>AISafety-SafeRL2023</addr-line>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>Lockheed Martin, Advanced Technology Laboratories</institution>
          ,
          <addr-line>Arlington, VA</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
        <aff id="aff4">
          <label>4</label>
          <institution>Umeå University</institution>
          ,
          <country country="SE">Sweden</country>
        </aff>
        <aff id="aff5">
          <label>5</label>
          <institution>Universitat Politècnica de València</institution>
          ,
          <country country="ES">Spain</country>
        </aff>
        <aff id="aff6">
          <label>6</label>
          <institution>University of Liverpool</institution>
          ,
          <addr-line>Liverpool</addr-line>
          ,
          <country country="UK">United Kingdom</country>
        </aff>
        <aff id="aff7">
          <label>7</label>
          <institution>University of York</institution>
          ,
          <country country="UK">United Kingdom</country>
        </aff>
        <aff id="aff8">
          <label>8</label>
          <institution>We summarize the IJCAI-23 Joint Workshop on Artificial Intelligence Safety and Safe Reinforcement</institution>
        </aff>
      </contrib-group>
      <abstract>
        <p />
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>7 KDT JU, Belgium
Huascar.Espinoza@kdt-ju.europa.eu
11 David Bossens, University of Southampton</p>
      <p>davidmbossens@gmail.com
13 Sebastian Tschiatschek, University of Vienna</p>
      <p>sebastian.tschiatschek@univie.ac.at
14 Anqi Liu, Johns Hopkins University</p>
      <p>ataliu@cs.jhu.edu
1 Workshop series website: https://www.aisafetyw.org/
Copyright © 2023 for this paper by its authors. Use permitted under
Creative Commons License Attribution 4.0 International (CC
BY 4.0).
Intelligence (IJCAI-23) on August 21-22, 2023 in
Macau, China.</p>
    </sec>
    <sec id="sec-2">
      <title>Introduction</title>
      <p>Safety in Artificial Intelligence (AI) is increasingly
becoming a substantial part of AI research, deeply
intertwined with the ethical, legal and societal issues
associated with AI systems. Even if AI safety is considered
a design principle, there are varying levels of safety,
diverse sets of ethical standards and values, and varying
degrees of liability, for which we need to deal with
trade-offs or alternative solutions. These choices can only
be analyzed holistically if we integrate technological and
ethical perspectives into the engineering problem, and
consider both the theoretical and practical challenges for
AI safety. This view must cover a wide range of AI
paradigms, considering systems that are specific for a
particular application, and also those that are more general,
which may lead to unanticipated risks. We must bridge the
short-term with the long-term perspectives, idealistic goals
with pragmatic solutions, operational with policy issues,
and industry with academia, in order to build, evaluate,
deploy, operate and maintain AI-based systems that are
truly safe.</p>
      <p>Safe Reinforcement Learning (Safe RL) is a specialized
domain within the broader field of reinforcement learning
that emphasizes the importance of ensuring safety during
the learning and decision-making processes. The primary
objective of Safe RL is to develop algorithms and systems
that can learn and make decisions without causing harm to
themselves, the environment, or other entities. This
encompasses avoiding physical damage, breaches of
ethical standards, and violations of societal norms or legal
regulations. In essence, Safe RL seeks to strike a balance
between exploration and exploitation in learning, ensuring
that an RL agent doesn't take actions that could lead to
irreversible negative consequences, especially in critical
applications like aerospace, robotics, and other
safety-critical systems.</p>
      <p>The IJCAI-23 Joint Workshop on Artificial Intelligence
Safety and Safe Reinforcement Learning
(AISafety-SafeRL2023) seeks to explore new ideas in AI
safety with a particular focus on addressing the following
questions:
● What is the status of existing approaches for ensuring
AI and Machine Learning (ML) safety and what are the
gaps?
● How can we engineer trustworthy AI software
architectures?
● How can we make AI-based systems more ethically
aligned?
● What safety engineering considerations are required to
develop safe human-machine interaction?
● What AI safety considerations and experiences are
relevant from industry?
● How can we characterize or evaluate AI systems
according to their potential risks and vulnerabilities?
● How can we develop solid technical visions and new
paradigms about AI safety?
● How do metrics of capability and generality, and
trade-offs with performance, affect safety?
These are the main topics of the series of AISafety
workshops which this year have been enriched by a
particular focus on Reinforcement Learning techniques,
their challenges, solutions and perspectives. Overall, the
series aims to achieve a holistic view of AI and safety
engineering, taking ethical and legal issues into account, in
order to build trustworthy intelligent autonomous
machines.</p>
    </sec>
    <sec id="sec-3">
      <title>Program</title>
      <p>The Program Committee (PC) received 19 submissions.
Each paper was peer-reviewed by at least two PC
members, by following a single-blind reviewing process.
The committee decided to accept 10 full papers, resulting
in an overall paper acceptance rate of 52%.</p>
      <p>The AISafety-SafeRL2023 program was organized in
five thematic sessions, two keynote and three (invited)
talks. The thematic sessions followed a highly interactive
format. They were structured into short pitches and a group
debate panel slot to discuss both individual paper
contributions and shared topic issues. Three specific roles
were part of this format: session chairs, presenters and
session discussants.
● Session Chairs introduced sessions and participants.</p>
      <p>The Chair moderated sessions and plenary discussions,
monitored time, and moderated questions and
discussions from the audience.
● Presenters gave a 10-minute paper talk and participated
in the debate slot.
● Presenters gave a 10-minute paper talk and participated
in the debate slot.
● Invited speakers gave a 25-minute talk on a relevant
topic to the workshop.
● Contributed talk speakers gave a 15-minute talk on a
relevant topic to the workshop
● Session Discussants gave a critical review of the
session papers, and participated in the plenary debate.
Presentations and papers were grouped by topic as follows:</p>
      <sec id="sec-3-1">
        <title>Session 1: Robustness of AI via OoD and</title>
      </sec>
      <sec id="sec-3-2">
        <title>Unknown-Unknowns Dectection</title>
        <p>● Diffusion Denoised Smoothing for Certified and
Adversarial Robust Out Of Distribution, Nicola Franco,
Daniel Korth, Jeanette Miriam Lorenz, Karsten Roscher
and Stephan Günnemann
● Unsupervised Unknown Unknown Detection in Active
Learning, Prajit T. Rajendran, Huascar Espinoza, Agnes
Delaborde and Chokri Mraidha</p>
      </sec>
      <sec id="sec-3-3">
        <title>Session 2: AI Robustness, Adversarial Attacks and</title>
      </sec>
      <sec id="sec-3-4">
        <title>Reinforcemnt Learning</title>
        <p>● PerCBA: Persistent Clean-label Backdoor Attacks on
Semi-Supervised Graph Node Classification, Xiao
Yang, Gaolei Li, Chaofeng Zhang, Meng Han and Wu
Yang
● Distribution-restrained Softmax Loss for the Model
Robustness, Chen Li, Hao Wang, Jinzhe Jiang, Xin
Zhang, Yaqian Zhao and Weifeng Gong
● Fear Field: Adaptive constraints for safe environment
transitions in Shielded Reinforcement Learning, Haritz
Odriozola-Olalde, Nestor Arana, Arexolaleiba, Maider
Zamalloa, Jon Perez, Cerrolaza, Jokin Arozamena and
Rodríguez</p>
      </sec>
      <sec id="sec-3-5">
        <title>Session 3: AI Governance and Policy/Value Alignment</title>
        <p>● An open source perspective on AI and alignment with
the EU AI Act Diego Calanzone, Andrea Coppari,
Riccardo Tedoldi, Giulia Olivato and Carlo Casonato</p>
      </sec>
      <sec id="sec-3-6">
        <title>Session 4: SafeRL</title>
        <p>● Yanan Sui: Embodied safe optimization for the
restoration of human motor functions
● Thiago Simao: Ensuring the offline reliability and
online safety of reinforcement learning agents
● Filip Cano: Search-based Testing of Reinforcement</p>
        <p>Learning
● Martin Kurezca: Monte Carlo Tree Search with
Function Approximation for Risk-constrained Planning
and Reinforcement Learning
● Ruoqi Zhang: Risk-sensitive Actor-free Policy via</p>
        <p>Convex Optimisation
● Weiye Zhao: State-wise Constrained Policy</p>
        <p>Optimization</p>
      </sec>
      <sec id="sec-3-7">
        <title>Session 5: AI Trustworthiness, Explainability and</title>
      </sec>
      <sec id="sec-3-8">
        <title>Testing</title>
        <p>● Empirical Optimal Risk to Quantify Model
Trustworthiness for Failure Detection, Shuang Ao,
Stefan Rueger and Advaith Siddharthan
● Weight-based Semantic Testing Approach for Deep
Neural Networks, Amany Alshareef, Nicolas Berthier,
Sven Schewe and Xiaowei Huang
● AI for Safety: How to use Explainable Machine
Learning Approaches for Safety Analyses Iwo
Kurzidem, Simon Burton and Philipp
AISafety was pleased to have several additional
inspirational researchers as invited speakers:</p>
      </sec>
      <sec id="sec-3-9">
        <title>Keynotes</title>
        <p>● Paul Lukowicz, Safety risks of AI: Intelligence,</p>
        <p>Complexity and Stupidity
● François Terrier, No Trust without regulation! European
challenge on regulation, liability and standards for
trusted AI</p>
      </sec>
      <sec id="sec-3-10">
        <title>Invited Talks</title>
        <p>● Yanan Sui: Embodied safe optimization for the
restoration of human motor functions
● Thiago Simao: Ensuring the offline reliability and
online safety of reinforcement learning agents</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Acknowledgements</title>
      <p>We thank all researchers who submitted papers to AISafety
2023 and congratulate the authors whose papers were
selected for inclusion into the workshop program and
proceedings.</p>
      <p>We especially thank our distinguished PC members for
reviewing the submissions and providing useful feedback
to the authors:
● Simos Gerasimou, University of York, UK
● Jonas Nilson, NVIDIA, USA
● Brent Harrison, University of Kentucky, USA
● Alessio R. Lomuscio, Imperial College London, UK
● Brian Tse, Affiliate at University of Oxford, China
● Michael Paulitsch, Intel, Germany
● Ganesh Pai, NASA Ames Research Center, USA
● Rob Alexander, University of York, UK
● Vahid Behzadan, University of New Haven, USA
● Chokri Mraidha, CEA LIST, France
● Ke Pei, Huawei, China
● Orlando Avila-García, Arquimea Research Center,</p>
      <p>Spain
● I-Jeng Wang, Johns Hopkins University, USA
● Chris Allsopp, Frazer-Nash Consultancy, UK
● Andrea Orlandini, ISTC-CNR, Italy
● Agnes Delaborde, LNE, France
● Morayo Adedjouma, CEA LIST, France
● Rasmus Adler, Fraunhofer IESE, Germany
● Roel Dobbe, TU Delft, The Netherlands
● Vahid Hashemi, Audi, Germany
● Juliette Mattioli, Thales, France
● Bonnie W. Johnson, Naval Postgraduate School, USA
● Roman V. Yampolskiy, University of Louisville, USA
● Jan Reich, Fraunhofer IESE, Germany
● Fateh Kaakai, Thales, France
● Francesca Rossi, IBM and University of Padova, USA
● Javier Ibañez-Guzman, Renault, France
● Jérémie Guiochet, LAAS-CNRS, France
● Raja Chatila, Sorbonne University, France
● François Terrier, CEA LIST, France
● Mehrdad Saadatmand, RISE Research Institutes of
Sweden, Sweden
● Alec Banks, Defence Science and Technology</p>
      <p>Laboratory, UK
● Roman Nagy, Argo AI, Germany
● Nathalie Baracaldo, IBM Research, USA
● Toshihiro Nakae, DENSO Corporation, Japan
● Gereon Weiss, Fraunhofer IKS, Germany
● Philippa Ryan Conmy, Adelard, UK
● Stefan Kugele, Technische Hochschule Ingolstadt,</p>
      <p>Germany
● Colin Paterson, University of York, UK
● Davide Bacciu, Università di Pisa, Italy
● Timo Sämann, Valeo, Germany
● Sylvie Putot, Ecole Polytechnique, France
● John Burden, University of Cambridge, UK
● Sandeep Neema, DARPA, USA
● Fredrik Heintz, Linköping University, Sweden
● Simon Fürst, BMW Group, Germany
● Mario Gleirscher, University of Bremen, Germany
● Mandar Pitale, NVIDIA, USA
● Leon Kester, TNO, The Netherlands
● Bernhard Kaiser, ANSYS, Germany</p>
      <p>organization for
framework for</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <article-title>Finally we thank the IJCAI-23 providing an excellent AISafety-SafeRL2023.</article-title>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>