<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>The AAAI-22 Workshop on Artificial Intelligence Safety (SafeAI 2022)</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Gabriel Pedroza</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>José Hernández-Orallo</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Xin Cynthia Chen</string-name>
          <xref ref-type="aff" rid="aff4">4</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Xiaowei Huang</string-name>
          <email>xiaowei.huang@liverpool.ac.uk</email>
          <xref ref-type="aff" rid="aff6">6</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Huáscar</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Espinoza</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Mauricio Castillo-Effen</string-name>
          <email>mauricio.castillo-effen@lmco.com</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>John McDermid</string-name>
          <email>john.mcdermid@york.ac.uk</email>
          <xref ref-type="aff" rid="aff5">5</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Richard Mallah</string-name>
          <email>richard@futureoflife.org</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Seán S. ÓhÉigeartaigh</string-name>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Université Paris-Saclay</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>CEA LIST</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>France gabriel.pedroza@cea.fr</string-name>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Future of Life Institute</institution>
          ,
          <country country="US">USA</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Lockheed Martin, Advanced Technology Laboratories</institution>
          ,
          <addr-line>Arlington, VA</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Universitat Politècnica de València</institution>
          ,
          <country country="ES">Spain</country>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>University of Cambridge</institution>
          ,
          <addr-line>Cambridge</addr-line>
          ,
          <country country="UK">United Kingdom</country>
        </aff>
        <aff id="aff4">
          <label>4</label>
          <institution>University of Hong Kong</institution>
          ,
          <country country="CN">China</country>
        </aff>
        <aff id="aff5">
          <label>5</label>
          <institution>University of Liverpool</institution>
          ,
          <addr-line>Liverpool</addr-line>
          ,
          <country country="UK">United Kingdom</country>
        </aff>
        <aff id="aff6">
          <label>6</label>
          <institution>University of York</institution>
          ,
          <country country="UK">United Kingdom</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>We summarize the AAAI-22 Workshop on Artificial Intelligence Safety (SafeAI 2022)1, virtually held at the Thirty-Sixth AAAI Conference on Artificial Intelligence on February 28.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>5 ECSEL JU, Belgium</title>
      <p>Huascar.Espinoza@ecsel.europa.eu</p>
      <sec id="sec-1-1">
        <title>Introduction</title>
        <p>Safety in Artificial Intelligence (AI) is increasingly
becoming a substantial part of AI research, deeply
intertwined with the ethical, legal and societal issues
associated with AI systems. Even if AI safety is considered
a design principle, there are varying levels of safety,
diverse sets of ethical standards and values, and varying
1 Workshop series website: http://safeaiw.org/
Copyright © 2022 for this paper by its authors. Use permitted under
Creative Commons License Attribution 4.0 International (CC
BY 4.0).
degrees of liability, for which we need to deal with
trade-offs or alternative solutions. These choices can only
be analyzed holistically if we integrate technological and
ethical perspectives into the engineering problem, and
consider both the theoretical and practical challenges for
AI safety. This view must cover a wide range of AI
paradigms, considering systems that are specific for a
particular application, and also those that are more general,
which may lead to unanticipated risks. We must bridge the
short-term with the long-term perspectives, idealistic goals
with pragmatic solutions, operational with policy issues,
and industry with academia, in order to build, evaluate,
deploy, operate and maintain AI-based systems that are
truly safe.</p>
        <p>The AAAI-22 Workshop on Artificial Intelligence
Safety (SafeAI 2022) seeks to explore new ideas in AI
safety with a particular focus on addressing the following
questions:
● What is the status of existing approaches for ensuring
AI and Machine Learning (ML) safety and what are the
gaps?
● How can we engineer trustworthy AI software
architectures?
● How can we make AI-based systems more ethically
aligned?
● What safety engineering considerations are required to
develop safe human-machine interaction?
● What AI safety considerations and experiences are
relevant from industry?
● How can we characterize or evaluate AI systems
according to their potential risks and vulnerabilities?
● How can we develop solid technical visions and new
paradigms about AI safety?
● How do metrics of capability and generality, and
trade-offs with performance, affect safety?
These are the main topics of the series of SafeAI
workshops. They aim to achieve a holistic view of AI and
safety engineering, taking ethical and legal issues into
account, in order to build trustworthy intelligent
autonomous machines. The first edition of SafeAI was held
in January 27, 2019, in Honolulu, Hawaii (USA) as part of
the Thirty-Third AAAI Conference on Artificial
Intelligence (AAAI-19), and the second edition was held in
February 7, 2020 in New York City (USA) also as part of
AAAI. This fourth edition was held online (because of the
COVID-19 situation) at the Thirty-Sixth AAAI Conference
on Artificial Intelligence on February 28, virtually.</p>
      </sec>
      <sec id="sec-1-2">
        <title>Program</title>
        <p>The Program Committee (PC) received 53 submissions.
Each paper was peer-reviewed by at least two PC
members, by following a single-blind reviewing process.
The committee decided to accept 18 full papers, 3 talks and
12 posters, resulting in a full-paper acceptance rate of
34.0% and an overall acceptance rate of 62.3%.</p>
        <p>The SafeAI 2022 program was organized in five
thematic sessions, two keynotes and three (invited) talks.</p>
        <p>The thematic sessions followed a highly interactive
format. They were structured into short pitches and a group
debate panel slot to discuss both individual paper
contributions and shared topic issues. Three specific roles
were part of this format: session chairs, presenters and
session discussants.
● Session Chairs introduced sessions and participants. The</p>
        <p>Chair moderated sessions and plenary discussions,
monitored time, and moderated questions and
discussions from the audience.
● Presenters gave a 10 minutes paper talk and participated
in the debate slot.
● Session Discussants gave a critical review of the session
papers, and participated in the plenary debate.</p>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>Papers were grouped by topic as follows:</title>
      <sec id="sec-2-1">
        <title>Session 1: Bias, Fairness and Value Alignment</title>
        <p>● The Problem of Behaviour and Preference Manipulation
in AI Systems, Hal Ashton and Matija Franklin.
● IFBiD: Inference-Free Bias Detection, Ignacio Serna,
Daniel DeAlcala, Aythami Morales Moreno, Julian
Fierrez and Javier Ortega-Garcia.
● Blackbox Post-Processing for Multiclass Fairness,</p>
        <p>Preston Putzel and Scott Lee.</p>
      </sec>
      <sec id="sec-2-2">
        <title>Session 2: Interpretability and Accountability</title>
        <p>● A Gray Box Model for Characterizing Driver Behavior,
Soyeon Jung, Ransalu Senanayake and Mykel
Kochenderfer.
● Defining and Identifying the Legal Culpability of Side</p>
        <p>Effects using Causal Graphs, Hal Ashton.</p>
      </sec>
      <sec id="sec-2-3">
        <title>Session 3: Robustness and Uncertainty</title>
        <p>● Efficient Adversarial Sequence Generation for RNN
with Symbolic Weighted Finite Automata, Mingjun Ma,
Dehui Du, Yuanhao Liu, Yanyun Wang and Yiyang Li.
● A Study on Mitigating Hard Boundaries of
Decision-Tree-based Uncertainty Estimates for AI
Models, Pascal Gerber, Lisa Jöckel and Michael Kläs.
● Quantifying the Importance of Latent Features in Neural
Networks, Amany Alshareef, Nicolas Berthier, Sven
Schewe and Xiaowei Huang.
● Maximum Likelihood Uncertainty Estimation:
Robustness to Outliers, Deebul Nair, Nico
Hochgeschwender and Miguel Olivares-Mendez.</p>
      </sec>
      <sec id="sec-2-4">
        <title>Session 4: Safe Reinforcement Learning</title>
        <p>● Reinforcement Learning With Imperfect Safety
Constraints, Jin Woo Ro, Gerald Lüttgen and Diedrich
Wolter.
● Do Androids Dream of Electric Fences? Safety-Aware
Reinforcement Learning with Latent Shielding, Peter
He, Borja Leon and Francesco Belardinelli.
● HiSaRL: A Hierarchical Framework for Safe
Reinforcement Learning, Zikang Xiong, Ishika Agarwal
and Suresh Jagannathan.
● A Game-Theoretic Perspective on Risk-Sensitive
Reinforcement Learning, Mathieu Godbout, Maxime
Heuillet, Sharath Chandra Raparthy, Rupali Bhati and
Audrey Durand.</p>
      </sec>
      <sec id="sec-2-5">
        <title>Session 5: AI Testing and Assessment</title>
        <p>
          ● Beyond Test Accuracy: The Effects of Model
Compression on CNNs, Adrian Schwaiger, Kristian
Schwienbacher and Karsten Roscher.
● Differential Assessment of Black-Box AI Agents,
Rashmeet Kaur Nayyar, Pulkit Verma and Siddharth
Srivastava.
          <xref ref-type="bibr" rid="ref1">(Note that this paper is out of the
proceedings of SafeAI2022.)</xref>
          ● Using Adaptive Stress Testing to Identify Paths to
Ethical Dilemmas in Autonomous Systems, Ann-Katrin
Reuel, Mark Koren, Anthony Corso and Mykel J.
        </p>
        <p>Kochenderfer.</p>
        <p>SafeAI was pleased to have several additional inspirational
researchers as invited speakers:</p>
      </sec>
      <sec id="sec-2-6">
        <title>Keynote</title>
        <p>● Matthew Dwyer (University of Virginia),</p>
        <p>Distribution-aware Test Adequacy for Neural Networks
● Ganesh Pai (KBR / NASA Ames Research Center),
Towards Certification of Machine Learning in
Aeronautical Applications</p>
      </sec>
      <sec id="sec-2-7">
        <title>Invited Talks</title>
        <p>● Shiri Dori-Hacohen (University of Connecticut),</p>
        <p>Quantifying Misalignment Between Agents
● Roel Dobbe (TU Delft), A System Safety Perspective
for Developing and Governing Artificial Intelligence
● Bonnie W. Johnson (Naval Postgraduate School), Safety
in AI-Enabled Warfare Decision Aids</p>
      </sec>
      <sec id="sec-2-8">
        <title>Posters</title>
        <p>
          Posters were presented with 2-minute pitches. Most posters
have also been included as short papers within this volume.
● Near-Term AI as an Existential Risk Factor, Ben
Bucknall and Shiri Dori-Hacohen.
          <xref ref-type="bibr" rid="ref1">(Note that this paper
is out of the proceedings of SafeAI2022.)</xref>
          ● The Dilemma Between Data Transformations and
Adversarial Robustness for Time Series Application
Systems, Sheila Alemany and Niki Pissinou.
● Interpretable Local Tree Surrogate Policies, John Mern,
Sidhart Krishnan, Anil Yildiz, Kyle Hatch and Mykel J.
        </p>
        <p>Kochenderfer.
● Oases of Cooperation: An Empirical Evaluation of
Reinforcement Learning in the Iterated Prisoner’s
Dilemma, Peter Barnett and John Burden.
● Leveraging Multi-task Learning for Umambiguous and
Flexible Deep Neural Network Watermarking, Fangqi
Li, Lei Yang, Shilin Wang and Alan Wee-Chung Liew.
● Human-in-the-loop Learning for Safe Exploration
through Anomaly Prediction and Intervention, Prajit T
Rajendran, Huascar Espinoza, Agnes Delaborde and
Chokri Mraidha.
● Safety Aware Reinforcement Learning by Identifying
Comprehensible Constraints in Expert Demonstrations,
Leopold Müller, Lars Böcking and Michael Färber.
● Combining Data-Driven and Knowledge-Based AI
Paradigms for Engineering AI-Based Safety-Critical
Systems, Juliette Mattioli, Gabriel Pedroza, Souhaiel
Khalfaoui and Bertrand Leroy.
● Is it all a cluster game? – Exploring Out-of-Distribution
Detection based on Clustering in the Embedding Space,
Poulami Sinhamahapatra, Rajat Koner, Karsten Roscher
and Stephan Günnemann.
● A Practical Overview of Safety Concerns and
Mitigation Methods for Visual Deep Learning
Algorithms, Saeed Bakhshi Germi and Esa Rahtu.
● Comparing Vision Transformers and Convolutional Nets
for Safety Critical Systems, Michal Filipiuk and Vasu
Singh.
● A Framework to Argue Quantitative Safety Targets in
Assurance Cases for AI/ML Components Combining
Design and Runtime Safety Measures, Michael Klaes,
Lisa Jöckel, Rasmus Adler and Jan Reich. (Note that
this paper is out of the proceedings of SafeAI2022.)</p>
      </sec>
      <sec id="sec-2-9">
        <title>Special Sessions</title>
        <p>● EnnCore: End-to-End Conceptual Guarding of Neural
Architectures, Edoardo Manino, Danilo Carvalho, Yi
Dong, Julia Rozanova, Xidan Song, Andre Freitas,
Gavin Brown, Mikel Luján, Xiaowei Huang, and Lucas
Cordeiro.
● The wall of safety for AI: approaches in the
Confiance.ai program, Bertrand Braunschweig, François
Terrier, and Rodolphe Gelin.</p>
        <sec id="sec-2-9-1">
          <title>Acknowledgements</title>
          <p>We thank all researchers who submitted papers to SafeAI
2022 and congratulate the authors whose papers and
posters were selected for inclusion into the workshop
program and proceedings.</p>
          <p>We especially thank our distinguished PC members for
reviewing the submissions and providing useful feedback
to the authors:
● Stuart Russell, UC Berkeley, USA
● Raja Chatila, Sorbonne University, France
● Francesca Rossi, IBM and University of Padova, USA
● Roman V. Yampolskiy, University of Louisville, USA
● Gereon Weiss, Fraunhofer ESK, Germany
● Roman Nagy, Argo AI, Germany
● Nathalie Baracaldo, IBM Research, USA
● Chokri Mraidha, CEA LIST, France</p>
        </sec>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <article-title>Finally we thank the AAAI-22 organization for providing an excellent framework for SafeAI 2022</article-title>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>