<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>The AAAI-21 Workshop on Artificial Intelligence Safety (SafeAI 2021)</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Huáscar Espinoza​</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>José Hernández-Orallo​</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Xin Cynthia Chen​</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Seán S. ÓhÉigeartaigh​</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Xiaowei Huang​</string-name>
          <email>xiaowei.huang@liverpool.ac.uk</email>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Mauricio Castillo-Effen​</string-name>
          <email>mauricio.castillo-effen@lmco.com</email>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Richard Mallah​</string-name>
          <email>richard@futureoflife.org</email>
        </contrib>
        <contrib contrib-type="author">
          <string-name>John McDermid​</string-name>
          <email>john.mcdermid@york.ac.uk</email>
        </contrib>
        <contrib contrib-type="author">
          <string-name>​ CEA LIST</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Gif-sur-Yvette</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>France huascar.espinoza@cea.fr</string-name>
        </contrib>
      </contrib-group>
      <abstract>
        <p>We summarize the AAAI-21 Workshop on Artificial Intelligence Safety (SafeAI 2021)​1​, virtually held at the Thirty-Fifth AAAI Conference on Artificial Intelligence on February 8.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>Safety in Artificial Intelligence (AI) is increasingly
becoming a substantial part of AI research, deeply
intertwined with the ethical, legal and societal issues
associated with AI systems. Even if AI safety is considered
a design principle, there are varying levels of safety,
diverse sets of ethical standards and values, and varying
degrees of liability, for which we need to deal with</p>
      <sec id="sec-1-1">
        <title>1 Workshop series website: ​http://safeaiw.org/</title>
        <p>Copyright © 2021 for this paper by its authors. Use permitted under
Creative Commons License Attribution 4.0 International (CC
BY 4.0).
trade-offs or alternative solutions. These choices can only
be analyzed holistically if we integrate technological and
ethical perspectives into the engineering problem, and
consider both the theoretical and practical challenges for
AI safety. This view must cover a wide range of AI
paradigms, considering systems that are specific for a
particular application, and also those that are more general,
which may lead to unanticipated risks. We must bridge the
short-term with the long-term perspectives, idealistic goals
with pragmatic solutions, operational with policy issues,
and industry with academia, in order to build, evaluate,
deploy, operate and maintain AI-based systems that are
truly safe.</p>
        <p>The AAAI-21 Workshop on Artificial Intelligence
Safety (SafeAI 2021) seeks to explore new ideas in AI
safety with a particular focus on addressing the following
questions:
● What is the status of existing approaches for ensuring
AI and Machine Learning (ML) safety and what are the
gaps?
● How can we engineer trustworthy AI software
architectures?
● How can we make AI-based systems more ethically
aligned?
● What safety engineering considerations are required to
develop safe human-machine interaction?
● What AI safety considerations and experiences are
relevant from industry?
● How can we characterize or evaluate AI systems
according to their potential risks and vulnerabilities?
● How can we develop solid technical visions and new
paradigms about AI safety?
● How do metrics of capability and generality, and
trade-offs with performance, affect safety?
These are the main topics of the series of SafeAI
workshops. They aim to achieve a holistic view of AI and
safety engineering, taking ethical and legal issues into
account, in order to build trustworthy intelligent
autonomous machines. The first edition of SafeAI was held
in January 27, 2019, in Honolulu, Hawaii (USA) as part of
the Thirty-Third AAAI Conference on Artificial
Intelligence (AAAI-19), and the second edition was held in
February 7, 2020 in New York City (USA) also as part of
AAAI. This third edition was held online (because of the
COVID-19 situation) at the Thirty-Fifth AAAI Conference
on Artificial Intelligence on February 8, virtually.</p>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>Program</title>
      <p>The Program Committee (PC) received 44 submissions.
Each paper was peer-reviewed by at least two PC
members, by following a single-blind reviewing process.
The committee decided to accept 13 full papers, 1 talks and
11 posters, resulting in a full-paper acceptance rate of
29.5% and an overall acceptance rate of 56.8%.</p>
      <p>The SafeAI 2021 program was organized in four
thematic sessions, one keynote and two (invited) talks.</p>
      <p>The thematic sessions followed a highly interactive
format. They were structured into short pitches and a group
debate panel slot to discuss both individual paper
contributions and shared topic issues. Three specific roles
were part of this format: session chairs, presenters and
session discussants.
● Session Chairs introduced sessions and participants. The
Chair moderated sessions and plenary discussions,
monitored time, and moderated questions and
discussions from the audience.
● Presenters gave a 10 minute paper talk and participated
in the debate slot.
● Session Discussants gave a critical review of the session
papers, and participated in the plenary debate.</p>
      <sec id="sec-2-1">
        <title>Papers were grouped by topic as follows:</title>
        <sec id="sec-2-1-1">
          <title>Session 1: Dynamic Safety and Anomaly Assessment</title>
          <p>● Feature Space Singularity for Out-of-Distribution
Detection, Haiwen Huang, Zhihan Li, Lulu Wang,
Sishuo Chen, Xinyu Zhou and Bin Dong.
● An Evaluation of “Crash Prediction Networks” (CPN)
for Autonomous Driving Scenarios in CARLA
Simulator, Saasha Nair, Sina Shafaei, Daniel Auge and
Alois Knoll.
● From Black-box to White-box: Examining Confidence
Calibration under different Conditions, Franziska
Schwaiger, Maximilian Henne, Fabian Küppers, Felippe
Schmoeller Roza, Karsten Roscher and Anselm
Haselhoff.</p>
        </sec>
        <sec id="sec-2-1-2">
          <title>Session 2: Safety Considerations for the Assurance of</title>
        </sec>
        <sec id="sec-2-1-3">
          <title>AI-based Systems</title>
          <p>● The Utility of Neural Network Test Coverage Measures,</p>
          <p>Rob Ashmore and Alec Banks.
● Safety Properties of Inductive Logic Programming,</p>
          <p>Gavin Leech, Nandi Schoots and Joar Skalse.
● A Hybrid-AI Approach for Competence Assessment of
Automated Driving functions, Jan-Pieter Paardekooper,
Mauro Comi, Corrado Grappiolo, Ron Snijders, Willeke
van Vught and Rutger Beekelaar.</p>
        </sec>
        <sec id="sec-2-1-4">
          <title>Session 3: Adversarial</title>
        </sec>
        <sec id="sec-2-1-5">
          <title>Trustworthiness</title>
        </sec>
        <sec id="sec-2-1-6">
          <title>Machine</title>
        </sec>
        <sec id="sec-2-1-7">
          <title>Learning and</title>
          <p>● Adversarial Robustness for Face Recognition: How to
Introduce Ensemble Diversity among Feature
Extractors?, Takuma Amada, Kazuya Kakizaki, Seng
Pei Liew, Toshinori Araki, Joseph Keshet and Jun
Furukawa.
● Multi-Modal Generative Adversarial Networks Make
Realistic and Diverse but Untrustworthy Predictions
When Applied to Ill-posed Problems, John Hyatt and
Michael Lee.
● DeepFakesON-Phys: DeepFakes Detection based on
Heart Rate Estimation, Javier Hernandez-Ortega, Ruben
Tolosana, Julian Fierrez and Aythami Morales.</p>
        </sec>
        <sec id="sec-2-1-8">
          <title>Session 4: Safe Autonomous Agents</title>
          <p>● What Criminal and Civil Law Tells Us About Safe RL
Techniques to Generate Law-abiding Behaviour, Hal
Ashton.
● Performance of Bounded-Rational Agents With the
Ability to Self-Modify, Jakub Tětek, Marek Sklenka and
Tomáš Gavenčiak.
● Deep CPT-RL: Imparting Human-Like Risk Sensitivity
to Artificial Agents, Jared Markowitz, Marie Chau and
I-Jeng Wang.
● Challenges for Using Impact Regularizers to Avoid
Negative Side Effects, David Lindner, Kyle Matoba and
Alexander Meulemans.</p>
          <p>SafeAI was pleased to have several additional inspirational
researchers as invited speakers:</p>
        </sec>
        <sec id="sec-2-1-9">
          <title>Keynote</title>
        </sec>
        <sec id="sec-2-1-10">
          <title>Invited Talks</title>
          <p>● Not finalized at the date of publishing
● Juliette Mattioli (Thales, France) and Rodolphe Gelin
(Renault, France). Methods and Tools for Trusted AI: an
Urgent Challenge for Industry
● Sandhya Saisubramanian (University of Massachusetts
Amherst, USA), Challenges and Directions in Avoiding
Negative Side Effects
Posters were presented with 2-minute pitches. Most posters
have also been included as short papers within this volume.</p>
        </sec>
        <sec id="sec-2-1-11">
          <title>Posters</title>
          <p>● Towards an Ontological Framework for Environmental
Survey Hazard Analysis of Autonomous Systems,
Christopher Harper and Praminda Caleb-Solly.
● Overestimation learning with guarantees, Adrien</p>
          <p>Gauffriau, François Malgouyres and Mélanie Ducoffe.
● On the Use of Available Testing Methods for
Verification &amp; Validation of AI-based Software and
Systems, Franz Wotawa.
● Runtime Decision Making Under Uncertainty in
Autonomous Vehicles, Vibhu Gautam, Youcef
Gheraibia, Rob Alexander and Richard Hawkins.
● Negative Side Effects and AI Agent Indicators:
Experiments in SafeLife, John Burden, Jose
Hernandez-Orallo and Sean O'Heigeartaigh.
● Time for AI (Ethics) Maturity Model Is Now, Ville
Vakkuri, Marianna Jantunen, Erika Halme, Kai-Kristian
Kemell, Anh Nguyen-Duc, Tommi Mikkonen and
Pekka Abrahamsson.
● AI-Blueprint for Deep Neural Networks, Ernest</p>
          <p>Wozniak, Henrik Putzer and Carmen Carlan.
● Neural Criticality: Validation of Convolutional Neural</p>
          <p>Networks, Vaclav Divis and Marek Hruz.
● Adversarial Attacks for Tabular Data: Application to
Fraud Detection and Imbalanced Data, Francesco
Cartella, Orlando Anunciacao, Yuki Funabiki, Daisuke
Yamaguchi, Toru Akishita and Olivier Elshocht.
● Correct-by-Construction Multi-Label Classification
Networks, Eleonora Giunchiglia and Thomas
Lukasiewicz.
● Classification Confidence Scores with Point-wise
Guarantees, Nivasini Ananthakrishnan, Shai Ben-David
and Tosca Lechner.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Acknowledgements</title>
      <p>We thank all researchers who submitted papers to SafeAI
2021 and congratulate the authors whose papers and
posters were selected for inclusion into the workshop
program and proceedings.</p>
      <p>We especially thank our distinguished PC members for
reviewing the submissions and providing useful feedback
to the authors:
● Stuart Russell, UC Berkeley, USA
● Francesca Rossi, IBM and University of Padova, Italy
● Raja Chatila, Sorbonne University, France
● Roman V. Yampolskiy, University of Louisville, USA
● Gereon Weiss, Fraunhofer ESK, Germany
● Mark Nitzberg, Center for Human-Compatible AI, USA
● Roman Nagy, Autonomous Intelligent Driving GmbH,</p>
      <p>Germany
● François Terrier, CEA LIST, France
● Hélène Waeselynck, LAAS-CNRS, France
● Siddartha Khastgir, University of Warwick, UK
● Orlando Avila-García, Atos, Spain
● Nathalie Baracaldo, IBM Research, USA
● Peter Eckersley, Partnership on AI, USA
● Andreas Theodorou, Umeå University, UK
● Emmanuel Arbaretier, Apsys-Airbus, France
● Yang Liu, Webank, China
● Philip Koopman, Carnegie Mellon University, USA
● Chokri Mraidha, CEA LIST, France
● Heather Roff, Johns Hopkins University, USA
● Bernhard Kaiser, ANSYS, Germany
● Brent Harrison, University of Kentucky, USA
● José M. Faria, Safe Perspective, UK
● Toshihiro Nakae, DENSO Corporation, Japan
● John Favaro, Trust-IT, Italy
● Rob Ashmore, Defence Science and Technology</p>
      <p>Laboratory, UK
● Jonas Nilsson, NVIDIA, USA
● Michael Paulitsch, Intel, Germany
● Philippa Ryan Conmy, Adelard, UK
● Ramya Ramakrishnan, Massachusetts Institute of</p>
      <p>Technology, USA
● Stefan Kugele, Technical University of Munich,</p>
      <p>Germany
● Victoria Krakovna, Google DeepMind, UK
● Richard Cheng, California Institute of Technology, USA
● Javier Ibañez-Guzman, Renault, France
● Mehrdad Saadatmand, RISE SICS, Sweden
● Alessio R. Lomuscio, Imperial College London, UK
● Rick Salay, University of Waterloo, Canada
● Jérémie Guiochet, LAAS-CNRS, France
● Sandhya Saisubramanian, University of Massachusetts</p>
      <p>Amherst, USA
● Mario Gleirscher, University of York, UK
● Guy Katz, Hebrew University of Jerusalem, Israel
● Chris Allsopp, Frazer-Nash Consultancy, UK
● Daniela Cancila, CEA LIST, France
● Vahid Behzadan, University of New Haven, USA
● Simos Gerasimou, University of York, UK
● Brian Tse, Affiliate at University of Oxford, China
● Peter Flach, University of Bristol, UK
● Gopal Sarma, Broad Institute of MIT and Harvard, USA
● Huáscar Espinoza, CEA LIST, France
● Seán Ó hÉigeartaigh, University of Cambridge, UK
● Xiaowei Huang, University of Liverpool, UK
● José Hernández-Orallo, Universitat Politècnica de</p>
      <p>València, Spain
● Mauricio Castillo-Effen, Lockheed Martin, USA
● Xin Cynthia Chen, University of Hong Kong, China
● Richard Mallah, Future of Life Institute, USA
● John McDermid, University of York, United Kingdom</p>
      <sec id="sec-3-1">
        <title>As well as the additional reviewers:</title>
        <p>● Fabio Arnez,
● Dashan Gao
● Anbu Huang</p>
        <p>We thank Juliette Mattioli, Rodolphe Gelin and Sandhya
Saisubramanian for their inspiring talks.</p>
        <p>Finally we thank the AAAI-21 organization for
providing an excellent framework for SafeAI 2021.</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list />
  </back>
</article>