<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>The AAAI-19 Workshop on Artificial Intelligence Safety (SafeAI 2019)</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Huáscar Espinoza</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Seán Ó hÉigeartaigh</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Xiaowei Huang</string-name>
          <email>xiaowei.huang@liverpool.ac.uk</email>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff4">4</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>José Hernández-Orallo</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Mauricio Castillo-Effen</string-name>
          <email>mauricio.castillo-effen@lmco.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>CEA LIST</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Gif-sur-Yvette</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>France huascar.espinoza@cea.fr</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Lockheed Martin, Advanced Technology Laboratories</institution>
          ,
          <addr-line>Arlington, VA</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>This preface introduces the AAAI-19 Workshop on Artificial Intelligence Safety (SafeAI 2019), held at the Thirty- Third AAAI Conference on Artificial Intelligence on January 27</institution>
          ,
          <addr-line>2019 in Honolulu, Hawaii</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Universitat Politècnica de València</institution>
          ,
          <addr-line>Valencia</addr-line>
          ,
          <country country="ES">Spain</country>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>University of Cambridge</institution>
          ,
          <addr-line>Cambridge</addr-line>
          ,
          <country country="UK">United Kingdom</country>
        </aff>
        <aff id="aff4">
          <label>4</label>
          <institution>University of Liverpool</institution>
          ,
          <addr-line>Liverpool</addr-line>
          ,
          <country country="UK">United Kingdom</country>
        </aff>
      </contrib-group>
      <abstract>
        <p />
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>Safety in Artificial Intelligence (AI) should not be an
option, but a design principle. However, there are varying
levels of safety, diverse sets of ethical standards and
values, and varying degrees of liability, for which we need to
deal with trade-offs or alternative solutions. These choices
can only be analyzed holistically if we integrate the
technological and ethical perspectives into the engineering
problem, and consider both the theoretical and practical
challenges for AI safety. This view must cover a wide
range of AI paradigms, considering systems that are
specific for a particular application, and also those that are more
general, which may lead to unanticipated risks. We must
bridge the short-term with the long-term perspectives,
idealistic with pragmatic solutions, operational with policy
issues, and industry with academia, to build, evaluate,
deploy, operate and maintain AI-based systems that are truly
safe.</p>
      <p>The AAAI-19 Workshop on Artificial Intelligence
Safety (SafeAI 2019) seeks to explore new ideas on AI safety
with particular focus on addressing the following
questions:
 What is the status of existing approaches in ensuring AI
and Machine Learning (ML) safety and what are the
gaps?
 How can we engineer trustable AI software
architectures?
 How can we make AI-based systems more ethically
aligned?
 What safety engineering considerations are required to
develop safe human-machine interaction?
 What AI safety considerations and experiences are
relevant from industry?
 How can we characterize or evaluate AI systems
according to their potential risks and vulnerabilities?
 How can we develop solid technical visions and new
paradigms about AI Safety?
 How do metrics of capability and generality, and
tradeoffs with performance affect safety?
The main interest of SafeAI 2019 is to look holistically at
AI and safety engineering, jointly with the ethical and legal
issues, to build trustable intelligent autonomous machines.
The first edition of SafeAI was held in January 27, 2019, in
Honolulu, Hawaii (USA) as part of the Thirty-Third AAAI
Conference on Artificial Intelligence (AAAI-19).</p>
    </sec>
    <sec id="sec-2">
      <title>Program</title>
      <p>The Program Committee (PC) received 33 submissions, in
the following categories:
 Short position papers – 13 submissions.
 Full scientific contributions – 18 submissions.
 Proposals of technical talks – 2 submissions.</p>
      <p>Each of the papers was peer-reviewed by at least two PC
members, by following a single-blind reviewing process.
The committee decided to accept 12 papers (4 position
papers and 8 scientific papers) and 1 talk, resulting in an
overall acceptance rate of 39%. We additionally invited
two talks, which were not submitted to the call, and
accepted 10 submissions as short papers for poster
presentation.</p>
      <p>The SafeAI 2019 program was organized in five
thematic sessions, two keynote and three (invited) talks.</p>
      <p>The thematic sessions followed a highly interactive
format. They were structured into short pitches and a common
panel slot to discuss both individual paper contributions
and shared topic issues. Three specific roles were part of
this format: session chairs, presenters and session
discussants.
 Session Chairs introduced sessions and participants. The
Chair moderated session and plenary discussions, took
care of the time, and gave the word to speakers in the
audience during discussions.
 Presenters gave a paper pitch in 10 minutes and then
participated in the debate slot.
 Session Discussants prepared the discussion of
individual papers and the plenary debate. The discussant gave a
critical review of the session papers.</p>
      <p>The mixture of topics has been carefully balanced, as
follows:</p>
      <sec id="sec-2-1">
        <title>Session 1: Safe Planning and Operation of Autonomous</title>
      </sec>
      <sec id="sec-2-2">
        <title>Systems</title>
        <p> Minimizing the Negative Side Effects of Planning with
Reduced Models, Sandhya Saisubramanian and Shlomo
Zilberstein.
 Robust Motion Planning and Safety Benchmarking in
Human Workspaces, Shih-Yun Lo, Shani Alkoby and
Peter Stone.
 Enter the Matrix: Safely Interruptible Autonomous
Systems via Virtualization, Mark Riedl and Brent Harrison.</p>
      </sec>
      <sec id="sec-2-3">
        <title>Session 2: New Paradigms in AI and AGI Safety</title>
        <p> Integrative Biological Simulation, Neuropsychology,
and AI Safety, Gopal Sarma, Adam Safron and Nick
Hay.</p>
      </sec>
      <sec id="sec-2-4">
        <title>Session 3: Safety in Automated Driving</title>
        <p> How Many Operational Design Domains, Objects, and</p>
        <p>Events?, Philip Koopman and Frank Fratrik.
 Monitoring Safety of Autonomous Vehicles with Crash
Prediction Networks, Saasha Nair, Sina Shafaei, Stefan
Kugele, Mohd Hafeez Osman and Alois Knoll.</p>
      </sec>
      <sec id="sec-2-5">
        <title>Session 4: Safety-Related AI Requirements and Characteristics</title>
        <p> Requirements Assurance in Machine Learning, Alec</p>
        <p>Banks and Rob Ashmore.
 Surveying Safety-relevant AI Characteristics, Jose
Hernandez-Orallo, Fernando Martínez-Plumed, Shahar
Avin and Seán Ó hÉigeartaigh.</p>
      </sec>
      <sec id="sec-2-6">
        <title>Session 5: Adversarial Machine Learning</title>
        <p> Detecting Backdoor Attacks on Deep Neural Networks
by Activation Clustering, Bryant Chen, Wilka Carvalho,
Nathalie Baracaldo, Heiko Ludwig, Benjamin Edwards,
Taesung Lee, Ian Molloy and Biplav Srivastava.
 DPATCH: An Adversarial Patch Attack on Object
Detectors, Xin Liu, Huanrui Yang, Ziwei Liu, Linghao
Song, Yiran Chen and Hai Li.
 Attacks on Machine Learning: Lurking Danger for
Accountability, Katja Auernhammer, Ramin Tavakoli
Kolagari and Markus Zoppelt
Additionally, SafeAI was proud to bring great inspirational
speakers:</p>
      </sec>
      <sec id="sec-2-7">
        <title>Keynotes</title>
      </sec>
      <sec id="sec-2-8">
        <title>Invited Talks</title>
        <p> Dr. Sandeep Neema (DARPA), Assured Autonomy.
 Prof. Francesca Rossi (IBM and University of Padova),</p>
        <p>Ethically Bounded AI.
 Dr. Peter Eckersley (Partnership on AI), Impossibility
and Uncertainty Theorems in AI Value Alignment (or
why your AGI should not have a utility function).
 Dr. Ian Goodfellow (Google Brain), Adversarial
Robustness for AI Safety.
 Prof. Alessio R. Lomuscio (Imperial College London),
Reachability Analysis for Neural Agent-Environment
Systems.
 Towards Robust End-to-End Alignment, Lê Nguyên
Hoang.</p>
        <p>Posters were presented in 2-minute pitches and most of
them are also part of this volume as short papers.
ii</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Acknowledgements</title>
      <p>We thank all those who submitted papers to SafeAI 2019
and congratulate the authors whose papers and posters
were selected for inclusion into the workshop program and
proceedings.</p>
      <p>We specially thank our distinguished PC members, for
reviewing the submissions and providing useful feedback
to the authors:
 Stuart Russell, UC Berkeley, USA
 Francesca Rossi, IBM and University of Padova, Italy
 Raja Chatila, ISIR – Sorbonne University, France
 Roman V. Yampolskiy, University of Louisville, USA
 Nozha Boujemaa, DATAIA Institute &amp; INRIA, France
 Mark Nitzberg, Center for Human-Compatible AI, USA
 François Terrier, CEA LIST, France
 Orlando Avila-García, Atos, Spain
 Kobi Leins, Australian National University, Australia
 Rob Alexander, University of York, UK
 Peter Eckersley, Electronic Frontier Foundation, USA
 Andreas Theodorou, University of Bath, UK
 Philip Koopman, Carnegie Mellon University, USA
 Chokri Mraidha, CEA LIST, France
 Heather Roff, Google DeepMind, UK
 Bernhard Kaiser, ANSYS, Germany
 Brent Harrison, University of Kentucky, USA
 John Favaro, INTECS, Italy
 Rob Ashmore, Defence Science and Technology
Laboratory, UK
 Martin Vechev, ETH Zurich, Switzerland
 Jonas Nilsson, Zenuity, Sweden
 Philippa Ryan, Adelard, UK
 Florent Kirchner, CEA LIST, France
 Andrew Banks, LDRA, UK
 Stefan Kugele, Technical University of Munich,
Germany
 Marta Kwiatkowska, University of Oxford, UK
 Virginia Dignum, TU Delft, Netherlands
 Yi Zeng, Institute of Automation, Chinese Academy of</p>
      <p>Sciences, China
 Timo Latvala, Space Systems Finland, Finland
 Mehrdad Saadatmand, RISE SICS, Sweden
 Rick Salay, University of Waterloo, Canada
 Jérémie Guiochet, LAAS-CNRS, France
 Mario Gleirscher, University of York, UK
 Chris Allsopp, Frazer-Nash Consultancy, UK
 Daniela Cancila, CEA LIST, France
 Huáscar Espinoza, CEA LIST, France
 Seán Ó hÉigeartaigh, University of Cambridge, UK
 Xiaowei Huang, University of Liverpool, UK
 José Hernández-Orallo, Universitat Politècnica de</p>
      <p>València, Spain
 Mauricio Castillo-Effen, Lockheed Martin, USA
As well as the additional reviewers:
 Zac Hatfield-Dodds, Australian National University,</p>
      <p>Australia
 Patrik Hoyer, Space Systems Finland, Finland
 Paolo Panaroni, Intecs, Italy
 Renaud Sirdey, CEA LIST, France</p>
      <p>We thank Dr. Sandeep Neema, Prof. Francesca Rossi,
Dr. Peter Eckersley, Dr. Ian Goodfellow, and Prof.
Alessio R. Lomuscio for their interesting talks on the current
challenges of safety and ethics for AI and autonomy.</p>
      <p>Finally, yet importantly, we thank the AAAI-19
organization for providing an excellent framework for SafeAI
2019.</p>
    </sec>
  </body>
  <back>
    <ref-list />
  </back>
</article>