<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>The IJCAI-PRICAI-20 Workshop on Artificial Intelligence Safety (AISafety 2020)</article-title>
      </title-group>
      <contrib-group>
        <aff id="aff0">
          <label>0</label>
          <institution>Commissariat à l ́Energie Atomique</institution>
          ,
          <country country="FR">France</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Future of Life Institute</institution>
          ,
          <country country="US">USA</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Huáscar Espinoza</institution>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>Loockheed Martin</institution>
          ,
          <country country="US">USA</country>
        </aff>
        <aff id="aff4">
          <label>4</label>
          <institution>Univeristy of York</institution>
          ,
          <country country="UK">UK</country>
        </aff>
        <aff id="aff5">
          <label>5</label>
          <institution>Universitat Politècnica de València</institution>
          ,
          <country country="ES">Spain</country>
        </aff>
        <aff id="aff6">
          <label>6</label>
          <institution>University of Cambridge</institution>
          ,
          <country country="UK">UK</country>
        </aff>
        <aff id="aff7">
          <label>7</label>
          <institution>University of Hong Kong</institution>
          ,
          <country country="CN">China</country>
        </aff>
        <aff id="aff8">
          <label>8</label>
          <institution>University of Liverpool</institution>
          ,
          <country country="UK">UK</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>This preface introduces the Second Workshop on Artificial Intelligence Safety (AISafety 2020), held at the 29th International Joint Conference on Artificial Intelligence and the 17th Pacific Rim International Conference on Artificial Intelligence (IJCAI-PRICAI) in January 2021, Japan.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>In the last decade, there has been a growing concern on risks
of Artificial Intelligence (AI). Safety is becoming
increasingly relevant as humans are progressively side-lined from
the decision/control loop of intelligent and learning-enabled
machines. In particular, the technical foundations and
assumptions on which traditional safety engineering principles
are based, are inadequate for systems in which AI algorithms,
and in particular Machine Learning (ML) algorithms, are
interacting with people and/or the environment at
increasingly higher levels of autonomy. We must also consider the
connection between the safety challenges posed by
present-day AI systems, and more forward-looking research
focused on more capable future AI systems, up to and
including Artificial General Intelligence (AGI).</p>
      <p>The IJCAI-PRICAI-20 Workshop on Artificial Intelligence
Safety (AISafety 2020) seeks to explore new ideas on AI
safety with particular focus on addressing the following
questions:
• How can we engineer trustworthy AI system
architectures?
• Do we need to specify and use bounded morality in
engineering more ethically-aligned AI-based systems?
• What is the status of existing approaches in ensuring AI
and ML safety and what are the gaps?
• How to evaluate AI safety?
• What AI safety considerations and experiences are
relevant from industry?
• What safety engineering considerations are required to
develop safe human-machine interaction in automated
decision-making systems?
• How can we characterise or evaluate AI systems
according to their potential risks and vulnerabilities?
• How can we develop solid technical visions and paradigm
shift articles about AI Safety?
• How do metrics of capability and generality affect the
level of risk of a system and how trade-offs can be found
with performance?
• How do AI system feature for example ethics,
explainability, transparency, and accountability relate to, or
contribute to, its safety?
The main interest of AISafety 2020 is to look holistically at
AI and safety engineering, jointly with the ethical and legal
issues, to build trustable intelligent autonomous machines.
The second edition of AISafety will be held in January 2021,
in Japan, as part of the 29th International Joint Conference on
Artificial Intelligence and the 17th Pacific Rim International
Conference on Artificial Intelligence (IJCAI-PRICAI). The
AISafety workshop is organized as a “sister workshop” to
two other workshops: WAISE1 and SafeAI2.</p>
      <p>Copyright © 2020 for the individual papers by the papers' authors.
Copyright © 2020 for the volume as a collection by its editors. This
volume and its papers are published under the Creative Commons
License Attribution 4.0 International (CC BY 4.0)
1 https://www.waise.org
2 http://www.safeaiw.org</p>
      <p>As part of this workshop, we also host discussions related
to the AI Safety Landscape initiative3. This initiative aims at
defining an AI safety landscape providing a “view” of the
current needs, challenges and state of the art and the practice
of this field.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Programme</title>
      <p>The Programme Committee (PC) received 25 submissions, in
the following categories:
• Short position papers – 4 submission.
• Full scientific contributions – 20 submissions.
• Proposals of technical talks – 1 submission.</p>
      <p>Each of the papers was peer-reviewed by at least three PC
members, by following a single-blind reviewing process. The
committee decided to accept 11 papers and 1 talk, resulting in
an overall acceptance rate of 48%. We additionally accepted
6 submissions as poster presentations, (5 of which are
included in this proceedings, as poster papers.</p>
      <p>AISafety 2020 has been planned as a two-day workshop
with general AI Safety topics in the first day and AI Safety
Landscape talks and discussions during the second day. Since
the workshop has been delayed, together with
IJCAI-PRICAI-20, from July 2020 to January 2021, due to
the COVID-19 pandemic, we do not have yet the full list of
invited talks for the first day and no specific talk allocated to
the second day. The exact date and format (in-person or
virtual conference) is still under discussion by
IJCAI-PRICAI organizers at the date of publication of
AISafety-20 Proceedings.
2.1. First Workshop Day</p>
      <p>The AISafety 2020 programme will be organized in four
thematic sessions, one keynote and at least three invited
talks.</p>
      <p>The thematic sessions will be structured into short talks
and a common panel slot to discuss both individual paper
contributions and shared topic issues. Three specific roles are
part of this format: session chairs, presenters and session
discussants.
• Session Chairs introduce sessions and participants. The
Chair moderates session and plenary discussions, takes
care of the time, and gives the word to speakers in the
audience during discussions.
• Presenters give a paper talk in 10 minutes and then
participate in the debate slot.
• Session Discussants prepare the discussion of individual
papers and the plenary debate. The discussant gives a
critical review of the session papers.</p>
      <p>The mixture of topics has been carefully balanced, as
follows:
3 https://www.ai-safety.org</p>
      <sec id="sec-2-1">
        <title>Session 1: Adversarial Machine Learning</title>
        <p>• Understanding the One Pixel Attack: Propagation Maps
and Locality Analysis. Danilo Vasconcellos Vargas and
Jiawei Su.
• Error-Silenced Quantization: Bridging Robustness and
Compactness. Zhicong Tang, Yinpeng Dong and Hang
Su.
• Evolving Robust Neural Architectures to Defend from
Adversarial Attacks. Shashank Kotyan and Danilo
Vasconcellos Vargas.</p>
      </sec>
      <sec id="sec-2-2">
        <title>Session 2: AI Safety Landscape</title>
        <p>• Update Report: AI Safety Landscape Initiative, by</p>
        <p>Landscape Chairs [without paper].
• Safety of Artificial Intelligence: A Collaborative Model.</p>
        <p>John McDermid and Yan Jia.</p>
      </sec>
      <sec id="sec-2-3">
        <title>Session 3: Safe and Value-Aligned Learning in Decision</title>
      </sec>
      <sec id="sec-2-4">
        <title>Making</title>
        <p>• Choice Set Misspecification in Reward Inference. Rachel</p>
        <p>Freedman, Rohin Shah and Anca Dragan.
• Safety Augmentation in Decision Trees. Sumanta Dey,</p>
        <p>Pallab Dasgupta and Briti Gangopadhyay.
• Aligning with Heterogenous Preferences for Kidney
Exchange. Rachel Freedman.</p>
      </sec>
      <sec id="sec-2-5">
        <title>Session 4: DNN Testing and Runtime Monitoring</title>
        <p>• Is Uncertainty Quantification in Deep Learning Sufficient
for Out-of-Distribution Detection? Adrian Schwaiger,
Poulami Sinhamahapatra, Jens Gansloser and Karsten
Roscher.
• DeepSmartFuzzer: Reward Guided Test Generation For
Deep Learning. Samet Demir, Hasan Ferit Eniser and
Alper Sen.
• A Comparison of Uncertainty Estimation Approaches in
Deep Learning Components for Autonomous Vehicle
Applications. Fabio Arnez, Huascar Espinoza, Ansgar
Radermacher and François Terrier.
• Increasing the Trustworthiness of Deep Neural Networks
via Accuracy Monitoring. Zhihui Shao, Jianyi Yang and
Shaolei Ren..</p>
        <p>Additionally, AISafety has currently allocated one invited
talk, and plans to invite one Keynote speaker and at least two
additional invited talks.</p>
      </sec>
      <sec id="sec-2-6">
        <title>Invited Talk</title>
        <p>• Nathalie Baracaldo (IBM Research). Security and Privacy
Challenges in Federated Learning.</p>
        <p>Six posters will be presented in 2-minutes pitches. Five of
them will also be part of this volume as poster papers.
• Robustness as inherent property of datapoints.
Catalin-Andrei Ilie, Marius Popescu, and Alin Stefanescu.
• An Efficient Adversarial Attack on Graph Structured</p>
        <p>Data. Zhengyi Wang and Hang Su [without paper].
• Towards Safe and Reliable Robot Task Planning. Snehasis</p>
        <p>Banerje.
• Extracting Money from Causal Decision Theorists.
Caspar Oesterheld and Vincent Conitzer.
• Ethically Compliant Planning in Moral Autonomous
Systems. Justin Svegliato, Samer Nashed and Shlomo
Zilberstein.
• Bayesian Model for Trustworthiness Analysis of Deep
Learning Classifiers. Andrey Morozov, Emil Valiev,
Michael Beyer, Kai Ding, Lydia Gauerhof and Christoph
Schorn.
2.2. Second Workshop Day: Landscape
The second-day workshop (AI Safety Landscape) sessions
will be organized into by-invitation talks and panels with
structured discussions. The by-invitation talks will focus on
diverse topics contributing to understand the AI Safety
Landscape scientific and technical challenges, industrial and
academic opportunities, as well as gaps and pitfalls.</p>
        <p>One important ambition of this initiative is to align and
synchronize the proposed activities and outcomes with other
related initiatives. The AI Safety Landscape work will
follow-up in future meetings and workshops.
3</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Acknowledgements</title>
      <p>We thank all those who submitted papers to AISafety 2020
and congratulate the authors whose papers and posters were
selected for inclusion into the workshop program and
proceedings.</p>
      <p>We specially thank our distinguished PC members, for
reviewing the submissions and providing useful feedback to
the authors:
• Rick Salay, University of Toronto, Canada
• Ganesh Pai, NASA Ames Research Center, USA
• Hélène Waeselynck, CNRS LAAS, France
• Rob Alexander, University of York, UK
• Vahid Behzadan, Kansas State University, USA
• Simon Fürst, BMW, Germany
• Chokri Mraidha, CEA LIST, France
• Orlando Avila-García, Atos, Spain
• Rob Ashmore, Defence Science and Technology
Laboratory, UK
• I-Jeng Wang, Johns Hopkins University, USA
• Chris Allsopp, Frazer-Nash Consultancy, UK
• Francesca Rossi, IBM and University of Padova, Italy
• Ramana Kumar, Google DeepMind, UK
• Javier Ibañez-Guzman, Renault, France
• Jérémie Guiochet, LAAS-CNRS, France
• Raja Chatila, Sorbonne University, France
• Hang Su, Tsinghua University, China
• François Terrier, CEA LIST, France
• Mehrdad Saadatmand, RISE SICS, Sweden
• Alec Banks, Defence Science and Technology
Laboratory, UK
• Gopal Sarma, Broad Institute of MIT and Harvard, USA
• Philip Koopman, Carnegie Mellon University, USA
• Roman Nagy, Autonomous Intelligent Driving, Germany
• Nathalie Baracaldo, IBM Research, USA
• Toshihiro Nakae, DENSO Corporation, Japan
• Peter Flach, University of Bristol, UK
• Richard Cheng, California Institute of Technology, USA
• José M. Faria, Safe Perspective, UK
• Ramya Ramakrishnan, Massachusetts Institute of
Technology, USA
• Gereon Weiss, Fraunhofer ESK, Germany
• Douglas Lange, Space and Naval Warfare Systems Center</p>
      <p>Pacific, USA
• Philippa Ryan Conmy, Adelard, UK
• Stefan Kugele, Technical University of Munich, Germany
• Colin Paterson, University of York, UK
• Ashley Llorens, Johns Hopkins University, USA
• Huáscar Espinoza, Commissariat à l´Energie Atomique,</p>
      <p>France
• John McDermid, University of York, UK
• Xiaowei Huang, University of Liverpool, UK
• Mauricio Castillo-Effen, Lockheed Martin, USA
• Xin Cynthia Chen, University of Hong Kong, China
• José Hernández-Orallo, Universitat Politècnica de
València, Spain
• Seán Ó hÉigeartaigh, University of Cambridge, UK
• Richard Mallah, Future of Life Institute, USA</p>
      <p>Finally, yet importantly, we thank the
IJCAI-PRICAI-20 organization for providing an excellent
framework for AISafety 2020.</p>
    </sec>
  </body>
  <back>
    <ref-list />
  </back>
</article>