<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>The IJCAI-24 Workshop on Artificial Intelligence Safety (AISafety2024)</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Gabriel Pedroza</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Xiaowei Huang</string-name>
          <email>xiaowei.huang@liverpool.ac.uk</email>
          <xref ref-type="aff" rid="aff4">4</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Xin Cynthia Chen</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Fabio Arnez</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Huascar Espinoza</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Hernández-Orallo</string-name>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Mauricio Castillo-Effen</string-name>
          <email>mauricio.castillo-effen@lmco.com</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Richard Mallah</string-name>
          <email>richard@futureoflife.org</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>John McDermid</string-name>
          <email>john.mcdermid@york.ac.uk</email>
          <xref ref-type="aff" rid="aff5">5</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Andreas</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Theodorou</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>ANSYS</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>France gabriel.pedroza@ansys.com</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>ETH Zurich</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Switzerland xin.chen@inf.ethz.ch</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>CEA LIST</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>France fabio.arnez@cea.fr</string-name>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Future of Life Institute</institution>
          ,
          <country country="US">USA</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Lockheed Martin, Advanced Technology Laboratories</institution>
          ,
          <addr-line>Arlington, VA</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Universitat Politècnica de Catalunya</institution>
          ,
          <country country="ES">Spain</country>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>Universitat Politècnica de València</institution>
          ,
          <country country="ES">Spain</country>
        </aff>
        <aff id="aff4">
          <label>4</label>
          <institution>University of Liverpool</institution>
          ,
          <addr-line>Liverpool</addr-line>
          ,
          <country country="UK">United Kingdom</country>
        </aff>
        <aff id="aff5">
          <label>5</label>
          <institution>University of York</institution>
          ,
          <country country="UK">United Kingdom</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>We summarize the IJCAI-24 Workshop on Artificial Intelligence Safety (AISafety2024)1, held at the 33nd International Joint Conference on Artificial Intelligence (IJCAI-24) on August 4, 2024 in Jeju, South Korea.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>5 Chips JU, Belgium</p>
      <p>Huascar.Espinoza@kdt-ju.europa.eu</p>
    </sec>
    <sec id="sec-2">
      <title>Introduction</title>
      <p>Safety in Artificial Intelligence (AI) is increasingly
becoming a substantial part of AI research, deeply
intertwined with the ethical, legal and societal issues
associated with AI systems. Even if AI safety is considered
1 Workshop series website: https://www.aisafetyw.org/
Copyright © 2024 for this paper by its authors. Use permitted under
Creative Commons License Attribution 4.0 International (CC
BY 4.0).
a design principle, there are varying levels of safety,
diverse sets of ethical standards and values, and varying
degrees of liability, for which we need to deal with
trade-offs or alternative solutions. These choices can only
be analyzed holistically if we integrate technological and
ethical perspectives into the engineering problem, and
consider both the theoretical and practical challenges for
AI safety. This view must cover a wide range of AI
paradigms, considering systems that are specific for a
particular application, and also those that are more general,
which may lead to unanticipated risks. We must bridge the
short-term with the long-term perspectives, idealistic goals
with pragmatic solutions, operational with policy issues,
and industry with academia, in order to build, evaluate,
deploy, operate, and maintain AI-based systems that are
truly safe.
The IJCAI-24 Workshop on Artificial Intelligence Safety
seeks to explore new ideas in AI safety with a particular
focus on addressing the following questions:
● How can we engineer trustable AI software</p>
      <p>architectures?
● Do we need to specify and use bounded morality in
system engineering to make AI-based systems more
ethically aligned?
● What is the status of existing approaches in ensuring AI
and ML safety and what are the gaps?
● What safety engineering considerations are required to
develop safe human-machine interaction in automated
decision-making systems?
● What AI safety considerations and experiences are
relevant from industry?
● How can we characterise or evaluate AI systems
according to their potential risks and vulnerabilities?
● How can we develop solid technical visions and
paradigm shift articles about AI Safety?
● How do metrics of capability and generality affect the
level of risk of a system and how trade-offs can be found
with performance?
● How do AI systems feature for example ethics,
explainability, transparency, and accountability relate to,
or contribute to, its safety?
● How to evaluate AI safety?
● How to safeguard GenAI/LLMs/ML?
These are the main topics of the series of AISafety
workshops which this year have been enriched by a
particular focus on Reinforcement Learning techniques,
their challenges, solutions and perspectives. Overall, the
series aims to achieve a holistic view of AI and safety
engineering, taking ethical and legal issues into account, in
order to build trustworthy intelligent autonomous
machines.</p>
    </sec>
    <sec id="sec-3">
      <title>Program</title>
      <p>The Program Committee (PC) received 11 submissions.
Each paper was peer-reviewed by at least two PC
members, by following a single-blind reviewing process.
The committee decided to accept 8 full papers, resulting in
an overall paper acceptance rate of 72%.</p>
      <p>The AISafety2024 program was organized in four
thematic sessions, two keynote, and one (invited) talk. The
thematic sessions followed a highly interactive format.
They were structured into short pitches and a group debate
panel slot to discuss both individual paper contributions
and shared topic issues. Three specific roles were part of
this format: session chairs, presenters and session
discussants.
● Session Chairs introduced sessions and participants.</p>
      <p>The Chair moderated sessions and plenary discussions,
monitored time, and moderated questions and
discussions from the audience.
● Presenters gave a 10-minute paper talk and participated
in the debate slot.
● Session Discussants gave a critical review of the
session papers, and participated in the plenary debate.
● Invited speakers gave a 25-minute talk on a relevant
topic to the workshop.
●</p>
      <p>Keynote speakers gave a 45-minute talk on a relevant
topic to the workshop.</p>
      <p>Presentations and papers were grouped by topic as follows:</p>
      <sec id="sec-3-1">
        <title>Session 1: AI Safety Assessment: Validation Methods and Techniques</title>
        <p>● ReLESS: A Framework for Assessing Safety in Deep
Learning Systems, Anita Raja, Nan Jia, Raffi
Khatchadourian
● Enhancing Autonomous Vehicle Safety through
N-version Machine Learning Systems, Qiang Wen,
Julio Mendonça, Fumio Machida, Marcus Völp</p>
      </sec>
      <sec id="sec-3-2">
        <title>Session 2: AI Robustness: Adversarial Learning and</title>
      </sec>
      <sec id="sec-3-3">
        <title>Security/Privacy</title>
        <p>● Hyper-parameter Tuning for Adversarially Robust</p>
        <p>Models, Pedro Mendes, Paolo Romano, David Garlan
● Low-Latency Privacy-Preserving Deep Learning</p>
        <p>Design via Secure MPC, Ke Lin, Yasir Glani, Ping Luo</p>
      </sec>
      <sec id="sec-3-4">
        <title>Session 3: Safety of Generative AI: OoD and Human vs</title>
      </sec>
      <sec id="sec-3-5">
        <title>Machine Generative Detection</title>
        <p>● Detecting Out-of-Distribution Text Using Topological
Features of Transformer-Based Language Models, Anj
Simmons, Andres Pollano, Anupam Chaudhuri
● The Impact of Prompts on Zero-Shot Detection of
AI-Generated Text, Kouichi Sakurai, Kaito Taguchi,
Yujie Gu</p>
      </sec>
      <sec id="sec-3-6">
        <title>Session 4: AI Robustness: Resilience to Noise and Soft</title>
      </sec>
      <sec id="sec-3-7">
        <title>Errors</title>
        <p>● Global Clipper: Enhancing Safety and Reliability of
Transformer-based Object Detection Models, Qutub
Syed, Michael Paulitsch, Karthik Pattabiraman,
Korbinian Hagn, Fabian Oboril, Cornelius Buerkle,
Kay-Ulrich Scholl, Gereon Hinz, Alois Knoll
● Neural Vicinal Risk Minimization: Noise-Robust
Distillation for Noisy Labels, Hyounguk Shon,
Seunghee Koh, Yunho Jeon, Junmo Kim
AISafety was pleased to have several
inspirational researchers as invited speakers:
additional</p>
      </sec>
      <sec id="sec-3-8">
        <title>Keynotes</title>
        <p>● Konstantin Dmitriev, The Evolvement of AI/ML
Aviation Regulations and Illustration of Some Practical
Aspects through an End-to-End Certification Case
Study
● Loïc Cantat, Journey and Findings of the Research</p>
        <p>Program Confiance.ai</p>
      </sec>
      <sec id="sec-3-9">
        <title>Invited Talks</title>
        <p>● Yonah Welker, Ability-Centered AI and Policy
(Transatlantic Safety Dialogue and Designated Groups)</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Acknowledgements</title>
      <p>We thank all researchers who submitted papers to AISafety
2024 and congratulate the authors whose papers were
selected for inclusion into the workshop program and
proceedings.</p>
      <p>We especially thank our distinguished PC members for
reviewing the submissions and providing useful feedback
to the authors:
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●</p>
      <p>John Burden, University of Cambridge, UK
Fredrik Heintz, Linköping University, Sweden
Simon Fürst, BMW Group, Germany
Mandar Pitale, NVIDIA, USA
Leon Kester, TNO, The Netherlands
Bernhard Kaiser, ANSYS, Germany
Xingyu Zhao, University of Warwick, UK</p>
      <p>Javier Garcia, Universidad Carlos III de Madrid, Spain</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <article-title>Finally, we thank the IJCAI-24 organization for providing an excellent framework for AISafety2024.</article-title>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>