<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">The IJCAI-ECAI-22 Workshop on Artificial Intelligence Safety (AISafety2022)</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Gabriel</forename><surname>Pedroza</surname></persName>
							<email>gabriel.pedroza@cea.fr</email>
							<affiliation key="aff0">
								<orgName type="institution">CEA LIST</orgName>
								<address>
									<country key="FR">France</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Xin</forename><forename type="middle">Cynthia</forename><surname>Chen</surname></persName>
							<affiliation key="aff1">
								<orgName type="institution">University of Hong Kong</orgName>
								<address>
									<country key="CN">China</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">José</forename><surname>Hernández-Orallo</surname></persName>
							<affiliation key="aff2">
								<orgName type="institution">Universitat Politècnica de València</orgName>
								<address>
									<country key="ES">Spain</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Xiaowei</forename><surname>Huang</surname></persName>
							<email>xiaowei.huang@liverpool.ac.uk</email>
							<affiliation key="aff3">
								<orgName type="institution">University of Liverpool</orgName>
								<address>
									<settlement>Liverpool</settlement>
									<country key="GB">United Kingdom</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Huascar</forename><surname>Espinoza</surname></persName>
							<email>huascar.espinoza@ecsel.europa.eu</email>
							<affiliation key="aff4">
								<orgName type="institution">KDT JU</orgName>
								<address>
									<country key="BE">Belgium</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Richard</forename><surname>Mallah</surname></persName>
							<email>richard@futureoflife.org</email>
							<affiliation key="aff5">
								<orgName type="department">Future of Life Institute</orgName>
								<address>
									<country key="US">USA</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">John</forename><surname>Mcdermid</surname></persName>
							<email>john.mcdermid@york.ac.uk</email>
							<affiliation key="aff6">
								<orgName type="institution">University of York</orgName>
								<address>
									<country key="GB">United Kingdom</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Mauricio</forename><surname>Castillo-Effen</surname></persName>
							<email>mauricio.castillo-effen@lmco.com</email>
							<affiliation key="aff7">
								<orgName type="department">Lockheed Martin</orgName>
								<orgName type="institution">Advanced Technology Laboratories</orgName>
								<address>
									<settlement>Arlington</settlement>
									<region>VA</region>
									<country key="US">USA</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">The IJCAI-ECAI-22 Workshop on Artificial Intelligence Safety (AISafety2022)</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">8003941A644151622EA04262707F62D0</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T23:22+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<abstract/>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Introduction</head><p>Safety in Artificial Intelligence (AI) is increasingly becoming a substantial part of AI research, deeply intertwined with the ethical, legal and societal issues associated with AI systems. Even if AI safety is considered a design principle, there are varying levels of safety, diverse sets of ethical standards and values, and varying degrees of liability, for which we need to deal with trade-offs or alternative solutions. These choices can only be analyzed holistically if we integrate technological and ethical perspectives into the engineering problem, and consider both the theoretical and practical challenges for AI safety. This view must cover a wide range of AI paradigms, considering systems that are specific for a particular application, and also those that are more general, which may lead to unanticipated risks. We must bridge the short-term with the long-term perspectives, idealistic goals with pragmatic solutions, operational with policy issues, and industry with academia, in order to build, evaluate, deploy, operate and maintain AI-based systems that are truly safe.</p><p>The IJCAI-ECAI-22 Workshop on Artificial Intelligence Safety (AISafety 2022) seeks to explore new ideas in AI safety with a particular focus on addressing the following questions: </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Program</head><p>The Program Committee (PC) received 26 submissions. Each paper was peer-reviewed by at least two PC members, by following a single-blind reviewing process. The committee decided to accept 13 full papers and 6 short presentations, resulting in a full-paper acceptance rate of 50% and an overall acceptance rate of 73%.</p><p>The AISafety 2022 program was organized in six thematic sessions, one (invited) special session, two keynote and four (invited) talks. The special session was given flexibility to structure its program and format.</p><p>The thematic sessions followed a highly interactive format. They were structured into short pitches and a group debate panel slot to discuss both individual paper contributions and shared topic issues. Three specific roles were part of this format: session chairs, presenters and session discussants.</p><p>• Session Chairs introduced sessions and participants.</p><p>The Chair moderated sessions and plenary discussions, monitored time, and moderated questions and discussions from the audience. • Presenters gave a 10-minute paper talk and participated in the debate slot. The short presentations are given 5 minutes for each paper. • Session Discussants gave a critical review of the session papers, and participated in the plenary debate.</p><p>Presentations and papers were grouped by topic as follows:</p><p>Session 1: AI Ethics: Fairness, Bias, and Accountability </p></div><figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Machine learning for safety-critical AI •</head><label></label><figDesc>Revisiting the Evaluation of Deep Neural Networks for Pedestrian Detection, Patrick Feifel, Benedikt Franke, Arne Raulf, Friedhelm Schwenker, Frank Bonarens and Frank Köster. • Improvement of Rejection for AI Safety through Loss-Based Monitoring, Daniel Scholz, Florian Hauer, Klaus Knobloch and Christian Mayr.</figDesc><table><row><cell cols="5">• Let it RAIN for Social Good, Mattias Brännström,</cell></row><row><cell cols="4">Andreas Theodorou and Virginia Dignum.</cell><cell></cell></row><row><cell cols="5">• Accountability and Responsibility of Artificial</cell></row><row><cell cols="5">Intelligence Decision-making Models in Indian Policy</cell></row><row><cell cols="4">Landscape, Palak Malhotra and Amita Misra.</cell><cell></cell></row><row><cell cols="5">• Assessing Demographic Bias Transfer from Dataset to</cell></row><row><cell cols="5">Model: A Case Study in Facial Expression Recognition,</cell></row><row><cell cols="5">Iris Dominguez-Catena, Daniel Paternain and Mikel</cell></row><row><cell>Galar.</cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell cols="5">Session 2: Short Presentations -Safety Assessment of</cell></row><row><cell cols="2">AI-enabled systems</cell><cell></cell><cell></cell><cell></cell></row><row><cell cols="5">• A Hierarchical HAZOP-Like Safety Analysis for</cell></row><row><cell cols="3">Learning-Enabled Systems,</cell><cell cols="2">Yi Qi, Philippa Ryan</cell></row><row><cell cols="5">Conmy, Wei Huang, Xingyu Zhao and Xiaowei Huang.</cell></row><row><cell cols="5">• Increasingly Autonomous CPS: Taming Emerging</cell></row><row><cell cols="5">Behaviors from an Architectural Perspective, Jerome</cell></row><row><cell cols="3">Hugues and Daniela Cancila.</cell><cell></cell><cell></cell></row><row><cell cols="5">• CAISAR: A platform for Characterizing Artificial</cell></row><row><cell>Intelligence</cell><cell>Safety</cell><cell>and</cell><cell>Robustness,</cell><cell>Julien</cell></row><row><cell cols="5">Girard-Satabin, Michele Alberti, François Bobot,</cell></row><row><cell cols="4">Zakaria Chihani and Augustin Lemesle.</cell><cell></cell></row><row><cell>Session 3: Special</cell><cell></cell><cell></cell><cell></cell><cell></cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Session : TAILOR -Towards Trustworthy AI •</head><label></label><figDesc>Foundations of Trustworthy AI*, Francesca Pratesi. • Panel on Trustworthy AI*, Fosca Giannotti, Pilipp Slusallek, Giuseppe De Giacomo, Hector Geffner, Holger Hoos.</figDesc><table /><note>*Presentations without papers.</note></figure>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgements</head><p>We thank all researchers who submitted papers to AISafety 2022 and congratulate the authors whose papers were selected for inclusion into the workshop program and proceedings.</p><p>We especially thank our distinguished PC members for reviewing the submissions and providing useful feedback to the authors:</p></div>
			</div>

			<div type="annex">
<div xmlns="http://www.tei-c.org/ns/1.0"><p>We thank Gary Marcus, Thomas A. Henzinger, Elizabeth Adams, Luis Aranda, Simos Gerasimou, and Zakaria Chihani for their inspiring talks.</p><p>Finally we thank the IJCAI-ECAI-22 organization for providing an excellent framework for AISafety 2022.</p></div>			</div>
			<div type="references">

				<listBibl/>
			</div>
		</back>
	</text>
</TEI>
