=Paper= {{Paper |id=Vol-3615/xpreface |storemode=property |title=None |pdfUrl=https://ceur-ws.org/Vol-3615/xpreface.pdf |volume=Vol-3615 }} ==None== https://ceur-ws.org/Vol-3615/xpreface.pdf
Preface to the Proceedings of BEWARE-23


   The second edition of the BEWARE workshop (https://sites.google.com/view/
beware2023), co-located with the AIxIA 2023 conference, was held in Rome on November
6, 2023. The workshop focused on the emerging ethical aspects of AI, particularly addressing
Bias, Risk, Explainability, and the role of Logic and Logic Programming. The event brought
together a diverse group of researchers and practitioners to discuss and explore solutions for
ethical decision-making in AI.
   This year, the workshop saw significant participation, with 9 accepted high-quality submis-
sions, 5 as long papers and 4 as short papers (out of a total of 11 submitted papers), rigorously
peer-reviewed by our Program Committee. We were honored to have Teresa Scantamburlo
(University of Venice) as our invited speaker, who delivered a thought-provoking talk on “Moral
Exercises for Responsible AI” , with the following abstract:

       Artificial Intelligence (AI) scholars and practitioners have developed a vast range
       of tools and methodologies to address the ethical challenges raised by critical AI
       applications. Nevertheless, the mere availability of ethical toolkits is not sufficient to
       ensure responsible AI practices. In this presentation, I will introduce the idea of moral
       exercises to stimulate and guide meaningful conversations around choices and actions
       interrogating the AI actors along the whole AI-life cycle. Built upon classical and
       modern moral theories, moral exercises are structured activities that can offer guidance
       in moral reflection and judgment in AI-related tasks. In line with other approaches,
       such as value sensitive design, they suggest the need to rediscover the dimensions of
       personal commitment and intersubjectivity in responsible design practices, highlighting
       the value of community and dialogue. To explore the practical implications of moral
       exercises, I will consider critical choices in the field of algorithmic fairness and bias.

The talk was based on “Moral Exercises for Human Oversight of Algorithmic Decision-Making”
by Teresa Scantamburlo, Silvia Crafa and Giovanni Grandi which was presented during the
main track conference – feel free to check it out at https://www.aixia2023.cnr.it if you would
like to dive into the details of our invited speaker’s work.
   We also would like to extend our heartfelt thanks to all participants and the Program Com-
mittee for their outstanding efforts in making the BEWARE 2023 a successful and enriching
event:

    • Damiano Azzolini – University of Ferrara
    • Paolo Baldi – University of Salento
    • Massimiliano Badino – University of Verona

2nd Workshop on Bias, Ethical Al, Explainability and the role of Logic and Logic Programming, BEWARE-23, co-located
with AlxIA 2023, Roma Tre University, Roma, Italy, November 6, 2023
         © 2023 Copyright for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).
    • Guido Boella – University of Turin
    • Stefania Costantini – University of L’Aquila
    • Fabio Aurelio D’Asaro – University of Verona
    • Luigi Di Caro – University of Turin
    • Abeer Dyoub – University of L’Aquila
    • Giovanni De Gasperis – University of L’Aquila
    • Roberta Ferrario – CNR Trento
    • Mattia Fumagalli – University of Bozen–Bolzano
    • Ekaterina Kubyshkina – University of Milan
    • Ludovica Marinucci – University of Salerno
    • Francesca A. Lisi – University of Bari
    • Davide Posillipo – Alkemy
    • Giuseppe Primiero – University of Milan
    • Daniele Porello – University of Genoa
    • Francesca Pratesi – ISTI-CNR, Pisa
    • Roberto Prevete – University of Naples Federico II
    • Daniele Chiffi – Polytechnic University of Milan
    • Teresa Scantamburlo – University of Venice Ca’Foscari
    • Viola Schiaffonati – Polytechnic University of Milan
    • Giovanni Sartor – University of Bologna
    • Matteo Spezialetti – University of L’Aquila
    • Guglielmo Tamburrini – University of Naples Federico II

The dedication and expertise of all involved have been instrumental in advancing our
understanding of ethical AI. We anticipate the continued evolution of this dialogue in the
next edition of BEWARE, where we hope to delve even deeper into these crucial topics. Our
gratitude goes to everyone who contributed to the success of this year’s event.


Verona, December 2023


The Organizers,

Guido Boella
Fabio Aurelio D’Asaro
Abeer Dyoub
Laura Gorrieri
Francesca Alessandra Lisi
Chiara Manganini
Giuseppe Primiero
Acknowledgments
We would like to acknowledge the contribution of the BRIO project (Bias, RIsk and Opacity
in AI, MUR PRIN Project 020SSKZ7R), the AI Aware project, the SIpEIA organization, and
the FAIR foundation for their invaluable support and collaboration in making the BEWARE-
2023 workshop a success. The event, held in collaboration with the FAIR foundation (https:
//future-ai-research.it), benefited immensely from the contributions and expertise provided by
these esteemed organizations. Their commitment to advancing the field of ethical AI has played
a pivotal role in enriching the discussions and outcomes of this year’s workshop.