=Paper= {{Paper |id=Vol-3069/longpaper02 |storemode=property |title=Teaching AI Ethics to Engineering Students: Reflections on Syllabus Design and Teaching Methods |pdfUrl=https://ceur-ws.org/Vol-3069/FP_02.pdf |volume=Vol-3069 |authors=Lauri Tuovinen,Anna Rohunen }} ==Teaching AI Ethics to Engineering Students: Reflections on Syllabus Design and Teaching Methods== https://ceur-ws.org/Vol-3069/FP_02.pdf
        Proceedings of the Conference on Technology Ethics 2021 - Tethics 2021




     Teaching AI Ethics to Engineering Students: Reflections
           on Syllabus Design and Teaching Methods


                                           Long paper


           Lauri Tuovinen1[0000-0002-7916-0255] and Anna Rohunen[0000-0002-4896-7056]

                                University of Oulu, Oulu, Finland
                                    1
                                     lauri.tuovinen@oulu.fi



        Abstract. The importance of ethics in artificial intelligence is increasing, and this
        must be reflected in the contents of computer engineering curricula, since the
        researchers and engineers who develop artificial intelligence technologies and
        applications play a key part in anticipating and mitigating their harmful effects.
        However, there are still many open questions concerning what should be taught
        and how. In this paper we suggest an approach to building a syllabus for a course
        in ethics of artificial intelligence, make some observations concerning effective
        teaching methods and discuss some particular challenges that we have
        encountered. These are based on the pilot implementation of a new course that
        aimed to give engineering students a comprehensive overview of the ethical and
        legislative aspects of artificial intelligence, covering both knowledge of issues
        that the students should be aware of and skills that they will need in order to
        competently deal with those issues in their work. The course was well received
        by the students, but also criticized for its high workload. Substantial difficulties
        were experienced in trying to inspire the students to engage in discussions and
        debates among themselves, which may limit the effectiveness of the course in
        building the students’ ethical argumentation skills unless a satisfactory solution
        is found. Several promising ideas for future development of our teaching
        practices can be found in the literature.

        Keywords: artificial intelligence, ethics education, course syllabi, teaching
        practices


 1      Introduction

 Applications of artificial intelligence (AI) involve many ethical challenges. The
 increasing autonomy of AI-based systems to make decisions that may have harmful
 consequences to living beings raises questions concerning the safety, transparency and
 accountability of such systems. Individuals’ rights to privacy and self-determination
 are threatened by the collection of vast amounts of personal data to be analyzed using
 AI algorithms for purposes such as surveillance and psychological manipulation. There
 are various efforts to regulate the use of AI either currently underway or already




                    Copyright © 2021 for this paper by its authors.                             19
Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0)
       Proceedings of the Conference on Technology Ethics 2021 - Tethics 2021




 implemented, but given the rate at which AI technology is advancing, it seems doubtful
 that policymaking alone will be enough to curb morally questionable uses of AI
 effectively. At the forefront of this progress are the AI researchers and engineers, so as
 a complementary approach, we should aim to foster a strong sense of professional ethics
 among this community. This, in turn, means that ethics education should be included
 in AI engineering curricula at higher education institutions.
    It is much less obvious, however, what exactly should be taught and how. For
 example, if we consider autonomous cars, there are some things that can be imparted
 to the students as facts, such as the safety record of these vehicles so far and the current
 legislation governing their development and deployment, but arguably the majority of
 interesting issues are subject to ambiguity and debate. These include relevant ethical
 principles and their interpretation and prioritization (e.g. how an autonomous vehicle
 should weigh the safety of passengers against that of other people), broader societal
 implications (e.g. effects of the proliferation of autonomous vehicles on the
 transportation sector), and implications of possible game-changing future
 developments (e.g. artificial general intelligence). A comprehensive syllabus thus
 needs to strike a balance between technical and philosophical topics, as well as between
 practical topics that the students can apply immediately and more theoretical ones that
 enable them to keep up with future developments.
    Finding such a balance is a considerable challenge, especially when the time devoted
 to ethics in the curriculum is limited. Furthermore, the non-technical nature of much of
 the subject matter must be reflected in teaching and evaluation methods, which may
 require both students and teachers to adopt unfamiliar ways of thinking, since courses
 designed to build engineering skills typically do not expose them to philosophical
 concepts or methods. In this paper we examine the problem of teaching AI ethics to
 university-level engineering students based on the pilot implementation of a new course
 in the spring term of 2021. The course was offered at the University of Oulu, Finland,
 to any interested students as a non-mandatory part of a computer engineering
 curriculum, with emphasis on practical knowledge and skills that the students can use
 to identify and address ethical issues likely to be relevant to them in the present or the
 near future, but also some coverage of more theoretical topics.
    In the remainder of the paper, we first present a review of related work in Section 2,
 focusing on surveys and reports of existing courses dealing with AI ethics. In Section
 3 we give an overview of the objectives, topics and implementation of our own AI
 ethics course. In Section 4 we propose a framework for categorizing study topics and
 an approach to syllabus design based on the framework. In Section 5 we offer some
 reflections on teaching methods and practices, based on what we learned from the pilot
 implementation. In Section 6 we look at feedback received from students who took the
 course, and in Section 7 we present a discussion and conclusions.




                    Copyright © 2021 for this paper by its authors.                             20
Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0)
       Proceedings of the Conference on Technology Ethics 2021 - Tethics 2021




 2      Related work

 Fiesler et al. (2020) have reviewed general courses on technology ethics in computer
 science education, with a subset of AI-specific courses. In their study, they have
 analyzed 115 syllabi with respect to course content and learning outcomes. Of these
 courses, 10 were AI-specific, while 55 included some content for the topic AI and
 algorithms and 27 for AI and robots. The authors also identified some course topics
 closely related to AI, such as privacy and surveillance with 61 courses. Their analysis
 results show a high variability across technology ethics courses with respect to their
 content, which they deem unsurprising considering the lack of standards, as well as the
 disciplinary breadth of the syllabi reviewed. However, they highlight the fact that the
 observed variability has potential to enable computing ethics educators to learn from
 each other and to finally begin to create norms around the learning outcomes. Despite
 the variability, some patterns in the syllabi reveal some critical topics, such as privacy,
 algorithms, inequality and justice.
     Building on this work and another previous study (Saltz et al., 2019), Garrett et al.
 (2020) have compiled a new dataset of established syllabi, specifically focusing on AI
 ethics courses. They have analyzed 51 courses with AI ethics content in all, including
 31 standalone AI ethics courses and 20 technical AI or ML courses with AI ethics topics
 integrated into them. With respect to standalone AI ethics courses, they have identified
 the following topics categories: bias, automation and robots, law and policy,
 consequences of algorithms, philosophy/morality, privacy, future of AI, history of AI.
 When it comes to technical AI and ML courses with AI ethics topics, the most common
 topics identified were bias, fairness and privacy.
     Garrett et al. (2020) further identified some common teaching practices. For
 example, the available reading lists for standalone AI ethics courses showed that the
 majority of these courses included news articles as reading assignments; therefore, it
 seems that incorporating current events into the course material is a common way to
 illustrate consequences of AI usage. In some technical courses, the non-technical
 content focused on societal considerations or “technology for social good”. However,
 including ethical or social implications in technical courses may be inhibited by the fact
 that there is too much material to cover. Based on these two reviews by Fiesler et al.
 and Garrett et al., it seems that AI ethics teaching in higher education has often been
 organized as standalone courses so far, but in parallel with these, AI ethics topics are
 also increasingly being integrated into technical AI courses.
     Burton et al. (2018) have developed a course where they have employed science
 fiction as a tool to teach AI ethics. Through this approach, they aim to move from an
 authority-based view to knowledge (typical of fields with a strong practical component
 and an established body of knowledge) to equipping students with skills to cope also
 with the unforeseen ethical issues in their future work in technology development and
 deployment. During the course, the students analyze science fiction stories and brief
 articles using ethical theories as both evaluative and descriptive tools to recognize
 problems and consider possible solutions from multiple perspectives. The authors
 suggest that their course assignments help develop capacity for attention and critical
 thought in a manner that serves the students in their professional lives, enabling them




                    Copyright © 2021 for this paper by its authors.                            21
Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0)
       Proceedings of the Conference on Technology Ethics 2021 - Tethics 2021




 to identify potential ethical risks related to a given technology or model, as well as to
 articulate their arguments about a given ethical approach and see past incomplete or
 specious defenses of potentially unethical projects.
    Henderson (2019) has designed and implemented a standalone module on data ethics
 and privacy, aiming to raise students’ awareness of current debates in computer science
 and to teach how to address these issues in practice. The module includes
 interdisciplinary content from ethics, law and computer science, and covers the
 following learning outcomes: be able to understand various conceptions of ethics and
 privacy; be able to critically analyze research literature at the intersection of computer
 science, philosophy and the law; be able to understand the effect of, and the source of,
 bias or discrimination in a data-intensive system; understand the need for, and
 optionally be able to carry out ethical, social, or privacy assessment of data-intensive
 projects. The module was delivered as seminar sessions, with the aim to allow deeper
 discussions compared to shorter lectures. The course assessment was based on an essay,
 peer instruction with essay-style questions on the weekly topics by the students
 themselves, and a data protection impact assessment task.
    Wilk (2019) proposes content and teaching strategies for a new standalone course
 “Computers, Ethics, Law, and Public Policy” that aims to increase computer science
 students’ ethical and legal awareness, as well as to promote critical thinking and skills
 needed in decision-making in their future work regarding ethical issues. He presents
 ethical, legal and public policy issues relevant to building and using intelligent systems.
 These include, for example, ethical and legal problems of algorithmic decision-making,
 autonomous systems, social media, fake news, journalism, privacy, and big data. The
 author also suggests the specific topics to be taught in the course and proposes teaching
 strategies supporting the course objectives; these include, for example, discussing
 ethical dilemmas and how to make ethical decisions, seminars with defender and
 opponent roles, balancing theory and practice through analyzing case studies based on
 ethical theories, and utilization of decision-making methodologies.
    Slavkovik (2020) has designed and implemented a standalone course derived from
 an existing course, “Research Topics in Artificial Intelligence”, where her goal was to
 give an overview of the core issues in AI ethics in such a way that it motivates the
 students to pursue further learning in this area. The course also aimed to familiarize the
 students with the research topic of machine ethics, as well as some of the research
 methods and practices in the AI field. The course topics included machine ethics,
 explainable AI, fair-accountable-transparent AI, and responsible AI. Learning
 outcomes of the course included, for example, identification of the basic ethical
 problems related to AI systems, understanding of the premises of the core moral
 theories, ability to appraise the ethical aspects of AI problems, and insight into the
 research process in machine ethics. Through assessment of the original course
 methodology, the author concluded that there was room for implementing more
 learning by teaching, learning by reflection, and learning by example. Scientific articles
 were used as learning and discussion material. The students were evaluated through an
 oral exam and a group project with intermediary assignments.
    Fink (2018) has investigated how to find a balance in covering the theory and
 practice as well as the philosophy and ethics in an introductory computer science AI




                    Copyright © 2021 for this paper by its authors.                            22
Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0)
       Proceedings of the Conference on Technology Ethics 2021 - Tethics 2021




 course. She presents three examples of assignments that can be utilized to teach
 concepts of technical, philosophical and ethical issues related to intelligent agents in
 this type of courses. In the ethical assignment, students explore Isaac Asimov’s Three
 Laws of Robotics and the ethical implications of AI through the movie “I, Robot”. The
 assignment focuses on this movie as its theme revolves around Asimov’s laws and what
 can happen when they are applied strictly with logic and not considering standard
 human sensitivities. Finally, the students need to address the question of the role of pure
 logic vs. emotion in ethical behavior, which the author states is a key theme in the
 success or failure of an intelligent agent to truly make decisions that are acceptable from
 the human perspective.
    Williams et al. (2020) have designed and implemented an experimental ethics-based
 curricular module for an undergraduate course, “Robot Ethics”. This module aims to
 teach usage of human-subjects research methods to investigate potential ethical
 concerns arising in human-robot interaction by engaging students in real experimental
 ethics research. The students participate in robot ethics research as experimenters,
 through which they simultaneously learn methodological approaches to experimental
 robot ethics and use these methods to engage with key theoretical concepts. There were
 three interdisciplinary learning objectives in this course: normative influence of
 technology with the aim of understanding how technologies may affect human
 behaviors due to their perception as moral and social agents; experimental ethics with
 the aim of understanding how human-subject experimentation can be used to explore
 the ethical implications of technology; and ethical research conduct with the aim of
 understanding ethical concerns that may arise in the design and execution of
 experimental ethics experiments. The students’ learning was assessed through an
 inclass quiz on the experiment’s details related to all three learning objectives.
    Furey and Martin (2018) have developed a module on ethical thinking about
 autonomous vehicles in an AI course. They suggest incorporating this type of
 introductory lesson about ethics into a one-semester AI course. Through a modular
 approach, students have an opportunity to connect specific AI topics to the related
 ethical implications. In this module, students become familiar with the use of thought
 experiment through the trolley problem, as well as learn to understand the complex
 nature of ethical dilemmas. Regarding the ethics module, the course assessment is
 carried out as follows: the final project paper of the course requires discussion of the
 ethical implications of the project idea, and in the final examination there is a question
 assessing the students’ understanding of the trolley problem.
    Shapiro et al. (2020) introduce a data ethics teaching method, Re-Shape, for data
 science education, with the aim to teach the ethical implications of data collection and
 use in a computing course. Through the tools and activities of the method, students
 collect, process and visualize their movement data. Based on these data, critical
 reflection and coordinated classroom activities are carried out about data, data privacy,
 and human-centered systems for data science. Building on the idea of cultivating care,
 students are engaged with the concept of responsibility to other and confronted with the
 idea that they are the “other” within systems that collect and use personal data.
    Based on the AI ethics teaching examples described in this section, we have made
 the following remarks:




                    Copyright © 2021 for this paper by its authors.                            23
Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0)
          Proceedings of the Conference on Technology Ethics 2021 - Tethics 2021




      •    The learning outcomes and objectives of the described courses and modules
           are relatively similar. They aim to increase students’ awareness of the ethical
           implications of AI usage, and to train the students to identify and address
           ethical issues related to the development and deployment of AI in practice.
      • To achieve the learning outcomes, students are typically taught the key
           concepts and theories of ethics in parallel with current ethical issues related to
           AI usage. Furthermore, critical thinking as well as skills to discuss and
           appraise ethical issues are often promoted through teaching methods that
           support learning of this type of skills (such as article or case study analyses,
           or class discussions). In some courses, the students are also equipped with
           skills to use research methods for exploring the ethical implications of AI or
           trained to use practical tools to address the ethical issues.
      • In higher education AI ethics teaching, a diverse set of methods and strategies
           have been reported. Standalone courses typically utilize assignments that
           comprise analysis of articles, stories or movies, discussions or seminars on the
           identified ethical issues, as well as consideration of how to solve the studied
           issues using suitable tools. AI ethics modules that can be integrated into other
           courses, for their part, aim to train students e.g. to use specific methods for
           investigating ethical issues, as well as to consider the ethical aspects of a
           specific application, project or activity.
      • It seems that assessment methods other than traditional written exams can be
           employed when teaching AI ethics; this may be an approach that matches well
           the reported learning outcomes.
    Like the majority of the AI ethics teaching examples presented above, our new
 course was carried out as a standalone course. Many of the existing courses and modules
 comprise topics and learning outcomes that are relatively similar to our ones,
 specifically with respect to the ethical issues of AI and the aim to provide the students
 with ethics skills needed in the development and deployment of AI systems in practice.
 Teaching methods and strategies seem to vary a great deal among the reviewed courses
 and modules, and an extensive set of these has been reported in the literature. Similarly
 to what Fiesler et al. (2020) suggested, we see this type of variability as an excellent
 opportunity to learn from other AI ethics educators.
    Borenstein and Howard (2021) have recently presented some recommendations on
 how to design AI ethics teaching with the aim to foster a professional mindset for AI
 developers and to seriously engage the students with ethical challenges. We find these
 recommendations highly relevant to the selection of suitable teaching methods and
 strategies when teaching AI ethics to engineering students. Specifically, the authors
 suggest three elements to familiarize students with ethical challenges: 1) teaching the
 ethical design of AI algorithms, 2) incorporating fundamental concepts of data science
 and the ethics of data acquisition through usage of real-world datasets requiring privacy,
 fairness and legal issues to be addressed, and 3) offering “ethics across the curriculum”
 through systematic inclusion of AI ethics into the curriculum. Based on these
 recommendations, for instance, topical examples of misbehaving AI algorithms could
 be elaborated in standalone AI ethics courses, while the application of ethical principles
 to system design could be considered in technical AI courses.




                    Copyright © 2021 for this paper by its authors.                             24
Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0)
       Proceedings of the Conference on Technology Ethics 2021 - Tethics 2021




 3      Course description

 The planning of the course, worth 5 ECTS credit points, began in 2020. The full title of
 the course was “AI Ethics, Privacy and Legislation”. Because of the COVID-19
 pandemic, the course was designed from the start to be taught remotely using the
 Moodle online learning platform, with lectures implemented as Zoom meetings. The
 learning outcomes for the course were defined as follows:
      • Students will be aware of the ethical and legislative aspects and conditions that
           need to be considered in the design and deployment of AI applications;
      • They will understand the ethical special characteristics of AI applications,
           compared to information technology applications in general;
      • They will be able to examine existing and hypothetical AI applications from
           an ethical viewpoint and identify potential issues;
      • They will be able to weigh the benefits of AI applications against their
           drawbacks also in a wider, societal context;
      • They will be able to apply ethical principles in AI application design.
    The course was lectured over a period of 8 weeks, with each week dedicated to one
 of 8 course themes. The themes are explained in Table 1. For each week of the course,
 two Zoom sessions were planned: a main lecture of approximately 1.5 hours and a
 supplementary session of approximately 45 minutes, the latter featuring a short
 presentation or demonstration on a specialized topic followed by discussion.
 Exceptions to this plan were week 5, when the supplementary session had to be
 cancelled due to scheduling conflicts, and week 8, when the supplementary session took
 the form of a tutorial for a design methodology.
    The course participants were evaluated based on a series of 8 smaller assignments,
 one for each course theme, and a larger final assignment spanning all themes. Passing
 all 9 assignments was required in order to pass the course. The lecturers reviewed each
 submitted assignment and either accepted it as such or sent it back to the student for
 revisions; in either case the student would be given some verbal feedback on their work.
 The format of the smaller assignments varied from week to week, but as a typical
 example, the student would be instructed to choose an AI application and write a short
 essay about it, addressing questions related to the theme of the week.
    For the final assignment, the students could pick one of two options. Students who
 attended a certain number of Zoom sessions could choose to write a lecture diary, in
 which they would discuss the topics covered by the lectures and supplementary sessions
 and reflect on what they had learned. As an alternative, students could choose to
 complete a final exercise, in which they would come up with an AI application concept
 and write a report discussing the ethical aspect of the application, guided by the 8 course
 themes. This second option was provided so that students who were unable to attend
 the lectures or preferred to work independently for other reasons could complete the
 course; to study the course themes, they had access to lecture slides as well as items of
 further reading curated by the lecturers. Each student had the opportunity to send in an




                    Copyright © 2021 for this paper by its authors.                            25
Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0)
          Proceedings of the Conference on Technology Ethics 2021 - Tethics 2021




 incomplete draft of their final assignment and receive feedback on it to help them in
 preparing the final submission.

                      Table 1. The 8 principal themes of the AI ethics course.
     Week       Theme                                Topics covered
                                                     Essential concepts in ethics and AI,
      1         Introduction to AI ethics            perspectives on AI ethics

                                                     Debates and controversies involving e.g.
      2         Controversial AI applications        autonomous weapons and surveillance

                                                     Role of data in AI, data ethics, privacy,
      3         AI and data                          data ownership, data philanthropy

                                                     Accountability and transparency in
      4         AI as decision-maker                 automated decision-making

                                                     Societal changes driven/facilitated by AI,
      5         AI and society                       AI divides, AI literacy, good AI society

                                                     Overview of AI regulation, active and
      6         AI and legislation                   proposed legislation relevant to AI

                                                     Overview of AI ethics guides, review of
      7         AI ethics guides                     some major institutional guides

                                                     Implementing AI ethics in practice by
      8         Ethical AI design                    following a design methodology



 4        Syllabus design

 Among the challenges of designing a balanced syllabus for an AI ethics course, there
 are two in particular that we wish to draw attention to. The first one of these is striking
 a balance between topics related to ethical reasoning and argumentation and those
 related to understanding AI technology and systems. Covering both areas is essential:
 the latter to enhance the students’ awareness of AI applications and their ethical
 implications, the former to enhance their ability to analyze these implications and to
 make justified decisions regarding ethical issues when encountering them in their work.
 We refer to this as the philosophy-technology axis.
    The other challenge is the rapid rate of progress in AI research and development,
 which raises the question of how we can ensure that at least some of the course content
 remains relevant to the students also in the future. For dealing with issues that are
 immediately relevant to today’s AI practitioners, the course can offer practical tools
 such as ethical design guidelines and methodologies, but the further into the future we
 look, the less we know about what the capabilities of AI will be, and therefore the less




                    Copyright © 2021 for this paper by its authors.                               26
Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0)
       Proceedings of the Conference on Technology Ethics 2021 - Tethics 2021




 likely it is that any concrete guidelines given today will be useful in dealing with the
 ethical implications of those capabilities. To help the students gain some degree of
 preparedness, the course should cover technology topics representing visions of the
 future of AI, as well as philosophy topics that are more theoretical in nature but also
 more stable over time. This we refer to as the practice-theory axis.
    If we consider these two axes as perpendicular dimensions, we can visualize them
 as shown in Fig. 1. We can thus divide possible study topics in AI ethics into four broad
 categories, corresponding to the four quadrants of the figure. Counterclockwise from
 the top left, these are as follows:
      • Timeless Foundations: established concepts, theories and traditions in ethics;
      • Practical Guidance: applied ethics principles and guidelines relevant to AI;
      • Here and Now: AI applications of the present and the near future and the
            ethical issues associated with them;
      • Beyond the Horizon: future potential of AI and the ethical implications of
            hypothetical scenarios.




 Fig. 1. Categorization of study topics in AI ethics based on two perpendicular axes.

    Fig. 2 shows a selection of possible topics to be discussed on an AI ethics course,
 placed approximately where they are located on these two axes. The grey ellipse
 represents an approximation of the area from where we took the bulk of the subject
 matter for our course. Notably, the course syllabus was centered on the major ethical
 issues identified in present-day AI applications – safety, bias, accountability,
 explainability, privacy – and on the topics in the immediate vicinity of these, which
 provide essential context for the current issues and/or means for dealing with them. The
 least attention was devoted to the topics at the top of the figure, i.e. the most abstract
 and/or futuristic ones such as metaethics and superintelligence.
    We argue that this represents a reasonable core syllabus for a practically oriented AI
 ethics course, and furthermore that this system for classifying study topics can be used
 as a practical framework for designing a syllabus for a course with a wider scope.
 Expanding the scope with respect to a given aspect of AI ethics can be thought of as
 stretching the grey scoping bubble; the visualization of affinities between topics is
 useful in ensuring that the course remains a coherent whole, since it suggests clusters




                    Copyright © 2021 for this paper by its authors.                           27
Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0)
       Proceedings of the Conference on Technology Ethics 2021 - Tethics 2021




 of closely related topics that it would make sense to include together (and possibly also
 to discuss together, so the framework provides some hints for planning the course
 structure as well). The overall shape of the bubble serves as an indicator of the
 balancedness of the syllabus, so if the bubble becomes strongly elongated in some
 direction, this suggests that reviewing the planned syllabus and possibly including some
 additional topics for the sake of balance should be considered.




          Fig. 2. Placement of topics on the philosophy-technology and practice-theory axes.



 5      Teaching methods

 A central challenge that we faced in the planning and execution of the course was
 presenting the subject matter to engineering students in such a way that it engages them
 to think about AI more philosophically than their other courses require them to. Toward
 this end, each week of teaching would follow the same general pattern, where the
 lecture would introduce the students to the theme of the week from a mostly theoretical
 perspective. The role of the supplementary session would then be to have the students
 discuss their thoughts with their peers, and the role of the assignment would be to have
 them apply what they had learned.
    Real-world examples of AI applications were used in the Zoom sessions to illustrate
 theoretical concepts, and in many of the assignments to provide the students with a
 context where they can demonstrate their understanding of the subject matter. Perhaps
 predictably, when instructed to choose an application to analyze, the students tended to
 choose ones that are close to their own everyday lives; for example, services offered by
 Facebook or Google were popular choices. The evaluation of the assignments was based
 mainly on the students’ ability to build and defend arguments using the concepts and
 theories discussed in the Zoom sessions, although some of them also included a fact-
 finding element. Furthermore, several assignments invited the students to show original
 thought by proposing ideas for addressing ethical issues associated with a given
 application or technology.




                    Copyright © 2021 for this paper by its authors.                            28
Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0)
       Proceedings of the Conference on Technology Ethics 2021 - Tethics 2021




     Another category of AI applications that engaged the students’ interest were those
 involving palpable risk to core values such as health, safety or liberty. This is suggested
 by the popularity of such applications (e.g. medical AI systems, autonomous
 vehicles/weapons, predictive policing applications) as assignment subjects, and also by
 active audience participation during the week 2 lecture on controversial AI applications.
 The first recorded case of a pedestrian killed by an autonomous car was used already in
 the first lecture as an introductory example, since it provides an effective way to
 illustrate a number of key points:
       • A malfunction in an AI system may have severe harmful consequences,
            including loss of life.
       • When this happens, it may be hard to ascertain why it happened and who is
            morally or legally responsible.
       • Even a hypothetical perfect AI system may encounter a situation where it has
            to make a value-based choice between multiple bad options.
       • Short-term harm may be inevitable in the pursuit of long-term good, and the
            justification of such harm is also debatable.
     Overall, the course achieved mixed success at best in inspiring discourse and debate
 among the students. This represents a missed opportunity for the students to learn
 valuable argumentation skills through peer-to-peer interaction. We can only speculate
 on the reasons here; presumably, the teaching methods employed were less than optimal
 for this purpose, but on the other hand, it may also be that many of the students (who
 were computer engineering majors) were not predisposed to working like this.
 Furthermore, it seems plausible that a physical classroom where the participants can
 better connect with one another would have been a more suitable environment for this.
 This was particularly evident in the final supplementary session, where the participants
 were given a tutorial for a design methodology based on a set of cards, designed to be
 printed out and manipulated as physical objects. Since this was not an option, the
 tutorial was implemented using an online whiteboard application, with the result that
 the students participated mainly by adding their own ideas to the whiteboard without
 interacting with one another on Zoom while doing it.
     Online lecturing did have the unanticipated advantage that the students could use
 Zoom chat to ask questions and exchange thoughts even while the lecturer is speaking.
 We also set up a message board on Moodle and encouraged the students to use it for
 discussions, aiming to engage also those students who feel less comfortable expressing
 their views live on Zoom, but this proved unsuccessful. Another benefit of remote
 teaching was that this made it convenient to include lectures and presentations by
 visiting experts in the course program. Visitors were featured in two of the main lectures
 and three of the supplementary sessions; since all of these were planned from the start
 to be implemented as Zoom meetings, the visitors could deliver their presentations from
 where they are normally based and no special arrangements were required, making
 physical distance a non-factor in deciding which external experts to invite as visitors.
 Besides contributing their expertise to the course, the visitors enabled the principal
 lecturers to focus their efforts on course themes closer to their own respective areas of
 expertise, resulting in a higher overall standard of teaching quality than would otherwise
 have been achievable.




                    Copyright © 2021 for this paper by its authors.                            29
Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0)
        Proceedings of the Conference on Technology Ethics 2021 - Tethics 2021




 6       Student feedback

 Feedback on the course was solicited from the participating students using two
 anonymous surveys. A standard feedback form, generated automatically for every
 course at the university, was filled in and submitted by 7 course participants. The form
 included 11 statements to be rated on a Likert scale from 1 (totally disagree) to 5 (totally
 agree). The results are summarized in Table 2. The students were also requested to
 evaluate their workload during the course, bearing in mind that 1 ECTS credit is
 equivalent to 27 hours of student work. The answer options and corresponding integer
 values were no workload (2), light workload (1), suitable workload (0), high workload
 (1) and very high workload (2); the answers were split between suitable workload and
 very high workload for an average of 1.14.

           Table 2. 11 statements about the course, rated by students on a Likert scale.
   Statement                                                                               Avg
   I achieved the course learning outcomes and course objectives.                          3.86

   Course content helped me to achieve learning outcomes.                                  3.57
   The course content supported my progression towards expertise in my field.              4.00
   Teaching methods supported learning and helped me to achieve learning outcomes.         3.71
   Course material supported my learning.                                                  3.57
   Instructions to the course tasks were clear.                                            3.86
   There was enough support and guidance in the course.                                    3.57
   Assessment methods and criteria supported my learning.                                  4.14
   There was enough time to complete the tasks in the course.                              4.14
   I have tried to advance my own learning outside the lectures (e.g. reading the lecture
   material, reading some literature connected to the topic or searching more             4.43
   information).

   I have tried to advance my own learning during the lectures by discussing the topic
   with other students, by asking questions from the lecturer, by starting discussion in   3.71
   the whole group or questioning the teaching.



    In addition to these, the standard form had open-ended questions on good practices,
 things to develop and other feedback. Notably, several of the answers to these questions
 included complaints about the workload of the course assignments. There were also
 individual criticisms concerning the course timetable and the supplementary session
 concept. Positive feedback was received on lectures, study materials, weekly
 assignments and guest lecturers.




                    Copyright © 2021 for this paper by its authors.                               30
Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0)
        Proceedings of the Conference on Technology Ethics 2021 - Tethics 2021




    A supplementary feedback form was designed to capture more detailed information
 on the students’ expectations and learning outcomes. This form was available in the
 course workspace on Moodle and was filled in and submitted by 7 course participants.
 Concerning expectations, the form included the following questions:
      • What were your reasons for signing up for the course?
      • What were your expectations for the course with respect to your personal and
           professional development?
      • Would you agree that the course fulfilled your aforementioned expectations?
    The first two of these were open-ended questions. Among the answers were some
 generic motivations such as finding the subject interesting and needing the credits to
 complete one’s studies, but several answers also indicated that the student was
 expecting knowledge of AI ethics to be an asset in jobs involving AI. The third question
 was to be answered on a Likert scale from 1 (strongly disagree) to 5 (strongly agree);
 the answers ranged from 3 to 5 for an average of 4.29. Optionally, the students could
 specify their answers to the third question; here one student indicated that they would
 have liked to learn more about the fundamentals of ethics, while another remarked that
 the exercises required a lot of effort and more writing than the student was used to, but
 that they were also good practice. Better visualization of lecture materials was also
 requested, and numeric grading instead of pass/fail was suggested as an incentive to put
 more effort into the assignments.
    Concerning learning outcomes, the form included four Likert-scale questions
 corresponding to the four basic categories of study topics identified in Section 4. The
 theory here was that a balanced syllabus would contribute to the students’ awareness
 of ethical issues in AI – both those that are relevant now and those that are still in the
 future – as well as their confidence in their ability to deal with these issues in their work.
 The results are summarized in Table 3.

        Table 3. 4 statements about course outcomes, rated by students on a Likert scale.
   Statement                                                                                Avg
   The course increased my awareness of ethical issues that are relevant to AI
   applications today.                                                                      4.71

   The course increased my confidence in my ability to competently address these
   current issues in my work, if and when I encounter them.                                 4.14

   The course increased my awareness of ethical issues that may arise in the future,
   given the development potential of AI technology.                                        4.43

   The course increased my confidence in my ability to competently address these
   future issues in my work, if and when they arise.                                        4.29



 7       Discussion and conclusions

 In this paper we discussed the teaching of AI ethics to engineering students, based on
 experience of lecturing a pilot course in the spring semester of 2021. We faced several




                    Copyright © 2021 for this paper by its authors.                                31
Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0)
       Proceedings of the Conference on Technology Ethics 2021 - Tethics 2021




 challenges in designing a comprehensive and balanced syllabus for the course, as well
 as in the selection of effective teaching methods. The positive feedback received from
 participants suggests that the course was at least a moderate success, but since the
 number of students who took the course was small and only a fraction of them answered
 the feedback surveys, the results mut be interpreted with caution.
    The aspect of the course that received the most criticism from the students was the
 workload. The most obvious candidate for an explanation is that the effort required to
 complete the assignments was underestimated when planning the course, but there are
 other possible contributing factors. For example, if the format of the assignments was
 such that it took the students out of their comfort zone – as engineering students they
 would be more familiar with e.g. programming assignments than essay writing – then
 this may have affected the students’ perception of the course workload. Another
 possible factor is that both the individual assignments and the course as a whole were
 graded on a simple pass/fail grading scale, so there were no rewards other than words
 of praise from the evaluators to be earned by surpassing the minimum requirements
 (as one student explicitly pointed out in their feedback). This is perhaps the most
 important issue to be addressed when planning future iterations of the course.
    As another student noted in their feedback, commenting on the workload, the course
 format required a lot of effort from the teachers as well, which is an important point
 since it raises a question concerning the scalability of the course. Out of an original 30
 students who signed up for the course, roughly half completed all the assignments – the
 high dropout rate perhaps another indicator that the students found the workload higher
 than expected – which kept the effort required to review the submissions manageable,
 but had all 30 completed the course, the allocated teaching resources would have been
 stretched to their limits. The issue here is that designing an alternative to the current
 assignment format that scales up for a significantly larger number of students while still
 effectively building and testing their ethical argumentation skills is far from trivial.
 Getting the students engaged in debates among themselves would both develop these
 skills and reduce the burden on the teachers, so again the crucial question is how to
 effectively facilitate such debates, given that there may be many students who are not
 comfortable with this type of work. This is another major aspect of the course that will
 need to be addressed in future development.
    Among the practices reported in the studies reviewed in Section 2, there are some
 that seem particularly useful and promising for the future development of our teaching.
 For example, as there is still a lack of established textbooks, we should maintain and
 regularly update our current reading list and make sure that it includes both scientific
 papers and news articles (the latter to illustrate the real-world consequences of AI use).
 Due to the rapid technological development and changes expected in the field, the need
 to update the list should be reviewed regularly. As engineering is a typical field with a
 strong authority-based view to knowledge, it could be fruitful to think even more about
 how to move from this type of thinking to providing our students with skills to deal
 with unforeseen ethical issues in their future work tasks. Some of the proposed methods
 in the reviewed studies could fulfill this need well and would fit our course
 implementation, such as seminars with formal defender and opponent roles or case
 study analyses. Finally, we may also consider whether we should extend AI ethics




                    Copyright © 2021 for this paper by its authors.                           32
Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0)
       Proceedings of the Conference on Technology Ethics 2021 - Tethics 2021




 teaching to reach a larger portion of the students, as well as how to make the connection
 between technical AI topics and their ethical implications clearer; in addition to our
 current standalone course, AI ethics modules or assignments could be integrated into
 technical AI courses starting from the early stages of engineering study programs.


 References
 Borenstein, J., & Howard, A. (2021). Emerging challenges in AI and the need for AI ethics
          education. AI and Ethics, 1(1), 61–65. https://doi.org/10.1007/s43681-020-00002-7
 Burton, E., Goldsmith, J., & Mattei, N. (2018). How to teach computer ethics through science
          fiction. Communications of the ACM, 61(8), 54-64. https://doi.org/10.1145/3154485
 Fiesler, C., Garrett, N., & Beard, N. (2020). What Do We Teach When We Teach Tech Ethics? A
           Syllabi Analysis. Proceedings of the 51st ACM Technical Symposium on Computer
           Science Education, 289-295. https://doi.org/10.1145/3328778.3366825
 Fink, P. (2018). Addressing the Technical, Philosophical, and Ethical Issues of Artificial
          Intelligence Through Active Learning Class Assignments. Proceedings of the AAAI
          Conference         on      Artificial    Intelligence, 32(1),     Article      1.
          https://ojs.aaai.org/index.php/AAAI/article/view/11401
 Furey, H., & Martin, F. (2019). AI education matters: A modular approach to AI ethics education.
           AI Matters, 4(4), 13-15. https://doi.org/10.1145/3299758.3299764
 Garrett, N., Beard, N., & Fiesler, C. (2020). More Than “If Time Allows”: The Role of Ethics in
           AI Education. Proceedings of the AAAI/ACM Conference on AI, Ethics, and
           Society, 272–278. https://doi.org/10.1145/3375627.3375868
 Henderson, T. (2019). Teaching Data Ethics: We’re going to ethics the heck out of this.
         Proceedings of the 3rd Conference on Computing Education Practice, 1-4.
         https://doi.org/10.1145/3294016.3294017
 Saltz, J., Skirpan, M., Fiesler, C., Gorelick, M., Yeh, T., Heckman, R., Dewar, N., & Beard, N.
             (2019). Integrating Ethics within Machine Learning Courses. ACM Transactions on
             Computing Education, 19(4), 32:1-32:26. https://doi.org/10.1145/3341164
 Shapiro, B. R., Meng, A., O’Donnell, C., Lou, C., Zhao, E., Dankwa, B., & Hostetler, A. (2020).
           Re-Shape: A Method to Teach Data Ethics for Data Science Education. Proceedings
           of the 2020 CHI Conference on Human Factors in Computing Systems, 1–13.
           https://doi.org/10.1145/3313831.3376251
 Slavkovik, M. (2020). Teaching AI Ethics: Observations and Challenges. Norsk IKT-
          Konferanse for Forskning Og Utdanning, 4, Article 4.
          https://ojs.bibsys.no/index.php/NIK/article/view/815
 Wilk, A. (2019). Teaching AI, Ethics, Law and Policy. ArXiv:1904.12470 [Cs].
         http://arxiv.org/abs/1904.12470
 Williams, T., Zhu, Q., & Grollman, D. (2020). An Experimental Ethics Approach to Robot Ethics
          Education. Proceedings of the AAAI Conference on Artificial Intelligence, 34(09),
          13428-13435. https://doi.org/10.1609/aaai.v34i09.7067




                    Copyright © 2021 for this paper by its authors.                                 33
Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0)