=Paper=
{{Paper
|id=Vol-2903/IUI21WS-TExSS-3
|storemode=property
|title=Machine Learning Explanations as Boundary Objects: How AI Researchers Explain and Non-Experts Perceive Machine Learning
|pdfUrl=https://ceur-ws.org/Vol-2903/IUI21WS-TExSS-3.pdf
|volume=Vol-2903
|authors=Amid Ayobia,Katarzyna Stawarz,Dmitri Katz,Paul Marshall,Taku Yamagata,Raúl Santos-Rodríguez,Peter Flach,Aisling Ann O'Kane
|dblpUrl=https://dblp.org/rec/conf/iui/AyobiSKMYSFO21
}}
==Machine Learning Explanations as Boundary Objects: How AI Researchers Explain and Non-Experts Perceive Machine Learning==
Machine Learning Explanations as Boundary Objects: How AI Researchers Explain and Non-Experts Perceive Machine Learning Amid Ayobia, Katarzyna Stawarzb, Dmitri Katzc, Paul Marshalla, Taku Yamagataa, Raúl Santos-Rodrígueza, Peter Flacha, and Aisling Ann O'Kanea a University of Bristol, Bristol, England b Cardiff University, Cardiff, Wales c The Open University, Milton Keynes, England Abstract Understanding artificial intelligence (AI) and machine learning (ML) approaches is becoming increasingly important for people with a wide range of professional backgrounds. However, it is unclear how ML concepts can be effectively explained as part of human-centred and multidisciplinary design processes. We provide a qualitative account of how AI researchers explained and non-experts perceived ML concepts as part of a co-design project that aimed to inform the design of ML applications for diabetes self-care. We identify benefits and challenges of explaining ML concepts with analogical narratives, information visualisations, and publicly available videos. Co-design participants reported not only gaining an improved understanding of ML concepts but also highlighted challenges of understanding ML explanations, including misalignments between scientific models and their lived self-care experiences and individual information needs. We frame our findings through the lens of Stars and Griesemer’s concept of boundary objects to discuss how the presentation of user-centred ML explanations could strike a balance between being plastic and robust enough to support design objectives and people’s individual information needs. Keywords 1 Explainable AI, AI literacy, Explanation, Diabetes, Boundary Objects 1. Introduction and Related Work technology [30], more recent work has sought to integrate the approaches drawing not only on human-centred but also participatory HCI Understanding artificial intelligence (AI) methodologies to understanding both how AI approaches is becoming increasingly important technology is being developed and how human- for industry practitioners with a wide range of AI interactions could be designed. “What I do professional backgrounds and academic know is that the future is not AI; it can only be researchers working in interdisciplinary fields, an AI enabled through HCI,” writes Harper such as human-computer interaction (HCI). [12], reflecting on the important role HCI could While HCI and AI research have often been play in the new age of AI. In particular, the HCI characterised as having quite distinct views of community has looked at practices of the relationship between humans and researchers, data scientists, user experiences Joint Proceedings of the ACM IUI 2021 Workshops, April 13-17, 2021, College Station, USA EMAIL: amid.ayobi@bristol.ac.uk (A. 1); stawarzk@cardiff.ac.uk (A. 2); dmitrikatz23@gmail.com (A. 3); p.marshall@bristol.ac.uk (A. 4); taku.yamagata@bristol.ac.uk (A. 5), enrsr@bristol.ac.uk (A. 6); peter.flach@bristol.ac.uk (A. 7); a.okane@bristol.ac.uk (A. 8) ORCID: 0000-0003-1104-0043 (A. 1); 0000-0001-9021-0615 (A. 2); 0000-0003-1345-7539 (A. 3); 0000-0003-2950-8310 (A. 4); 0000-0001-8624-7669 (A. 5); 0000-0001-9576-3905 (A. 6); 0000-0001-6857-5810 (A. 7); 0000-0001-8219-8126 (A. 8) Copyright © 2021 for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0). CEUR Wor Pr ks hop oceedi ngs ht I tp: // ceur - SSN1613- ws .or 0073 g CEUR Workshop Proceedings (CEUR-WS.org) designers, and end-users to bridge gaps demonstrated the utility of interactive machine between HCI and AI. learning to support people with disabilities in Pointing out that the manual work and creating and customising gesturally controlled human factors of ML research can be musical interfaces through a series of overlooked, Gillies et al. [10] and Clarke et al. participatory design workshops. Although [5] encourage researchers to draw on human- participants faced challenges in understanding centred approaches to investigate the situated the training process to build instruments, they and collaborative facets of ML practices and the managed to appropriate pre-trained instruments design of usable ML support tools. Taking up according to their capabilities. this call, Muller et al. [22] unpack how data scientists develop an intuitive sense of their 2. Method datasets and how they create ground truth values as part of their data work. However, this The objective of this study was to perceived agency of working with data also has its limits. For example, based on a contextual investigate how ML explanations were inquiry, Kaur et al. [15] find that data scientists presented and perceived as part of a co-design project that aimed to co-design ML-based over-trust ML interpretability tools and face challenges to accurately describe output data decision support concepts and co-create suitable machine learning approaches. The visualisations. As ML plays an increasingly important role project involved HCI researchers, AI in the design of products, not only data researchers, and industry practitioners, as well scientists but also designers engage with ML as fifteen participants with T1 diabetes. This [17]. However, designing human-AI paper focuses on one workshop that specifically interactions entails major challenges [6, 7, 11, mediated ML concepts to workshop 31, 32]. For example, design professionals participants. We did not aim to evaluate the report difficulties in understanding ML effectiveness or efficiency of the ML explanations. Instead, we investigated the capabilities, and recommend adopting data following research questions: science jargon, including the use of quantitative evaluation methods, to be able to contribute to a data-centric work culture [31]. Envisioning a • How did AI researchers explain ML variety of feasible AI experiences and rapidly concepts to co-design workshop prototyping realistic human-AI interactions are participants? further challenges that designers are faced with, • How did co-design workshop participants considering time extensive ML training perceive the presented ML explanations? workflows and a lack of data to design with [6, • What are the transferable implications for 32, 33]. Furthermore, designers can find it designing user-centred ML explanations? difficult to productively collaborate with AI engineers because of a lack of a shared The first author conducted 18 interviews via language and methodologies that help align phone and video conference systems. human-cantered design and machine learning Interviews involved eight people with T1D who work streams [11]. participated in the co-design project (referred to Moving on from how data scientists and as P1, P2, etc.), three HCI researchers (e.g. designers work with AI concepts and tools, HCI1), and three AI researchers (e.g. AI1). To prior work has drawn on participatory support recollection before the interviews, a approaches to investigate end-users’ slide deck was shared with participants perceptions and the ethical implications of AI including ML explanations used throughout the systems [9, 21, 23, 28]. In particular, Loi et al. workshop. Interview topics covered prior [17, 18] have highlighted that participatory experiences with AI/ML and perceptions of ML design approaches are suitable to address AI explanations. Interview questions were challenges and inform AI futures: participatory adjusted for each group of interviewees and design has been shown to be a powerful lasted approximately 30 minutes. The audio methodology to explore the design space of recordings were transcribed verbatim. This desirable technologies and foster mutual interview study received an ethical approval learning between multidisciplinary actors [2, from the Faculty Ethics Committee. 24, 26, 29]. For example, Katan et al. [13] have Data collection and analysis was conducted The concept of anomaly detection was in a staggered way according to project roles. A explained with the help of two line graphs (see qualitative data analysis software was used by Figure 1). The first line graph showed the first author to thematically code data [3]. As continuous blood glucose measures over time some participants were authors, each in milligrams per decilitre. Representing a interviewee was sent the representative quotes binary machine interpretation, the second line for the codes and explicitly agreed to their use graph highlighted four anomalies in the before group analysis was conducted. The data continuous blood glucose data of the first line corpus was iteratively analysed in an inductive graph. Participants reported being used to fashion drawing on open coding by all the reflect on line graphs when using different authors [3]. health and wellbeing applications [14]. However, they wished to hear narratives that 3. Findings described the real-word context and experiences of the person who collected the data to be able to relate and make sense of the We first report on how AI researchers anomaly explanation. For example, P8 made it explained ML concepts to participants as part clear that it is important not only to understand of a co-design workshop using different types the contributing factors of anomalies but also of explanations, including analogical how anomalies could be managed: narratives, data visualisations, and publicly available videos. We then describe how “What you’re not really seeing is why those workshop participants, including HCI anomalies are happening. […] if we’re researchers and people with diabetes, perceived talking about diabetes, I think the ‘why’ is the presented ML explanations and what just as important in order to understand how benefits and challenges they experienced. to tackle those anomalies.” (P8) Moreover, participants highlighted that 3.1. ML Explanations binary representations of anomalies (see Figure 1, second line graph) may be useful to explain Since the objective of the co-design project the concept of anomaly detection, however, involved the design of ML based applications potentially not suitable to support sense-making for diabetes self-management, AI researchers and decision-making in everyday life. They felt used different methods to explain ML more comfortable with data visualisations that approaches to workshop participants, including supported their agency in identifying and data visualisations, analogies and videos of dismissing anomalies based on their lived real-world AI applications. experience. For example, high blood glucose values in daily life were not necessarily anomalous if participants were able to make educated guesses about contributing contextual 3.1.1. Data Visualisation: factors and manage these situations. Anomaly Detection 3.1.2. Analogy: Reinforcement Learning Another ML concept that was explained as part of the co-design workshops was reinforcement learning. AI researchers mediated the concept of reinforcement learning with the help of the analogy of training a dog. Figure 1: line graphs to explain anomaly detection associated with dietary challenges people with diabetes can experience: Reinforcement Learning • RL is like a dog learning good behaviours • Give a cookie for good behaviour “I’ve got dogs and I give them treats, little • Initially, the dog behaves randomly, and dog treats. I think the use of the word cookie accidentally does something good, and receives a cookie. I found amusing shall we say. Because • Then it learns how to get a cookie cookies are not a reward for us diabetics. In (situation + action) fact, that’s a challenge.” (P10) 3.1.3. Video: Agent Behaviour Figure 2: analogy to explain reinforcement learning In addition, researchers used a seminal video “At first, it was a bit like, ‘What!?’ and then, [8], that is widely cited in the machine learning when it was explained, it was like, ‘Oh, yes, that community, to demonstrate how agents learn to makes sense,’” P5 remembered, indicating that play the game of hide-and-seek. The video understanding this analogy requires translating showed how agents developed strategies and the act of training a dog to the act of training a counterstrategies over time, such as jumping on software agent that aims to maximise reward in cubes and moving cubes to block doors. All a given environment. Participants reused the participants described the video as a well- analogy of training a dog in different contexts, produced, powerful and memorable exemplar such as P8 who wished to be able to use a semi- that mediated machine learning driven multi- automated self-tracking approach [4] that agent behaviour with advanced character empowers people to manually stop false design and an entertaining narrative: machine interpretations: “The way the video showed how they sort of “So, you could use the dog example again, developed and how they learned was really where it might be learning something which clear, and the characters are quite cute, so I necessarily isn’t correct, if that makes sense, think it was quite funny as well, at the same like it might find a pattern which you don’t time. Again, that was a great example to want it to learn. So, I think... I don’t think show how machine learning can work.” it’s a question of like manually versus (P5). automatic. I think they need to work together in some shape or form. […] there needs to However, similar to the analogy of training be some sort of manual input to tell the a dog, participants found it challenging to machine learning aspect, ‘Please don’t transfer the hide-and-seek game to their learn this.’” (P8) diabetes self-management practices, highlighting that machine learning explanations Participants also perceived limitations of need not only be abstracted but also transferred using the analogy of training a dog with to a personally meaningful and research- cookies. For example, P3’s account refers to the specific context: challenges of transferring anticipated emotions, such as the desire to learn, to machines and the “I’m not sure how to transfer that to a challenges of translating the analogy to the diabetic situation in a way, that particular design space of digital applications: format. I mean there must be one, I haven’t really thought that one through. […] what “If a machine has desire or it’s how you have you got to have? You’ve got to have explain the one for a cookie. I think that’s something whereby you’re correlating the bit I find it hard to get my head round eating or carb intake, exercise and taking with a machine […] So, I don’t know how insulin. So, those three factors, I think.” (P6) you reward an app like a machine” (P3). Furthermore, P10 politely critiqued the use of the term ‘cookie’ in the context of diabetes management, considering that cookies can be 3.2. Understanding of ML Participants described AI research and AI concepts, such as ML, as data driven algorithms Explanations that are written by humans and run on computers. “AI is computers that learn, that HCI researchers and workshop participants once you set certain criteria up or whatever, reported gaining an improved understanding of they can gain knowledge themselves without the presented ML approaches. Participants being told to gain knowledge, yeah. I think that explained that even though they might not fully is the simplest form,” explained P10, understand the “inner workings” (P3) of ML referencing the learning capabilities of AI approaches, it was important to gain some technologies. Participants with diabetes also knowledge of ML concepts to develop trust in reflected on potential limitations of ML the design process and potential ML approaches, including differences between implementations, though some noted the manual and automatic data collection, roles of importance of it being presented in data quality and potential limitations of understandable terms: predictive functionalities: “I don’t think you just blindly follow stuff, “If it’s showing information based on weeks particularly when designs are being made in and weeks of data-gathering and it’s the background […] it’s better to put it into basically giving you your average day, I terms that we could understand, which is mean, I suppose that could be useful. But quite difficult when it can be so complex, but then, if you suddenly change your physical I do think it’s quite important to give us activity, or you’re eating something at a some understanding of how and what’s time that you don’t usually eat something, going on in the background.” (P5) then I guess that could disrupt it.” (P8). HCI researchers and participants reported that learning about ML approaches as part of 4. Discussion the workshops changed their prior understanding of the benefits and limitations of Understanding AI approaches is becoming ML based technologies. “Before, it was kind of increasingly important for people with a wide like, you know, computers being able to think range of professional backgrounds in industrial for themselves or like have a sentience,” P8 and academic settings. We have provided a explained, exemplifying that some participants’ qualitative account of how AI researchers prior understanding of AI was based on science explained ML concepts to HCI researchers and fiction narratives that typically portray AI people with diabetes as part of a co-design technologies with potentially dangerous project that aimed to inform the design of ML autonomous and emotional capacities. applications for diabetes self-care. Here we Reflecting on their co-design workshop discuss our findings through the lens of Stars experiences, participants demonstrated and Griesemer’s concept of boundary objects to differing degrees of ML literacy in creative outline how the presentation of user-centred ways. For example, participants used existing ML explanations could strike a balance digital consumer services as examples to between being plastic and robust enough to explain ML functionality, such as support design objectives and people’s recommendations: individual information needs as part of multidisciplinary projects. “I think the term ‘artificial intelligence’ is a bit more specific than that, I think. It’s more 4.1. Framing ML Explanations as to do with machine learning, […] So it’s things like, you know, how Netflix decides Boundary Objects what you watch, kind of thing, or how you choose your recommendation. I think it’s Star and Griesemer’s [25] concept of algorithms, really.” (P3) boundary objects has been used as a theoretical lens to understand how various actors with different backgrounds, roles, and interests successfully collaborate as part of multidisciplinary endeavours. Boundary boundary objects - robustness and plasticity - objects are artefacts that facilitate imply for the design of ML explanations. communication and collaboration between multiple actors and are defined as: 4.2. Balancing Robustness and “objects which are both plastic enough to Plasticity adapt to local needs and the constraints of the several parties employing them, yet While the robustness of a ML explanation robust enough to maintain a common can be described with features, such as being identity across sites” (ibid, p. 393). algorithmically correct and transferable to different research settings, the plasticity of a In their study of how amateurs, ML explanation can be associated with professionals, and administrators collaborate in features, such as being adaptable to people’s a museum setting, Star and Griesemer lived experiences, reflective capacities, and distinguish between four types of boundary information needs. Design techniques, such as objects: (1) repositories provide a central personalisation and customisation are location where objects, such as samples, are particularly suitable to support people’s systematically stored and are available for individual needs and experiences of agency, people to be used; (2) ideal type is an object, such as sense of identify and ownership [1]. such as a diagram, that provides an abstracted A robust and plastic enough ML explanation representation that can be adapted by others; (3) support actors, such as a co-designer, product coincident boundaries are objects, such as manager, and end-user, in making sense of and tailored maps: they are defined by common acting on a ML explanation. (geographical) boundaries but can have In our study, we have observed that different contents, purposes, and styles; (4) participants made sense of ML explanations standardised forms are boundary objects that based on their prior knowledge of AI narratives are used as formal methods of communication and technologies, reused ML explanations, across different actors. While these four types such as the analogy of training a dog, as part of of boundary objects can be used in different co-design activities, and co-created mockups ways and can have different meanings for that visualised possible ML-based different actors from different social worlds, functionalities, such as predicting blood they typically support communication and glucose values. facilitate collaborations. Although boundary An important contributing factor for objects aim to resolve conflicts, they are not adopting a ML explanation was familiarity: neutral. The creation of boundary objects participants particularly valued the analogical requires carefully managing power narrative of training a dog, since it seemed to relationships to avoid forced use of predefined help bridge the unknown concept of representations that can cause systematic reinforcement learning and the known practice exclusion, discrimination, and injustice. of training a dog. Barriers to adopting and using In our case, AI researchers used different a ML explanation seemed to be a lack of types of ML explanations to support HCI abstraction and associations with people’s lived researchers and people with diabetes in co- self-care experiences. designing possible ML systems. To foster a shared understanding of ML concepts, they 4.3. Considering Sociocultural used analogical narratives to explain reinforcement learning, data visualisations to Contexts and Ethical Implications explain anomaly detection, and publicly available videos to explain multi-agent The sociocultural underpinning of boundary behaviour. These explanations can be objects suggests that co-designing a plastic and characterised as ideal types, based on Star and robust enough ML explanation involves not Griesemer’s types of boundary objects. only representing a specific ML concept Framing these ML explanations as boundary correctly and evaluating whether the ML objects poses the question what the theory of explanation was correctly understood, but also boundary objects and the key properties of gaining a holistic and non-judgemental understanding of how the ML explanation was appropriated and experienced within a certain Content could be presented in engaging context. For example, our qualitive inquiry has ways, as demonstrated by the creative revealed the importance of tailoring general presentation of AI as a monster metaphor [7], ML explanations to specific cases, such as self- the use of tangible cards in the context of data managing diabetes, to avoid misalignments protection regulations [20], and “inspirational between people’s lived experience and bits” [27] that expose dynamic properties of scientific concepts of ML. sensors to allow designers to understand and Conceptualising ML explanations as experience the properties of technology that boundary objects means to acknowledge that might be used in research and design projects. abstraction and ambiguity can lead to divergent viewpoints, misinterpretations, and 5. Conclusion misunderstandings. Our findings suggest that gaining a good enough understanding of ML We have provided a qualitative account of explanations can support participants in developing trust in design processes, data how AI researchers explained and non-experts collection and analysis technologies, and perceived ML concepts as part of a co-design project that aimed to inform the design of ML overarching research objectives. However, what a good enough understanding is and applications for diabetes self-care. We have identified benefits and challenges whether a good enough understanding of ML explanations and functionalities is ethically of explaining ML concepts with analogical responsible depends on contextual factors, such narratives, information visualisations, and as the sensitivity of a research setting. While publicly available videos. Co-design participants with diabetes sketched predictive participants reported not only gaining an functionalities during co-design activities, AI improved understanding of ML concepts but researchers highlighted fundamental also gaining trust in the co-design process of differences between the desirability and ML based technologies, data collection and analysis technologies, and overarching research feasibility of ML-driven systems considering objectives. However, co-design participants fatal implications of false predictions and recommendations in the case of continuous also highlighted challenges of understanding blood glucose monitoring and management. ML explanations, including misalignments between scientific models of ML and their lived self-care experiences and prior knowledge of 4.4. Applying User Experience AI and ML approaches. Design Methods Based on this understanding, we have framed our findings through the lens of Stars Developing a plastic and robust enough ML and Griesemer’s concept of boundary objects to explanation can require an iterative and discuss how the presentation of user-centred multidisciplinary design process with a detailed ML explanations could maintain a delicate understanding of ML approaches, user groups, balance between being plastic and robust and the intended purpose of a ML explanation. enough to support design objectives and Considering that design methods and tools people’s individual information needs as part of to facilitate co-design are recognised multidisciplinary projects. methodological contributions [2, 16], we encourage researchers and practitioners to 6. Acknowledgements explore the design space of “learner-centered” [19] ML explanations specifically for human- This project was funded by an Innovate UK centred technology projects. Such design-led Digital Catalyst Award - Digital Health. RSR is inquiries could explore how scientific ML partially funded by the UKRI Turing AI explanations could be intertwined with people’s Fellowship EP/V024817/1. Many thanks to all lived self-care experiences and their study participants and reviewers. information needs as co-designers. These explanation instruments could represent AI/ML at a layer of abstraction above specific algorithms and communicate not just of what AI/ML can do, but also what it cannot. 7. References DOI:https://doi.org/10.1007/s42486-020- 00028-0. [1] Ayobi, A. et al. 2020. Trackly: A [10] Gillies, M. et al. 2016. Human-Centred Customisable and Pictorial Self-Tracking Machine Learning. Proceedings of the App to Support Agency in Multiple 2016 CHI Conference Extended Abstracts Sclerosis Self-Care. Proceedings of the on Human Factors in Computing Systems - 2020 CHI Conference on Human Factors CHI EA ’16 (San Jose, California, USA, in Computing Systems (CHI ’20) (2020). 2016), 3558–3565. [2] Bødker, S. et al. 2000. Co-operative [11] Girardin, F. and Lathia, N. 2017. When Design—perspectives on 20 years with ‘the User Experience Designers Partner with Scandinavian IT Design Model.’ Data Scientists. AAAI Spring Symposia proceedings of NordiCHI (2000), 22–24. (2017). [3] Braun, V. and Clarke, V. 2006. Using [12] Harper, R.H.R. 2019. The Role of HCI in thematic analysis in psychology. the Age of AI. International Journal of Qualitative Research in Psychology. 3, 2 Human–Computer Interaction. 35, 15 (Sep. (Jan. 2006), 77–101. 2019), 1331–1344. DOI:https://doi.org/10.1191/1478088706qp DOI:https://doi.org/10.1080/10447318.201 063oa. 9.1631527. [4] Choe, E.K. et al. 2017. Semi-Automated [13] Katan, S. et al. 2015. Using Interactive Tracking: A Balanced Approach for Self- Machine Learning to Support Interface Monitoring Applications. IEEE Pervasive Development Through Workshops with Computing. 16, 1 (Jan. 2017), 74–84. Disabled People. Proceedings of the 33rd DOI:https://doi.org/10.1109/MPRV.2017.1 Annual ACM Conference on Human 8. Factors in Computing Systems - CHI ’15 [5] Clarke, M.F. et al. 2019. Better Supporting (Seoul, Republic of Korea, 2015), 251–254. Workers in ML Workplaces. Conference [14] Katz, D.S. et al. 2018. Data, Data Companion Publication of the 2019 on Everywhere, and Still Too Hard to Link: Computer Supported Cooperative Work Insights from User Interactions with and Social Computing (Austin TX USA, Diabetes Apps. Proceedings of the 2018 Nov. 2019), 443–448. CHI Conference on Human Factors in Computing Systems (New York, NY, USA, [6] Dove, G. et al. 2017. UX Design 2018), 503:1-503:12. Innovation: Challenges for Working with Machine Learning as a Design Material. [15] Kaur, H. et al. 2020. Interpreting Proceedings of the 2017 CHI Conference Interpretability: Understanding Data on Human Factors in Computing Systems Scientists’ Use of Interpretability Tools for (Denver Colorado USA, May 2017), 278– Machine Learning. Proceedings of the 288. 2020 CHI Conference on Human Factors in Computing Systems (Honolulu HI USA, [7] Dove, G. and Fayard, A.-L. 2020. Apr. 2020), 1–14. Monsters, Metaphors, and Machine Learning. Proceedings of the 2020 CHI [16] Kensing, F. and Blomberg, J. 1998. Conference on Human Factors in Participatory design: Issues and concerns. Computing Systems (Honolulu HI USA, Computer Supported Cooperative Work Apr. 2020), 1–17. (CSCW). 7, 3–4 (1998), 167–185. [8] Emergent Tool Use from Multi-Agent [17] Loi, D. et al. 2019. Co-designing AI Interaction: 2019. Futures: Integrating AI Ethics, Social https://openai.com/blog/emergent-tool- Computing, and Design. Companion use/. Accessed: 2020-09-17. Publication of the 2019 on Designing Interactive Systems Conference 2019 [9] Fu, Z. and Zhou, Y. 2020. Research on Companion - DIS ’19 Companion (San human–AI co-creation based on reflective Diego, CA, USA, 2019), 381–384. design practice. CCF Transactions on Pervasive Computing and Interaction. 2, 1 [18] Loi, D. et al. 2018. PD manifesto for AI (Mar. 2020), 33–41. futures. Proceedings of the 15th Participatory Design Conference on Short [26] Steen, M. 2013. Co-design as a process of Papers, Situated Actions, Workshops and joint inquiry and imagination. Design Tutorial - PDC ’18 (Hasselt and Genk, Issues. 29, 2 (2013), 16–28. Belgium, 2018), 1–4. [27] Sundström, P. et al. 2011. Inspirational [19] Long, D. and Magerko, B. 2020. What is bits: towards a shared understanding of the AI Literacy? Competencies and Design digital material. Proceedings of the 2011 Considerations. Proceedings of the 2020 annual conference on Human factors in CHI Conference on Human Factors in computing systems - CHI ’11 (Vancouver, Computing Systems (Honolulu HI USA, BC, Canada, 2011), 1561. Apr. 2020), 1–16. [28] Trewin, S. et al. 2019. Considerations for [20] Luger, E. et al. 2015. Playing the Legal AI fairness for people with disabilities. AI Card: Using Ideation Cards to Raise Data Matters. 5, 3 (Dec. 2019), 40–63. Protection Issues within the Design DOI:https://doi.org/10.1145/3362077.3362 Process. Proceedings of the 33rd Annual 086. ACM Conference on Human Factors in [29] Vartiainen, H. et al. 2020. Machine Computing Systems - CHI ’15 (Seoul, learning for middle-schoolers: Children as Republic of Korea, 2015), 457–466. designers of machine-learning apps. 2020 [21] Morrison, C. et al. 2017. Imagining IEEE Frontiers in Education Conference Artificial Intelligence Applications with (FIE) (Uppsala, Sweden, Oct. 2020), 1–9. People with Visual Disabilities using [30] Winograd, T. 2006. Shifting viewpoints: Tactile Ideation. Proceedings of the 19th Artificial intelligence and human–computer International ACM SIGACCESS interaction. Artificial Intelligence. 170, 18 Conference on Computers and Accessibility (Dec. 2006), 1256–1258. (Baltimore Maryland USA, Oct. 2017), 81– DOI:https://doi.org/10.1016/j.artint.2006.1 90. 0.011. [22] Muller, M. et al. 2019. How Data Science [31] Yang, Q. et al. 2018. Investigating How Workers Work with Data: Discovery, Experienced UX Designers Effectively Capture, Curation, Design, Creation. Work with Machine Learning. Proceedings Proceedings of the 2019 CHI Conference of the 2018 on Designing Interactive on Human Factors in Computing Systems - Systems Conference 2018 - DIS ’18 (Hong CHI ’19 (Glasgow, Scotland Uk, 2019), 1– Kong, China, 2018), 585–596. 15. [32] Yang, Q. et al. 2020. Re-examining [23] Muller, M. and Liao, Q.V. 2017. Exploring Whether, Why, and How Human-AI AI Ethics and Values through Participatory Interaction Is Uniquely Difficult to Design. Design Fictions. Human Computer Proceedings of the 2020 CHI Conference Interaction Consortium. (2017). on Human Factors in Computing Systems [24] Slattery, P. et al. 2020. Research co-design (Honolulu HI USA, Apr. 2020), 1–13. in health: a rapid overview of reviews. [33] Yang, Q. et al. 2019. Sketching NLP: A Health Research Policy and Systems. 18, 1 Case Study of Exploring the Right Things (Dec. 2020), 17. To Design with Language Intelligence. DOI:https://doi.org/10.1186/s12961-020- Proceedings of the 2019 CHI Conference 0528-9. on Human Factors in Computing Systems - [25] Star, S.L. and Griesemer, J.R. 1989. CHI ’19 (Glasgow, Scotland Uk, 2019), 1– Institutional ecology,translations’ and 12. boundary objects: Amateurs and professionals in Berkeley’s Museum of Vertebrate Zoology, 1907-39. Social studies of science. 19, 3 (1989), 387–420.