=Paper=
{{Paper
|id=Vol-3299/Paper09
|storemode=property
|title=Explainable and Responsible Process Prescriptive Analytics (Extended Abstract)
|pdfUrl=https://ceur-ws.org/Vol-3299/Paper09.pdf
|volume=Vol-3299
|authors=Alessandro Padella
|dblpUrl=https://dblp.org/rec/conf/icpm/Padella22
}}
==Explainable and Responsible Process Prescriptive Analytics (Extended Abstract)==
Explainable and Responsible Process Prescriptive Analytics (Extended Abstract) Alessandro Padella1 1 University of Padova Abstract Within the realm of Process Mining, Process-Aware Recommender systems (PAR systems) are information systems that aim to monitor process executions, predict their future behaviour, and find optimal corrective actions to reduce the risk of failure or to maximize a given reference Key Performance Indicator (KPI). The PAR system comprises Predictive Analytics and Prescriptive Analytics. The second part regards providing recommendations for fixing the execution of processes that are predicted to have undesired KPI values. While the research has focused on generating recommendations that aim to have better and better KPIs, my Ph.D. aims to combine the high scores of recommendations with their feasibility and fairness. Keywords Process Mining, Prescriptive Analytics, Recommender Systems, Explainable AI, Process Improvement, Bias removal 1. Introduction In the context of Process Mining, Process-Aware Recommender systems (hereafter shortened as PAR systems) are a class of Information Systems which aims to predict how the process instances are going to end and eventually recommend the corrective actions for improving their execution. This has been translated into developing frameworks that predict the outcome for each running instance of a process and suggest the next-best activity to perform, aiming to improve it when this is not satisfactory. In a general sense, an outcome is measured through the so-called Key Performance Indicator (KPI) [1, 2, 3]. While several proposals have recently been put forward to provide effective recommendations (cf.Section 3), less attention has been paid to ensure their fairness, practical feasibility wrt. the process constraints, and their comprehensibly for better engagement of process participants. This project aims to address these additional aspects. Bolzano 2022, International Conference on Process Mining, October 23-28, Bolzano, IT Envelope-Open alessandro.padella@phd.unipd.it (A. Padella) GLOBE https://github.com/Pado123 (A. Padella) © 2022 Copyright for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0). CEUR Workshop Proceedings http://ceur-ws.org ISSN 1613-0073 CEUR Workshop Proceedings (CEUR-WS.org) 41 2. Research Questions As mentioned, a PAR system’s starting point is defining the process outcomes. The outcome of a process is often measured through a customizable KPI function that, given an execution recorded in a log trace (which describes the life-cycle of a particular process instance, i.e., a case), looks at the activities executed and attribute values, returning a KPI value (e.g. the total execution time or the total cost of the procedure). Several PAR systems are limited to providing recommendations that only consist in suggesting what activity to do as next, irrespectively whether the activity is possible in that moment in time. Indeed, human process actors and other types of resources may not be available. Furthermore, the impact of the recommended activity on the process outcome may depend on the actual resource that performs the activity. From this, the following research question arises. Research question 1: How can we build a prescriptive business process analytics block that can provide more feasible recommendations? We indeed want to design a recommendation framework for Prescriptive Analytics to provide recommendations on combinations of activities and resources. It should analyze the past executions of the process and exploit some machine (or deep) learning techniques for generating recommendations. However, from the perspective of business reality, there is a wrong assumption: past execu- tions can be repeated at the moment in which we provide recommendations. Each company has its own set of rules and protocols and a pool of human resources. Over time and as the company evolves, these may change (e.g. some employees may resign, or some machines may be changed). The challenge then becomes to make the framework capable of understanding and incorpo- rating changes in organisational structures, thus avoiding making recommendations that are not actually executable. This leads to the second research question. Research question 2: How can we ensure that the context in which recommendations are generated and the context in which they are provided are compliant? Works [4] illustrate how some datasets, and so the models built on them, may be biased, especially regarding race and gender. The same problem can occur for PAR systems: it may, e.g., recommend corrective actions on cases with given characteristics. Also, PAR systems might decide to allocate activities to only certain process actors, who have better performances, thus ultimately causing them to be overloaded while others are seldom employed. This generates our third research question. Research question 3: Are the recommendations we are providing fair? Solving the research questions above, we focused on the quality of recommendations, an- swering the question “Who should perform what, for improving our KPI?”, trying to have an efficient and responsible framework. In addition, it is also crucial to accompany recommendations with an explanation of the rationale that brought the system to suggest this way. This increases the engagement and trust of the actors in the system, and thus the willingness to follow what is suggested. This brings to the fourth research question: Research question 4? How can we increase trust in our recommendations? We want to use some Explainable AI techniques to provide some explanations about not only 42 what we are suggesting but also why the algorithm took that decision. This can answer the question “Why should a resource perform the suggested activity?”. Finally, we want to understand whether the explanations provided help process actors trust our framework more. In doing so, we plan to perform some human evaluations with actual process actors, as well as we aim to use objective metrics to assess the explanation quality, e.g. discussed in [5]. Research question 5: Does the real users feel helped by our work? 3. Project Roadmap We have initially started addressing the first and the fourth research questions, integrating explanations in the prescriptive-analytics framework proposed in [1] and leveraging on Cat- Boost [6], a high-performance open source framework that has shown to provide more accurate predictions and with limited computation time if compared with the literature (see [7]). Explanations are given using the theory of Shapley values [8]. However, the explanations of the recommendations are different from those of predictions as discussed in the literature (see e.g. [3]). Indeed, the explanation of a prediction is translated into explaining how much each variable influences the final KPI value. For example, “The fact that the variable c u s t o m e r _ t y p e assumes the value G o l d contributes to decreasing the expected total time of the process instance by 120 hours”. On the other hand, the explanation of a recommendation is related to the recommended activity. Indeed, given the suggestion of a certain activity, the explanation represents how much performing that activity changes the contribution of each variable on the final KPI value. For example, “Performing the activity S e n d L e t t e r , the fact that the variable c u s t o m e r _ t y p e assumes the value Go l d goes from contributing to decreasing the expected total time by 120 hours to contributing to decreasing it by 230 hours.” Our work will be organized as follows: • Regarding research question 1, the literature in [2, 9] proposes other interesting Prescriptive-Analytics frameworks. However, just focus on recommending activities. We plan to test the goodness of recommendations using cases from the past for which compensatory actions are known and check whether the system recommends them. • For research question 2, we plan to analyze the work of De Smedt et al. in [10]. It presents a framework which aims to forecast the entire process model from historical event data, representing event data as multiple time series. We will start with this work and try to exploit similarity techniques for time-series data like Grid Representation and Matrix Distances from [11] for detecting the variation of the companies’ structure and resources. • Regarding research question 3, we aim to follow the directions indicated by Mannhardt in [12] and by van der Aalst in [13]. Specifically, we plan to specialize and adjust the de-biasing technique of Adversarial debiasing, discussed in [14]. His goal is to make the model-independent with respect to a certain variable. In it, two predictive models are trained: the first predicts your desired value, and the second takes as input the output of the first and infers what is the value of the variable whose influence we want to remove. The goal of the model-building phase will be to have the first classifier have the highest possible accuracy while that of the second has to be equivalent to that of a random, making the model capable of providing responsible recommendations. 43 • For Research question 5, we plan to work on a graphical user interface capable of rep- resenting the recommendations and their relative explanations. The literature in [5] proposes the approach of appropriate trust: an objective method which may overcome the problem of subjective evaluations of the explanations regarding the black-box models. Furthermore, we plan to combine this type of objective evaluation with a full evaluation through its associated graphical interface. From this, both real users and domain expert will test the system to which will be associated a satisfaction survey. Acknowledgments Acknowledgement. My PhD. scholarship is partly funded by IBM Italy, and by the BMCS Doctoral Program, University of Padua. This research is also supported by the Department of Mathematics, University of Padua, through the BIRD project “Data-driven Business Process Improvement” (code B I R D 2 1 5 9 2 4 / 2 1 ). References [1] M. de Leoni, M. Dees, L. Reulink, Design and evaluation of a process-aware recommender system based on prescriptive analytics, in: 2020 2nd International Conference on Process Mining (ICPM), 2020. [2] S. Weinzierl, S. Dunzer, S. Zilker, M. Matzner, Prescriptive business process monitoring for recommending next best actions, in: Business Process Management Forum, 2020. [3] R. Galanti, B. Coma-Puig, M. de Leoni, J. Carmona, N. Navarin, Explainable predictive process monitoring, in: Proceedings of the 2nd International Conference on Process Mining (ICPM 2020), 2020. [4] J. Buolamwini, T. Gebru, Gender shades: Intersectional accuracy disparities in commercial gender classification, in: Proceedings of the 1st Conference on Fairness, Accountability and Transparency, 2018. [5] H. Löfström, K. Hammar, U. Johansson, A meta survey of quality evaluation criteria in explanation methods, in: Proceedings of CAiSE Forum 2022, 2022. [6] A. V. Dorogush, V. Ershov, A. Gulin, Catboost: gradient boosting with categorical features support, in: Proceedings of the Workshop on ML Systems at NIPS 2017, 2017. [7] R. Galanti, M. de Leoni, M. Monaro, N. Navarin, A. Marazzi, B. Stasi, S. Maldera, An explainable decision support system for predictive process analytics, 2022. doi:1 0 . 4 8 5 5 0 / arXiv.2207.12782. [8] C. Molnar, Interpretable Machine Learning, 2022. Available online at https://christophm. github.io/interpretable-ml-book/. [9] Z. D. Bozorgi, I. Teinemaa, M. Dumas, M. L. Rosa, A. Polyvyanyy, Prescriptive Process Monitoring for Cost-Aware Cycle Time Reduction, in: 3rd International Conference on Process Mining (ICPM), 2021. [10] J. De Smedt, A. Yeshchenko, A. Polyvyanyy, J. Weerdt, J. Mendling, Process Model Forecasting Using Time Series Analysis of Event Sequence Data, 2021, pp. 47–61. doi:1 0 . 1 0 0 7 / 9 7 8 - 3 - 0 3 0 - 8 9 0 2 2 - 3 _ 5 . 44 [11] Y. Ye, J. Jiang, B. Ge, Y. Dou, K. Yang, Similarity measures for time series data classification using grid representation and matrix distance, Knowledge and Information Systems (2019). [12] F. Mannhardt, Responsible Process Mining, Springer International Publishing, 2022. [13] W. Aalst, Responsible data science: Using event data in a “people friendly” manner, 2017. doi:1 0 . 1 0 0 7 / 9 7 8 - 3 - 3 1 9 - 6 2 3 8 6 - 3 _ 1 . [14] D. Kenna, Using adversarial debiasing to remove bias from word embeddings, CoRR (2021). 45