<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Incorporating Explainable Artificial Intelligence (XAI) to aid the Understanding of Machine Learning in the Healthcare Domain</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Urja</forename><surname>Pawar</surname></persName>
							<email>urja.pawar@mycit.ie</email>
							<affiliation key="aff0">
								<orgName type="institution">Cork Institute of Technology</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Donna</forename><surname>O'shea</surname></persName>
							<email>donna.oshea@cit.ie</email>
							<affiliation key="aff0">
								<orgName type="institution">Cork Institute of Technology</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Susan</forename><surname>Rea</surname></persName>
							<email>susan.rea@cit.ie</email>
							<affiliation key="aff0">
								<orgName type="institution">Cork Institute of Technology</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Ruairi</forename><surname>O'reilly</surname></persName>
							<email>ruairi.oreilly@cit.ie</email>
							<affiliation key="aff0">
								<orgName type="institution">Cork Institute of Technology</orgName>
							</affiliation>
						</author>
						<title level="a" type="main">Incorporating Explainable Artificial Intelligence (XAI) to aid the Understanding of Machine Learning in the Healthcare Domain</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">36B4385023C5978F9AB0A2A6E919EE95</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-25T02:39+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Explainable Artificial Intelligence</term>
					<term>Healthcare</term>
					<term>Feature Importance</term>
					<term>Decision trees</term>
					<term>Explainable Underlying Workflow</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>In the healthcare domain, Artificial Intelligence (AI) based systems are being increasingly adopted with applications ranging from surgical robots to automated medical diagnostics. While a Machine Learning (ML) engineer might be interested in the parameters related to the performance and accuracy of these AI-based systems, it is postulated that a medical practitioner would be more concerned with the applicability, and utility of these systems in the medical setting. However, medical practitioners are unlikely to have the prerequisite skills to enable reasonable interpretation of an AI-based system. This is a concern for two reasons. Firstly, it inhibits the adoption of systems capable of automating routine analysis work and prevents the associated productivity gains. Secondly, and perhaps more importantly, it reduces the scope of expertise available to assist in the validation, iteration, and improvement of AI-based systems in providing healthcare solutions. Explainable Artificial Intelligence (XAI) is a domain focused on techniques and approaches that facilitate the understanding and interpretation of the operation of ML models. Research interest in the domain of XAI is becoming more widespread due to the increasing adoption of AI-based solutions and the associated regulatory requirements <ref type="bibr" target="#b0">[1]</ref>. Providing an understanding of ML models is typically approached from a Computer Science (CS) perspective <ref type="bibr" target="#b1">[2]</ref> with a limited research emphasis being placed on supporting alternate domains <ref type="bibr" target="#b2">[3]</ref>. In this paper, a simple, yet powerful solution for increasing the explainability of AI-based solutions to individuals from non-CS domains (such as medical practitioners), is presented. The proposed solution enables the explainability of ML models and the underlying workflows to be readily integrated into a standard ML workflow. Central to this solution are feature importance techniques that measure the impact of individual features on the outcomes of AI-based systems. It is envisaged that feature importance can enable a high-level understanding of a ML model and the workflow used to train the model. This could aid medical practitioners in comprehending AI-based systems and enhance their understanding of ML models' applicability and utility.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>Interpretability is the degree to which the rationale of a decision can be observed within a system <ref type="bibr" target="#b3">[4]</ref>. If a ML model's operation is readily understood then the model is interpretable. Explainability is the extent to which the internal operation of a system can be explained in human terms. XAI is comprised of methodologies for making AI systems interpretable and explainable <ref type="bibr" target="#b4">[5]</ref>.</p><p>The context of interpretability and explainability is generally considered domain-specific in an applied setting. For instance, a ML engineer and a medical practitioner would have a different perspective on what is "explainable" when viewing the same system. Interpretability from the perspective of the ML engineer relates to understanding the internal working of a system so that the technical parameters can be tuned to improve the overall performance. Interpretability from the medical practitioner's perspective would relate to a higher-level understanding of the internal operation of a system as it relates to the medical function it provides. Explainability for a ML engineer may relate to presenting technical information in an understandable format that enables effective evaluation of a system while explainability for medical practitioners may be more related to the rationale as to why a course of action is prescribed for a patient.</p><p>It is postulated that AI-based systems need to accommodate a medical practitioner's perspective to be considered explainable in a healthcare setting. This presents several challenges which are highlighted and addressed as part of this work:</p><p>Designing domain-agnostic systems with XAI and simultaneously accommodating multiple perspectives is a complex problem because explanations require a context of the domain (engineering, medicine, or healthcare) and can be useful for a targeted perspective but trivial for others. For instance, presenting interactive visualisations to explain layers of a neural network is beneficial for ML engineers but of less importance to the radiologists who use the neural network for analysing MRI scans.</p><p>The scope of interpretability and explainability for AI-based solutions is broader than the operation of a ML model. It also concerns the workflow adopted to train these models. The workflow can provide technical knowledge regarding the pre-processing steps, the ML models used, and the evaluation criteria (e.g. accuracy, precision) to the ML engineer. It can benefit medical practitioners with an overview of the underlying data, the model's interpretation of the data, and the performance metrics pertinent to medical diagnostics. For instance, the ML models that are used to predict based on a patient's medical record might be inappropriate if the underlying training data does not include records from similar demographics.</p><p>The subjective nature of XAI in medical setting presents challenges such as as the association of a trained model's knowledge with the medical features, the provisioning of explanations with regard to the underlying medical dataset <ref type="bibr" target="#b5">[6]</ref>, and an understanding of how the presence or absence of some medical features' information affects a model's performance and its interpretation of features.</p><p>There are several nuanced issues related to the challenges articulated. These include: (a) lack of explainability in underlying feature engineering processes to incorporate clinical expertise; (b) complexity in the integration of XAI approaches with existing ML workflows <ref type="bibr" target="#b0">[1]</ref>; (c) a lack of a high-level explainability of the data and the ML model <ref type="bibr" target="#b6">[7]</ref>; (d) and a lack of explainability of a model's operation in different medical settings.</p><p>A standard ML workflow consists of several stages: data collection, data pre-processing, modeling, training, evaluation, tuning, and deployment. XAI approaches should endeavor to integrate interpretability and explainability into the standard ML workflow. Feature Importance (FI) is a set of techniques that assign weightings(scores) to each feature indicating their relative importance in making a prediction or classification by a ML model <ref type="bibr" target="#b7">[8]</ref>. FI techniques are typically used as part of the data pre-processing to enhance feature selection.</p><p>Moving towards a solution: While addressing the challenges articulated in their totality is beyond the scope of this paper, addressing the nuanced issues outlined will provide the initial steps for a more complete solution to be derived and is the primary contribution of this work. In this paper, FI techniques are utilised as a means of enabling XAI. It is envisaged that FI will provide a simple but powerful means of integrating XAI into the standard ML workflow in a domain-agnostic manner. The approach can enable the explainability of a ML model as well as the underlying workflow whilst accommodating multiple perspectives. This is realised by three proposed approaches that utilise the associations between FI scores, FI techniques, the inclusion/exclusion of features, data augmentation techniques, and performance metrics. In doing so, it enables multiple levels of explainability encapsulating the operation of the ML model with different underlying datasets in different medical settings. The explainability derived is expected to enable the clinical validation of AI-based systems as discussed in the following sections.</p><p>The remainder of the paper is organised as follows: Section 2 presents related work with regards XAI, FI, and its utilisation in an applied ML setting. Section 3 outlines the proposed methodology for enabling XAI in a standard ML workflow. Section 4 outlines the results of the experimental work of the approaches proposed. Section 5 presents a discussion and concluding remarks arising from the work carried out to date.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Related Work</head><p>Preliminary work for making ML models used in clinical domains increasingly interpretable and explainable has been initiated in <ref type="bibr" target="#b0">[1,</ref><ref type="bibr" target="#b4">5,</ref><ref type="bibr" target="#b8">9,</ref><ref type="bibr" target="#b9">10]</ref>. The interpretability and explainability of ML models enable ML engineers to understand, and evaluate, a model's parameters (weights/coefficients) and hyper-parameters (inputsize, number of layers) with the model's outcomes (predictions/classifications). It can also enable medical practitioners to effectively comprehend and validate the output derived from ML models as per their medical expertise <ref type="bibr" target="#b0">[1,</ref><ref type="bibr" target="#b4">5]</ref>.</p><p>There exists a variety of XAI methods that are applicable to the medical domain. Ante-hoc XAI methods achieve interpretability without an additional step that makes them easier to adopt in existing ML workflows. They include inherently interpretable ML models such as Decision trees <ref type="bibr" target="#b10">[11]</ref>, Random Forests <ref type="bibr" target="#b10">[11]</ref> and Generalised Additive Models (GAMs) <ref type="bibr" target="#b8">[9]</ref>. They are typically used to achieve interpretability at the cost of lower performance scores as compared to complex ML models. However, their contribution towards enabling the explainability to non-CS perspectives in different domains is not extensively discussed in the literature <ref type="bibr" target="#b0">[1]</ref>.</p><p>In <ref type="bibr" target="#b11">[12]</ref>, an XAI-enabled framework to include clinical expertise in AI-based systems is proposed in an abstract format. This work also discusses the use of FI to enable the inclusion of clinical expertise when building AI-based solutions. In <ref type="bibr" target="#b10">[11]</ref> FI scores based on Decision trees were used to analyse the importance of features in classifying cervical cancer and achieving interpretability in the model. However, the interpretability in relation to the underlying dataset was not discussed. Also, the utilisation of the FI scores to enable explainability from the perspective of medical practitioners was not addressed. In <ref type="bibr" target="#b12">[13]</ref> Random forests were used to classify arrhythmia from time-series ECG data and FI scores were presented as a means of achieving interpretability. However, as the time-series ECG data has numeric values for each sampled record, the FI scores assigned to each time-stamped value were not useful as effective conclusions cannot be drawn by associating a FI score with a single amplitude value in a time-series ECG wave.</p><p>Post-hoc XAI methods are specifically designed for explainability and are applied after a ML model is trained. This makes post-hoc methods difficult to adopt but they are advantageous as they typically support multiple noninterpretable but performant classifiers <ref type="bibr" target="#b4">[5]</ref>. Local Interpretable Model-agnostic Explanations (LIME) is one of the commonly used post-hoc XAI methods that was developed to explain the predictions of any ML classifier by calculating FI scores based on some assumptions that don't always hold true across different types of classifiers <ref type="bibr" target="#b13">[14]</ref>. Shapley values are another post-hoc XAI technique that was initially introduced in game theory to present the average expected marginal contribution of a player in achieving a payout when all possible combinations of players are considered <ref type="bibr" target="#b14">[15]</ref>. In XAI, Shapley values are used to assign FI scores to features (players) in achieving predictions (payout) made by a model. In <ref type="bibr" target="#b15">[16]</ref> LIME and Shapley's FI scores were compared and it was found that Shapley's FI scores were more consistent when compared to LIME's. This consistency was derived on the basis of objective criteria including similarity, identity, and separability, which are important considerations when generating and providing explanations in a healthcare setting.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Methodology</head><p>The workflow adopted in this work is depicted in Figure <ref type="figure" target="#fig_0">1</ref>. It follows a standard ML workflow with the addition of the FI stage to enable post-hoc explainability. The FI scores are calculated without modifying prior stages of the workflow and are utilised to enable the explainability of the model and the inherent workflow. Three approaches are proposed that utilise FI scores to enable explainability and interpretability of a ML model and the underlying dataset: A1. Relative feature ranking: There needs to be a careful validation regarding features that are considered more or less relevant by ML models in healthcare <ref type="bibr" target="#b8">[9]</ref>. This approach derives FI scores using two distinct methods. The first is generated using Decision tree FI scores, the FI score of a feature based on its position in the conditional flow of a classification process, and the second is generated using Shapley values, based on weighting the feature's impact on the model's outcome. The derived FI scores are collated and sorted in descending order. This provides a high-level understanding of how a ML model ranks different features to be considered while deriving an outcome.</p><p>This enables a comparison between features that are considered important by the classification model (realised by the first approach) and the features that fluctuate the ML model's outcome (realised by the second approach). This enables explainability to be derived as it provides a relative ranking of the features as interpreted by a ML model along with their impact on the outcome. In the applied setting, this can be used by medical practitioners to gain an understanding of the features that are critical in formulating a medical diagnosis, highlighting features whose values cannot be ignored due to their high impact on the model's output.</p><p>A2. Feature importance in different medical settings: The availability of medical information in different medical settings is not uniform (e.g. lack of advanced medical tests in small clinics) and therefore, approaches followed by medical practitioners belonging to different medical settings differ. This reduces the associated utility of AI-based solutions. A gold-standard solution should be designed to include all the relevant data while providing multiple versions to acknowledge that different healthcare facilities will have different levels of access to this data. This realisation dramatically broadens the applicability and utility of the solution as it acknowledges the inclusion and exclusion of features in different settings.</p><p>This approach demonstrates the relative change of FI scores and performance metrics based on the inclusion/exclusion of features. This enables a broader un-derstanding of a ML model and highlights its suitability to different medical settings (e.g. a general practitioner in a clinic and an emergency room doctor in a hospital will have access to significantly different levels of data regarding an individual's health). If a ML model is trained upon a set of n features, explainability can be derived by training the model on all possible subsets (2 n ) of features and can enhance the understanding of how features are re-ranked, and performance is affected, based on the inclusion/exclusion of features.</p><p>In an applied setting, this approach is useful to medical practitioners as it aids their understanding based on inclusion/exclusion of clinical test results or medical information with an associated performance score. This enables an informed evaluation regarding the suitability of the AI-based solution on a per actor basis.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>A3. Understanding the Data:</head><p>The data on which a model was trained and how it was pre-processed can have significant consequences in a medical setting <ref type="bibr" target="#b5">[6,</ref><ref type="bibr" target="#b6">7]</ref>. As such, this approach demonstrates the association of FI scores and performance metrics with the data augmentation techniques that are used in a dataset. This association provides explainability by enabling an understanding of how differences in underlying data impact the performance metrics and the FI scores.</p><p>In an applied setting, the medical practitioners can validate the ranking of features as interpreted by a ML model trained on data augmented using different techniques and be able to associate it with the corresponding performance metrics. This furthers the interpretability of the underlying workflow used for processing the data and can enable better selection of augmentation techniques by incorporating clinical expertise along with the expected performance metrics.</p><p>The dataset and the modelling technique utilised for experimental work are discussed in Section 3.1 and 3.2 respectively. In Section 3.3, the two FI techniques: one based on Decision trees and the other based on Shapley values used in this work are discussed.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">Dataset</head><p>In this work, the "Cervical Cancer Risk Factors" dataset available from the UCI data repository is used <ref type="bibr" target="#b16">[17]</ref>. This dataset was used in <ref type="bibr" target="#b17">[18]</ref> to train different ML models to predict the occurrence of cervical cancer based on a person's health record. The performance of different models was compared based on accuracy, precision, recall, and F-score (harmonic mean of precision and recall) values <ref type="bibr" target="#b17">[18]</ref>. The work did not address the interpretability and explainability of ML models and the underlying workflow.</p><p>The dataset contains 36 feature attributes representing risk factors responsible for causing cervical cancer and the results of some preliminary and advanced medical tests. In the dataset, 803 out of 858 records have a negative Biopsy result while 55 have a positive result. The class-imbalance problem is addressed using Imbalanced-learn that offers many data sampling techniques to balance the number of the majority and minority classes <ref type="bibr" target="#b18">[19]</ref>. Table <ref type="table" target="#tab_0">1</ref>   <ref type="bibr" target="#b17">[18]</ref>. Legend: Number of Samples (Sam.), Biopsy results ratio -Positive (0): Negative (1).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">ML Model</head><p>Decision trees are graphs where nodes represent sets of data samples and edges represent conditions. Each node has an associated impurity factor indicating the diversity of classes/labels in that node. A node is pure if all the data samples present in it belong to the same class/label. In a classification problem, the conditions in the edges of Decision trees are designed to decrease the impurity. Therefore, from root node to leaf nodes, the impurity factor decreases and each leaf node should contain data samples that are classified under a single class/label. Decision trees are more interpretable as compared to complex models such as Support Vector Machines or Neural Networks <ref type="bibr" target="#b4">[5]</ref>. They also provide sufficient performance scores in the given dataset <ref type="bibr" target="#b17">[18]</ref>. In this work, a Decision tree was chosen as it achieves sufficient performance while retaining interpretability <ref type="bibr" target="#b0">[1]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3">Feature Importance (FI)</head><p>FI identifies the important features as considered by a ML model from a dataset for making a classification or prediction. In this paper, FI using decision trees and Shapley were used. When a decision tree is trained, FI scores can be calculated by measuring how much a feature contributes towards a decrease in the impurity <ref type="bibr" target="#b19">[20]</ref>. The FI scores obtained represent features considered important by the Decision tree model. Shapley values can be used to generate the FI score of a feature by first calculating a model's output including and excluding that feature to get the contribution of that feature alone. This contribution is then weighted in presence of all subsets of features. This whole process is summed for all the subsets of features to get a weighted and permuted FI score. The FI scores obtained represent the impact of different features on a model's outcome.</p><p>The main challenge in achieving explainability using FI was to present the FI scores generated after training the model with different data augmentation techniques and different feature sets in an integrated manner such that their association with the performance metrics (e.g. accuracy, F-scores) and relative ranking of the features can be effectively utilised in a domain-agnostic manner. This integration of information enables explainability of both: the ML model and the underlying data, thereby broadening the scope of explainability.</p><p>The three approaches outlined in Section 3 have been implemented. The value of the approaches and the derived explainability is demonstrated in this section. This is considered a contribution towards a long-term generic workflow for simplifying the integration of XAI in an applied setting such as healthcare.</p><p>A1: Relative Feature Ranking When the ML model was trained on data sampled using Random Over Sampling, FI scores assigned to different features were plotted as depicted in Figure <ref type="figure" target="#fig_1">2</ref>. Random over sampling provided higher accuracy than other sampling techniques <ref type="bibr" target="#b17">[18]</ref>, as such it was selected for this approach. In Figure <ref type="figure" target="#fig_1">2</ref>, the feature Schiller Test was omitted due to its high correlation with the biopsy results as it is an advanced medical test conducted to diagnose cervical cancer <ref type="bibr" target="#b10">[11]</ref>. It can be observed that the feature Hinselmann is the highest-ranked feature by both FI approaches. There is a similarity between the two sets of FI scores obtained as features considered important by the model (represented by Decision trees based FI) will automatically have a higher impact on its outcome (represented by Shapley based FI). The value derived from these FI scores is that a medical practitioner can understand and validate the ranking of features. This enables the incorporation of clinical expertise to improve feature engineering processes and achieve improved models for future use.</p><p>A2: FI in Different medical settings The random over sampled data was used to train multiple instances of the model, each time on a subset of features that excluded the highest-ranked feature from the previous instance. The resulting FI scores assigned to individual features are indicative of the impact the omission of the highest-ranked feature from a prior instance has on the ML model and is depicted in Figure <ref type="figure" target="#fig_2">3</ref>.</p><p>The change in the relative ranking of different features can be observed on the omission of the highest-ranked feature. For instance, when all the features were present (All features), Schiller (orange segment) was given the highest importance followed by Age (yellow segment). When feature Schiller is omitted (second bar), it was noticed that Hinselmann (grey segment) was given the highest importance instead of Age. Thus the inclusion or exclusion of a feature does not behave in an ordered fashion as dictated by a gold-standard approach that includes all features. Furthermore, the compounded omission of the highest-ranked features (left to right) significantly reduces the total sum of FI scores assigned to features in each instance (≥0.7 to ≤0.3). This is accompanied by a reduction in performance metrics such as F-scores (denoted at the top of each bar) and indicates less accurate models due to the absence of more important features or the presence of less important features. This approach warrants the derivation of multiple instances of a single model such that the relationship among features can be fully understood and can be validated with clinical expertise. Based on a threshold value of performance metrics, a medical practitioner can select a ML model that is trained with the features that are accessible to his/her medical setting and assigns appropriate importance scores to the available features while generating an outcome.</p><p>A3: Understanding Underlying Data The model was trained on the data sampled using different sampling techniques as discussed in Section 3.1 and FI scores were plotted corresponding to each of the sampled versions as depicted in Figure <ref type="figure" target="#fig_3">4</ref>. FI scores relating to a particular type of sampled data are assigned a particular color. Performance metrics corresponding to each of the sampling techniques are noted in the legend. This approach enables the interpretability of the underlying dataset by presenting the difference in FI scores when using data augmented using different techniques. For instance, in Figure <ref type="figure" target="#fig_3">4</ref>, there is a lack of similarity in the FI scores associated with under-sampling techniques (e.g. NCUS, RUS) as compared to over/combination-sampling techniques (e.g. S-TOM, ROS). The lesser the volume of data generated using under-sampling techniques the less diverse the values of a feature. This is evident when comparing the sorted ordering of FI scores in under-sampling techniques to over/combination sampling techniques.</p><p>As depicted in Figure <ref type="figure" target="#fig_3">4</ref>, the under-sampled data provided less accuracy and recall values ( 70-90%) compared to the over/combination-sampled data ( 93-97%). The association of performance metrics aligned with FI scores enables the explainability to validate the suitability of datasets from a domain-specific perspective. In contrast to over/combination sampled data, in the under-sampled data augmented using the NCUS technique (dark-red bars), Age is assigned a higher FI score than the Cytology test which would be considered an invalid approach as a Cytology test is a diagnostic aid with a high level of efficacy when detecting cervical cancer <ref type="bibr" target="#b20">[21]</ref>. A medical practitioner should disregard the use of NCUS data due to its invalid FI ranking along with the low-performance scores.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Conclusions and Future Work</head><p>XAI is a crucial tool for enabling medical practitioners to understand and evaluate AI-based solutions effectively in the healthcare domain. It provides additional benefits in the form of increased confidence in solutions being adopted amongst medical practitioners and increased exposure to the operation of the solutions.</p><p>In this paper, an alternative perspective regarding how FI scores can be integrated into a ML workflow is adopted. FI scores are used to surface pertinent information relating to associations between features, models, and data to provide explainability. This perspective is realised in three distinct approaches.</p><p>A1) A model/output-based perspective with regards to the relative ranking of a feature, this informs the medical practitioner which features the model considers most important and which features fluctuate the model outcome. A2) Relative feature ranking in different medical settings, this incorporates a hierarchical perspective which considers diagnostic capacity in the form of feature inclusion/exclusion aligning it more closely to the real world. This informs the medical practitioner how the model will perform and rank features in different medical settings enabling a more informed interpretation of a model's operation. A3) The impact of data augmentation approaches on the performance of a model and the validity of their FI scores in a medical setting. This informs the medical practitioner how suitable the augmented data is and how valid it is in a medical setting. The simple but powerful nature of FI enables the applicability of the three approaches proposed in a domain-agnostic manner.</p><p>It is intended to extend the work by developing a framework that automates the training and validation of models appropriate to the intended level of a hierarchy in order to enable explainability from a multi-level perspective. The workflow comprising that hierarchy will empirically evaluate the applicability of combining XAI and recommendations to increase operational efficacy.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Fig. 1 .</head><label>1</label><figDesc>Fig.1. Integration of FI stage in a standard ML workflow to enable derivation of explainability using three proposed approaches: A1, A2, and A3</figDesc><graphic coords="5,134.77,115.84,345.82,110.03" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Fig. 2 .</head><label>2</label><figDesc>Fig. 2. A1: Relative Feature Ranking using FI scores generated by Decision trees and Shapley based FI techniques. The two bars represent the FI scores assigned to individual features by the two FI techniques to provide the basis for the comparison.</figDesc><graphic coords="8,160.70,249.79,293.96,191.48" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Fig. 3 .</head><label>3</label><figDesc>Fig. 3. A2: Impact of excluding the highest ranked features on FI scores, F-scores, and Ordering of features based on FI score</figDesc><graphic coords="9,160.70,264.91,293.96,207.89" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Fig. 4 .</head><label>4</label><figDesc>Fig. 4. A3: Ranking of features (by Decision trees and Shapley based FI) on using different data sampling techniques and corresponding performance metrics. Performance metrics are noted as follows Accuracy (ACC ), Precision (Prec) and Recall (Rec).</figDesc><graphic coords="10,160.70,211.71,293.96,191.48" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 .</head><label>1</label><figDesc>denotes the number of records corresponding to positive and negative biopsy results after a sampling technique is applied. Number of samples in different data sampling techniques</figDesc><table><row><cell cols="3">Resampling Method</cell><cell></cell><cell cols="2">Sam. 0:1 Ratio</cell></row><row><cell cols="3">Random Over sampling (ROS)</cell><cell></cell><cell cols="2">1606 803:803</cell></row><row><cell cols="4">Adaptive Synthetic Over sampling (ASS)</cell><cell cols="2">1606 803:803</cell></row><row><cell cols="3">Random Under sampling (RUS)</cell><cell></cell><cell>110</cell><cell>55:55</cell></row><row><cell>Neighbourhood</cell><cell>cleaning</cell><cell>Under</cell><cell>sampling</cell><cell>725</cell><cell>670:55</cell></row><row><cell>(NCUS)</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell cols="6">SMOTEtomek Combination sampling (S-TOM) 1600 800:800</cell></row><row><cell cols="4">SMOTE edited nearest neighbours Combination</cell><cell cols="2">1429 652:777</cell></row><row><cell>sampling (S-ENN)</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row></table></figure>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Acknowledgement: This publication has emanated from research co-sponsored by McKesson and Science Foundation Ireland under Grant number SFI CRT 18/CRT/6222.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">What do we need to build explainable AI systems for the medical domain?</title>
		<author>
			<persName><forename type="first">Andreas</forename><surname>Holzinger</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Chris</forename><surname>Biemann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Constantinos</forename><forename type="middle">S</forename><surname>Pattichis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Douglas</forename><forename type="middle">B</forename><surname>Kell</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Ml</title>
		<imprint>
			<biblScope unit="page" from="1" to="28" />
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">What&apos;s inside the black-box? A genetic programming method for interpreting complex machine learning models</title>
		<author>
			<persName><forename type="first">P</forename><surname>Benjamin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Bing</forename><surname>Evans</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Mengjie</forename><surname>Xue</surname></persName>
		</author>
		<author>
			<persName><surname>Zhang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">GECCO 2019 -Proceedings of the 2019 Genetic and Evolutionary Computation Conference</title>
				<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="1012" to="1020" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Designing Theory-Driven User-Centric Explainable AI</title>
		<author>
			<persName><forename type="first">Danding</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Qian</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ashraf</forename><surname>Abdul</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Brian</forename><forename type="middle">Y</forename><surname>Lim</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems -CHI &apos;19</title>
				<meeting>the 2019 CHI Conference on Human Factors in Computing Systems -CHI &apos;19</meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="1" to="15" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<monogr>
		<author>
			<persName><forename type="first">Tim</forename><surname>Miller</surname></persName>
		</author>
		<title level="m">Explanation in artificial intelligence: Insights from the social sciences</title>
				<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">A Survey on Explainable Artificial Intelligence (XAI)</title>
		<author>
			<persName><forename type="first">Erico</forename><surname>Tjoa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Cuntai</forename><surname>Guan</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Towards Medical XAI</title>
		<imprint>
			<biblScope unit="volume">1</biblScope>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">The medical AI insurgency: what physicians must know about data to practice with intelligent machines</title>
		<author>
			<persName><forename type="first">Douglas</forename><surname>Miller</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">npj Digital Medicine</title>
		<imprint>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page">62</biblScope>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Rethinking pca for modern data sets: Theory, algorithms, and applications [scanning the issue</title>
		<author>
			<persName><forename type="first">Namrata</forename><surname>Vaswani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Yuejie</forename><surname>Chi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Thierry</forename><surname>Bouwmans</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the IEEE</title>
				<meeting>the IEEE</meeting>
		<imprint>
			<date type="published" when="2018">2018</date>
			<biblScope unit="volume">106</biblScope>
			<biblScope unit="page" from="1274" to="1276" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Feature importance for machine learning redshifts applied to sdss galaxies</title>
		<author>
			<persName><forename type="first">Ben</forename><surname>Hoyle</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Markus</forename><forename type="middle">Michael</forename><surname>Rau</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Roman</forename><surname>Zitlau</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Stella</forename><surname>Seitz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jochen</forename><surname>Weller</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Monthly Notices of the Royal Astronomical Society</title>
		<imprint>
			<biblScope unit="volume">449</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="1275" to="1283" />
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Intelligible models for healthcare: Predicting pneumonia risk and hospital 30-day readmission</title>
		<author>
			<persName><forename type="first">Rich</forename><surname>Caruana</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Yin</forename><surname>Lou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Johannes</forename><surname>Gehrke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Paul</forename><surname>Koch</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Marc</forename><surname>Sturm</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Noemie</forename><surname>Elhadad</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 21th ACM SIGKDD international conference on knowledge discovery and data mining</title>
				<meeting>the 21th ACM SIGKDD international conference on knowledge discovery and data mining</meeting>
		<imprint>
			<date type="published" when="2015">2015</date>
			<biblScope unit="page" from="1721" to="1730" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<title level="m" type="main">Explainable ai meets healthcare: A study on heart disease dataset</title>
		<author>
			<persName><forename type="first">Devam</forename><surname>Dave</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Het</forename><surname>Naik</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Smiti</forename><surname>Singhal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Pankesh</forename><surname>Patel</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Analysis of risk factors for cervical cancer based on machine learning methods</title>
		<author>
			<persName><forename type="first">X</forename><surname>Deng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Luo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Wang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">5th IEEE International Conference on Cloud Computing and Intelligence Systems (CCIS)</title>
				<imprint>
			<date type="published" when="2018">2018. 2018</date>
			<biblScope unit="page" from="631" to="635" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Explainable ai in healthcare</title>
		<author>
			<persName><forename type="first">U</forename><surname>Pawar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>O'shea</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Rea</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>O'reilly</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">2020 International Conference on Cyber Situational Awareness, Data Analytics and Assessment (CyberSA)</title>
				<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="1" to="2" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<monogr>
		<title level="m" type="main">Interpretable machine learning models for assisting clinicians in the analysis of physiological data</title>
		<author>
			<persName><forename type="first">Urja</forename><surname>Nisha</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ruairi O'</forename><surname>Pawar</surname></persName>
		</author>
		<author>
			<persName><surname>Reilly</surname></persName>
		</author>
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<monogr>
		<title level="m" type="main">Why Should I Trust You ?&quot; Explaining the Predictions of Any Classifier</title>
		<author>
			<persName><forename type="first">Marco</forename><surname>Tulio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ribeiro</forename></persName>
		</author>
		<author>
			<persName><forename type="first">Carlos</forename><surname>Guestrin</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2016">2016</date>
			<biblScope unit="page" from="1135" to="1144" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">A unified approach to interpreting model predictions</title>
		<author>
			<persName><forename type="first">M</forename><surname>Scott</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Su-In</forename><surname>Lundberg</surname></persName>
		</author>
		<author>
			<persName><surname>Lee</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Advances in neural information processing systems</title>
				<imprint>
			<date type="published" when="2017">2017</date>
			<biblScope unit="page" from="4765" to="4774" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Interpretability in healthcare a comparative study of local machine learning interpretability techniques</title>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">El</forename><surname>Shawi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Sherif</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Al-Mallah</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Sakr</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IEEE 32nd International Symposium on Computer-Based Medical Systems (CBMS)</title>
				<imprint>
			<date type="published" when="2019-06">2019. June 2019</date>
			<biblScope unit="page" from="275" to="280" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Transfer learning with partial observability applied to cervical cancer screening</title>
		<author>
			<persName><forename type="first">Kelwin</forename><surname>Fernandes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jaime</forename><forename type="middle">S</forename><surname>Cardoso</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jessica</forename><surname>Fernandes</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Iberian conference on pattern recognition and image analysis</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2017">2017</date>
			<biblScope unit="page" from="243" to="250" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">A comparative analysis of classification techniques for cervical cancer utilising at risk factors and screening test results</title>
		<author>
			<persName><forename type="first">Sean</forename><surname>Quinlan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Haithem</forename><surname>Afli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ruairi O'</forename><surname>Reilly</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">AICS</title>
		<imprint>
			<biblScope unit="page" from="400" to="411" />
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Imbalancedlearn: A python toolbox to tackle the curse of imbalanced datasets in machine learning</title>
		<author>
			<persName><forename type="first">Guillaume</forename><surname>Lemaître</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Fernando</forename><surname>Nogueira</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Christos</forename><forename type="middle">K</forename><surname>Aridas</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">The Journal of Machine Learning Research</title>
		<imprint>
			<biblScope unit="volume">18</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="559" to="563" />
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Induction of decision trees</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">Ross</forename><surname>Quinlan</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Machine learning</title>
		<imprint>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="81" to="106" />
			<date type="published" when="1986">1986</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">Cytology in the diagnosis of cervical cancer in symptomatic young women: a retrospective review</title>
		<author>
			<persName><forename type="first">Anita</forename><forename type="middle">Ww</forename><surname>Lim</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Rebecca</forename><surname>Landy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Alejandra</forename><surname>Castanon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Antony</forename><surname>Hollingworth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Willie</forename><surname>Hamilton</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Nick</forename><surname>Dudding</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Peter</forename><surname>Sasieni</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">the journal of the Royal College of General Practitioners</title>
		<imprint>
			<biblScope unit="volume">66</biblScope>
			<biblScope unit="issue">653</biblScope>
			<biblScope unit="page" from="e871" to="e879" />
			<date type="published" when="2016-12">dec 2016</date>
		</imprint>
	</monogr>
	<note>The British journal of general practice</note>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
