<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Shapley values and fairness ⋆</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Paolo</forename><surname>Giudici</surname></persName>
							<email>paolo.giudici@unipv.it</email>
							<affiliation key="aff0">
								<orgName type="institution">University of Pavia</orgName>
								<address>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Parvati</forename><surname>Neelakantan</surname></persName>
							<email>parvati@iitk.ac.in</email>
							<affiliation key="aff1">
								<orgName type="institution">Indian Institute of Technology</orgName>
								<address>
									<settlement>Kanpur</settlement>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Shapley values and fairness ⋆</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">C80EEE0AF6589F73E68B5ED12FD99804</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T19:37+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Explainable artificial intelligence</term>
					<term>Machine Learning</term>
					<term>Fairness</term>
					<term>Shapley values</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Fairness is a key requirement for artificial intelligence applications. The assessment of fairness is typically based on the marginal relationship between the response variable and a protected variable, such as gender, age, or race: the stronger the relationship, the lower the fairness. In the paper, we show that this type of reasoning is insufficient, and may lead to Simpson's paradox, for which a fair model may become unfair when conditioning on a control variable. We thus propose a novel method to assess fairness, based on conditional explainability.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Credit lending decisions may be unfair. Even when automatic decisions, based on machine learning, are taken, algorithmic injustice may arise. Indeed, the measurement of the fairness of the machine learning models employed for credit lending and, more generally, for similar binary classification models, has attracted much interest in the recent academic literature (see e.g. [1, 2, 3, 4, 5, 6]. We follow this stream of research and exploit explainability to achieve fairness in credit lending models and, specifically, in loan acceptance decisions. More precisely, we consider the difference between the Shapley values attributed to different predictors, conditioning on the different labels of a control variable. We apply our proposal to a real-world database containing 157,269 personal lending decisions and show that both logistic regression and random forest models are unfair. However, while the former remains unfair also at the marginal level, the latter does not so, providing an instance of Simpson's paradox. In this paper, we present a methodology to detect indirect bias, and the possible existence of Simpson's paradox, when machine learning models are considered for binary decisions, such as loan acceptance. This is indeed our main contribution to the extant literature: there are yet no papers that propose how to detect indirect bias, using explainable AI methods, from the output of machine learning models. Our contribution is to propose models that are fair and explainable not only marginally, with respect to a protected variable, but also conditionally, in different groups of a control variable. In the context of loan acceptance decisions, the motivation of our proposal is that fairness in loan acceptance should be evaluated not only by looking at whether the probability of accepting a loan depends on race (direct bias) but also by looking at whether the same probability changes conditionally on a control variable that is correlated with race (indirect bias). The statistical literature regards this phenomenon as an instance of Simpson's paradox: two variables may not be marginally dependent (no direct bias) but may be so when controlling for a third variable (indirect bias). In this paper, we present a methodology to detect indirect bias, and the possible existence of Simpson's paradox, when machine learning models are considered for binary decisions, such as loan acceptance. We present the data that motivate our approach in Section 2. Whereas in Section 3 we present our proposal, and in Section 4 we illustrate its application in the context of the real-world data presented in Section 2. Finally, Section 5 ends with some concluding remarks.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Data and Variables</head><p>We examine 157,269 loan applications from Home Mortgage Disclosure Act's (HDMA) website made in New York during 2017. The dependent target variable, Declined Loan, takes the value 1 if a loan application initially satisfies the approval requirements of Government Sponsored Enterprises or of Federal Housing Administrations (GSEs/FHA), though it subsequently fails in meeting the lenders' requirements; it takes the value 0 if the lender approves the loan. In detailing GSEs/FHA's initial acceptance of the borrower's application and its subsequent rejection by the bank, the HMDA dataset, which includes information on the applicant's race, eminently qualifies for our study. Our key independent variable of interest (the protected variable) is the information on the applicant's race. It takes the value 1 if the applicant is African American; and 0 if it is White American. Further independent variables are used as controls, including gender<ref type="foot" target="#foot_0">1</ref> , income<ref type="foot" target="#foot_1">2</ref> , amount of loan <ref type="bibr" target="#b2">3</ref> , purpose of loan <ref type="bibr" target="#b3">4</ref> , lien status<ref type="foot" target="#foot_4">5</ref> , and type of loan <ref type="bibr" target="#b5">6</ref> . Table <ref type="table" target="#tab_0">1</ref> reports the main summary statistics of the considered variables. From Table <ref type="table" target="#tab_0">1</ref> note that the dependent variable suffers from class imbalance. In other words, the number of observations that belong to the positive class (loan declined) is significantly lesser than those that belong to the negative class (loan approved). <ref type="bibr" target="#b6">7</ref> Models trained on such data, which prioritize the prevalent class over the minority class, may lead to an overly optimistic measure of accuracy [7]. While such models can predict loan approvals with a high level of accuracy, they often fail to accurately predict declined loans. To solve the problem of class imbalance, in this paper, we employ an under-sampling technique to meaningfully infer information from the data. This technique randomly discards observations from the majority class to balance the skewed distribution. <ref type="bibr" target="#b7">8</ref> We also examine the marginal correlations of the dependent variable and of the race explanatory variable with all others, to obtain preliminary insights on the explanatory power of the model. Tables <ref type="table" target="#tab_2">2 and 3</ref> report the correlations of the dependent variable "Declined loan" and the correlations of the "Applicant's race" variable with the remaining variables, respectively, and the corresponding t-statistics and p-values. From Table <ref type="table" target="#tab_1">2</ref>, note that most variables are correlated with the dependent variables and, among them, the applicant's race, regardless of its high imbalance (there are only about 8% African Americans in the sample). On the other hand, Table <ref type="table" target="#tab_2">3</ref> shows that Applicant's race is highly correlated with many other independent variables and, in particular, it is highly and positively correlated with the "Loan Amount". The combined effect of the dependencies shown in Table <ref type="table" target="#tab_1">2 and Table</ref>  <ref type="table" target="#tab_2">3</ref> suggests a possible case for the arising of Simpson's paradox. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Methodology</head><p>Derived from coalitional game theory, Shapley Values assume, for each instance of a prediction, that each feature value is a "player" in a game with the prediction as the payout (see, for example, [8]. Shapley value quantifies a feature's contribution in predicting a given instance's response value. Theoretically, Shapley values are the average marginal contribution of a feature value across all possible coalitions of features. They represent a "fair" distribution of the credit related to the difference between the specific prediction and the average prediction. This renders them a model-agnostic XAI method which can shed light on the models' internal logic. We follow [9] to compute the GSV XAI method. We first compute the Shapley values, using the [10] approximation method, for each feature at each instance. Owing to the Monotonicity property of Shapley Values, the local Shapley formulation can be easily extended to provide insights into the models' global behaviour. We, therefore, aggregate a feature's contributions across all instances to arrive at a measure that is interpreted as a measure of feature importance influencing the model behavior. A coalitional game is defined as a tuple &lt; 𝑁, 𝜈 &gt;, where 𝑁 = {1, 2, . . . , 𝑛} is a finite set of players and 𝜈 a characteristic function that assigns value to each subset of 𝑁 . [11] proved that unique values assigned to individual players could be estimated using the equation:</p><formula xml:id="formula_0">𝑆ℎ𝑎𝑝𝑙𝑒𝑦 𝑣𝑎𝑙𝑢𝑒 𝑖 (𝜈) = ∑︁ 𝑆⊆𝑁 ∖{𝑖},𝑠=|𝑆| (𝑛 − 𝑠 − 1)!𝑠! 𝑛! (𝜈(𝑆 ∪ {𝑖}) − 𝜈(𝑆))<label>(1)</label></formula><p>Thus, the marginal contribution of a predictor 𝑋 𝑘 , its Shapley value 𝜑, can be expressed as,</p><formula xml:id="formula_1">𝜑(𝑓 ˆ(𝑋 𝑖 )) = ∑︁ 𝑋 ′ ⊆𝐶(𝑋)∖𝑋 𝑘 |𝑋 ′ |!(𝐾 − |𝑋 ′ | − 1)! 𝐾! [𝑓 ˆ(𝑋 ′ ∪ 𝑋 𝑘 ) 𝑖 − 𝑓 ˆ(𝑋 ′ ) 𝑖 ]<label>(2)</label></formula><p>In this notation, 𝐶(𝑋) ∖ 𝑋 𝑘 is the set of all model configurations excluding variable 𝑋 𝑘 and 𝑓 ˆthe trained model. We aggregate the Shapley values of a feature across all the instances to arrive at the GSV measure. GSV is the average marginal contribution of a feature value across all possible coalitions of features. We present a methodology to detect indirect bias, and the possible existence of Simpson's paradox, when machine learning models are considered for binary decisions, such as loan acceptance. <ref type="bibr" target="#b8">9</ref> Indirect bias arises when conditioning on a control variable. For example, we can assess whether the distribution of GSV values differ when we condition on loans of high amount rather than of a low amount. To verify whether the difference in the GSV distributions is statistically significant, we can calculate Spearman's correlation coefficient between the importance ranks attributed to the variables in the two samples. Operationally, considering (without loss of generality) loan amount as a conditioning control variable, we can divide the dataset into low and high loan amount values and compute the GSV values on these subsamples. Specifically, a Loan amount less (greater) than the median amount is considered a low (high) loan amount. The choice of the loan amount as a control variable is justified by its high positive correlation with the race of the loan applicant. To verify whether the two distributions differ significantly we can set up a hypotheses test for Spearman's correlation, as follows: 𝐻 0 : The GSV values in the two loan amount subsamples are not correlated. 𝐻 1 : The GSV values in the two loan amount subsamples are correlated. When the null hypothesis is not rejected, the two distributions can be considered unrelated ("different"). Instead, when the null hypothesis is rejected, the two distributions assign similar ranks and can be considered dependent ("similar").</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Empirical findings</head><p>We first train a transparent logistic regression (LR) model on 70% of the data and apply GSV method to examine whether there is an incidence of racial discrimination. We then divide the dataset with respect to loan amount and train LR models on samples comprising high and low loan amounts, respectively. We then apply the GSV method on these trained models to examine whether Simpson's paradox arises. The table below reports the GSV results for the model comprising all the explanatory variables and trained on 70% of the dataset. Table <ref type="table" target="#tab_3">4</ref> shows that applicant race is ranked 1 out of all the variables and it explains 34% of the declined loans, hence showing evidence of racial discrimination. These results are consistent with the findings of [9]. The results for the LR model trained on the high loan amount dataset are reported in Table <ref type="table" target="#tab_4">5</ref>. From Table <ref type="table" target="#tab_4">5</ref>, we find that the GSV technique for the high loan amount sample confirms the importance of race in lending decisions. More precisely, the results indicate that applicant race is ranked 1 out of our 6 variables and that it explains 43% of the declined loans. In contrast, Loan purpose is ranked only 4 (it was 2 in the aggregated sample).The results for the LR model trained on the low loan amount dataset are reported in Table <ref type="table" target="#tab_5">6</ref>. From Table <ref type="table" target="#tab_5">6</ref>, we find that the GSV method ranks the applicant race only 2 of our 6 variables. In addition, the race variable explains only 12% of declined loans. Instead, the variable Loan purpose is ranked 1 and explains 59% of declined loans. Therefore, the results suggest a weak incidence of discrimination in the low loan amount sample, differently from what occurs for the high loan amount sample. We conclude that for the LR model, Simpson's paradox does not arise. The racial bias found when we consider high loan amounts remains true when we combine the two groups. For robustness, we conduct the Spearman test to verify whether the two populations are correlated. It turns out that Spearman's rank correlation is equal to −0.08, leading to a very large p-value, indicating that the null hypotheses cannot be rejected: there is a significant difference in the variable importance ranks of the two groups.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.">Random Forest</head><p>We now consider a black box random forest (RF) model on 70% of the data and apply the GSV method to examine whether there is an incidence of racial discrimination. We then divide the dataset with respect to loan amount and train RF models separately on the two samples, comprising high and low loan amounts, respectively. <ref type="bibr" target="#b9">10</ref> We then apply the GSV method on these trained RF models to examine whether a racial bias holds conditionally. The table below reports the GSV results for the RF model comprising all the explanatory variables and trained on 70% of the dataset. Table <ref type="table" target="#tab_6">7</ref> shows that applicant race is ranked 7 out of all the variables and it explains only 3% of the declined loans. Thus, using the GSV approach we obtain that the RF model gives low importance to the applicant's race. Hence, the RF model appears (marginally) as an ethically accountable model, differently from the LR model, a result consistent with the findings in [9]. We then fit a random forest model on the low loan amount sample and compute the GSV importances. The results are reported in Table <ref type="table" target="#tab_7">8</ref>. From Table <ref type="table" target="#tab_7">8</ref>, we do not find evidence of bias when we fit the RF model on the low loan amount sample. The XAI method ranks the applicant race 5 of our 6 variables and the race variable only explains 1% of the declined loans. These results imply that lenders may not discriminate against applicants based on their race if the amount of loan requested is low. This result is consistent with that obtained on low amount loans using the LR model. We then train the RF model on the high loan amount sample. The results are reported in Table <ref type="table" target="#tab_8">9</ref>. Table <ref type="table" target="#tab_8">9</ref> shows evidence of discrimination. Applicant race is ranked 1 and it explains 91% of the declined loans. This result contradicts the result obtained on the aggregated sample in Table <ref type="table" target="#tab_7">8</ref> and indicates that Simpson's paradox arises for RF models. There is no racial bias when we look at the aggregate sample, but the bias emerges when we condition on high loan amounts. We conclude that, when we condition on loan amounts, both LR and RF have a bias. Instead, when we look for fairness at the aggregate level, the bias disappears for random forests, but it remains for logistic regression. The difference may be due to the stronger importance of race attributed by the LR model for low amounts: 0.12 versus only 0.01 for the RF model. We further test whether the difference in the low and high loan amount results is statistically significant, employing Spearman rank correlation. We find that the Spearman rank correlation between the variable ranks attributed by the RF model in the two samples is equal to 0.02. This leads to a large p-value, which indicates that the null hypothesis of no correlation between the ranks cannot be rejected. Therefore, the difference between the two groups is not statistically significant, as was the case for the LR model.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Conclusions</head><p>We propose a novel method to assess fairness, based on conditional Shapley values. We apply our proposed model to the credit lending decisions contained in the well-known HDMA data repository of loan applications made in 2017 in New York State. Our results indicate that, using both LR and RF models, "High loan amount" loans receive a biased treatment in terms of race. However, when the group is aggregated with "Low amount" loans, the two models show different behaviour. While LR remains biased, RF becomes fair. This provides an instance of Simpson's paradox and indicates that, in assessing fairness, a conditional approach should be followed, rather than a marginal one, to avoid misleading conclusions. Our conditional approach reveals that both LR and RF models applied to our data, are biased. This is consistent with the fact that both learn from a vast amount of training data, in which decisions are made by human beings who are known to be unfair in credit lending decisions.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1</head><label>1</label><figDesc>Descriptive Statistics</figDesc><table><row><cell>Variable</cell><cell cols="3">Mean Max Std.Dev</cell></row><row><cell>Declined Loan</cell><cell>0.06</cell><cell>1</cell><cell>0.24</cell></row><row><cell>Applicant race</cell><cell>0.08</cell><cell>1</cell><cell>0.27</cell></row><row><cell>Applicant gender</cell><cell>0.65</cell><cell>1</cell><cell>0.48</cell></row><row><cell>Applicant income</cell><cell>4.6</cell><cell>11</cell><cell>0.74</cell></row><row><cell>Loan amount</cell><cell>5.3</cell><cell>9.9</cell><cell>0.88</cell></row><row><cell>Loan purpose</cell><cell>0.33</cell><cell>1</cell><cell>0.47</cell></row><row><cell>Lien status</cell><cell>0.97</cell><cell>1</cell><cell>0.16</cell></row><row><cell>Loan type</cell><cell>0.18</cell><cell>1</cell><cell>0.38</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2</head><label>2</label><figDesc>Correlation between Declined Loan and independent variables</figDesc><table><row><cell>Variable</cell><cell cols="2">Corr. t-statistic</cell><cell>p-value</cell></row><row><cell>Applicant race</cell><cell>0.04</cell><cell cols="2">17.43 &lt; 2.2e-16</cell></row><row><cell>Applicant gender</cell><cell>0.00</cell><cell>1.51</cell><cell>0.1323</cell></row><row><cell>Applicant income</cell><cell>0.01</cell><cell cols="2">3.93 1.069e-08</cell></row><row><cell>Loan amount</cell><cell>0.01</cell><cell>1.61</cell><cell>0.1082</cell></row><row><cell>Loan purpose</cell><cell>0.12</cell><cell cols="2">46.57 &lt; 2.2e-16</cell></row><row><cell>Lien status</cell><cell>0.01</cell><cell cols="2">4.63 3.671e-06</cell></row><row><cell>Loan type</cell><cell>0.03</cell><cell cols="2">13.25 &lt; 2.2e-16</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 3</head><label>3</label><figDesc>Correlation between Applicant race and other variables</figDesc><table><row><cell>Variable</cell><cell cols="2">Corr. t-statistic</cell><cell>p-value</cell></row><row><cell>Declined Loan</cell><cell>0.04</cell><cell cols="2">17.43 &lt; 2.2e-16</cell></row><row><cell>Applicant gender</cell><cell>-0.10</cell><cell cols="2">-37.85 &lt; 2.2e-16</cell></row><row><cell>Applicant income</cell><cell>-0.05</cell><cell cols="2">-20.70 &lt; 2.2e-16</cell></row><row><cell>Loan amount</cell><cell>0.05</cell><cell cols="2">20.11 &lt; 2.2e-16</cell></row><row><cell>Loan purpose</cell><cell>0.02</cell><cell cols="2">8.37 &lt; 2.2e-16</cell></row><row><cell>Lien status</cell><cell>-0.02</cell><cell cols="2">-5.96 2.518e-09</cell></row><row><cell>Loan type</cell><cell>0.16</cell><cell cols="2">62.52 &lt; 2.2e-16</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 4</head><label>4</label><figDesc>Explainable AI estimates in a transparent model for Declined Loans in the whole population</figDesc><table><row><cell>Variable</cell><cell cols="2">Global Shapley Rank</cell></row><row><cell>Applicant race</cell><cell>0.343770719</cell><cell>1</cell></row><row><cell>Applicant gender</cell><cell>0.048933083</cell><cell>6</cell></row><row><cell>Applicant income</cell><cell>0.095036921</cell><cell>4</cell></row><row><cell>Loan amount</cell><cell>0.063443748</cell><cell>5</cell></row><row><cell>Loan purpose</cell><cell>0.331693046</cell><cell>2</cell></row><row><cell>Lien status</cell><cell>0.116902396</cell><cell>3</cell></row><row><cell>Loan type</cell><cell>0.000220089</cell><cell>7</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Table 5</head><label>5</label><figDesc>Explainable AI estimates in a transparent model for Declined Loans in the high loan amount population</figDesc><table><row><cell>Variable</cell><cell cols="2">Global Shapley Rank</cell></row><row><cell>Applicant race</cell><cell>0.434362577</cell><cell>1</cell></row><row><cell>Applicant gender</cell><cell>0.125260673</cell><cell>3</cell></row><row><cell>Applicant income</cell><cell>0.072034107</cell><cell>5</cell></row><row><cell>Loan purpose</cell><cell>0.110704801</cell><cell>4</cell></row><row><cell>Lien status</cell><cell>0.004053909</cell><cell>6</cell></row><row><cell>Loan type</cell><cell>0.253583933</cell><cell>2</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_5"><head>Table 6</head><label>6</label><figDesc>Explainable AI estimates in a transparent model for Declined Loans -low loan amount population</figDesc><table><row><cell>Variable</cell><cell cols="2">Global Shapley Rank</cell></row><row><cell>Applicant race</cell><cell>0.128889574</cell><cell>2</cell></row><row><cell>Applicant gender</cell><cell>0.063403784</cell><cell>5</cell></row><row><cell>Applicant income</cell><cell>0.101432872</cell><cell>3</cell></row><row><cell>Loan purpose</cell><cell>0.598468759</cell><cell>1</cell></row><row><cell>Lien status</cell><cell>0.067086229</cell><cell>4</cell></row><row><cell>Loan type</cell><cell>0.040718781</cell><cell>6</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_6"><head>Table 7</head><label>7</label><figDesc>Explainable AI estimates in a Random Forest model for Declined Loans -whole population</figDesc><table><row><cell>Variable</cell><cell cols="2">Global Shapley Rank</cell></row><row><cell>Applicant race</cell><cell>0.032875723</cell><cell>7</cell></row><row><cell>Applicant gender</cell><cell>0.24566474</cell><cell>2</cell></row><row><cell>Applicant income</cell><cell>0.036333609</cell><cell>6</cell></row><row><cell>Loan amount</cell><cell>0.239419901</cell><cell>3</cell></row><row><cell>Loan purpose</cell><cell>0.309609827</cell><cell>1</cell></row><row><cell>Lien status</cell><cell>0.072667217</cell><cell>4</cell></row><row><cell>Loan type</cell><cell>0.063428984</cell><cell>5</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_7"><head>Table 8</head><label>8</label><figDesc>Explainable AI estimates in a Random Forest model for Declined Loans -low loan amount population</figDesc><table><row><cell>Variable</cell><cell cols="2">Global Shapley Rank</cell></row><row><cell>Applicant race</cell><cell>0.012392485</cell><cell>5</cell></row><row><cell>Applicant gender</cell><cell>0.002716161</cell><cell>6</cell></row><row><cell>Applicant income</cell><cell>0.071808511</cell><cell>2</cell></row><row><cell>Loan purpose</cell><cell>0.838784518</cell><cell>1</cell></row><row><cell>Lien status</cell><cell>0.059755545</cell><cell>3</cell></row><row><cell>Loan type</cell><cell>0.014542780</cell><cell>4</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_8"><head>Table 9</head><label>9</label><figDesc>Explainable AI estimates in a Random Forest model for Declined Loans -high loan amount population</figDesc><table><row><cell>Variable</cell><cell cols="2">Global Shapley Rank</cell></row><row><cell>Applicant race</cell><cell>0.910513384</cell><cell>1</cell></row><row><cell>Applicant gender</cell><cell>0.021595758</cell><cell>4</cell></row><row><cell>Applicant income</cell><cell>0.036790379</cell><cell>2</cell></row><row><cell>Loan purpose</cell><cell>0.023276865</cell><cell>3</cell></row><row><cell>Lien status</cell><cell>0.002392344</cell><cell>6</cell></row><row><cell>Loan type</cell><cell>0.005431269</cell><cell>5</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">The variable takes 1 if the applicant is male and 0 if female.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_1">Natural logarithm of the applicant's gross annual income the lender relies on when making the credit decision.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_2">Natural logarithm of the amount of the covered loan, or the amount applied for.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_3">The variable takes 1 if the purpose of seeking a loan was for refinancing the mortgage and 0 if purchasing a home.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_4">This variable takes the value 1 if the loan application is secured by 'first lien' and 0 for a subordinate lien. A first lien level of security indicates that the lender is the first to be paid when a borrower defaults and the property or asset is used as collateral for the debt.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="6" xml:id="foot_5">The variable takes the value 1 if the loan was insured by the FHA and if insured by a GSE.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="7" xml:id="foot_6">Notes. The Table presents the summary statistics. The observations are 157269 and the min for all the variables is 0.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="8" xml:id="foot_7">In reducing the majority class's size to match the minority class, however, under-sampling may forgo potentially useful information from the majority class.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="9" xml:id="foot_8"> Simpson's Paradox<ref type="bibr" target="#b11">[12]</ref> is a phenomenon in which a statistical dependence appears in different groups of data but disappears or reverses when these groups are combined by marginalisation.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="10" xml:id="foot_9">The data partition is exactly the same as that employed for the LR model.</note>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Fairness in credit scoring: Assessment, implementation and profit implications</title>
		<author>
			<persName><forename type="first">N</forename><surname>Kozodoi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Jacob</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Lessmann</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">European Journal of Operational Research</title>
		<imprint>
			<biblScope unit="volume">297</biblScope>
			<biblScope unit="page" from="1083" to="1094" />
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">On learning fairness and accuracy on multiple subgroups</title>
		<author>
			<persName><forename type="first">C</forename><surname>Shui</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">X</forename><surname>Ling</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Arbel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Gagné</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Advances in Neural Information Processing Systems</title>
		<imprint>
			<biblScope unit="volume">35</biblScope>
			<biblScope unit="page" from="34121" to="34135" />
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<monogr>
		<author>
			<persName><forename type="first">C</forename><surname>Hurlin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Pérignon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Saurin</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2205.10200</idno>
		<title level="m">The fairness of credit scoring models</title>
				<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Marrying fairness and explainability in supervised learning</title>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">A</forename><surname>Grabowicz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Perello</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Mishra</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">2022 ACM Conference on Fairness, Accountability, and Transparency</title>
				<imprint>
			<date type="published" when="2022">2022</date>
			<biblScope unit="page" from="1905" to="1916" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Explainability and fairness in machine learning: Improve fair end-to-end lending for kiva</title>
		<author>
			<persName><forename type="first">A</forename><surname>Stevens</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Deruyck</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Van Veldhoven</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Vanthienen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IEEE Symposium Series on Computational Intelligence (SSCI)</title>
				<imprint>
			<publisher>IEEE</publisher>
			<date type="published" when="2020">2020. 2020</date>
			<biblScope unit="page" from="1241" to="1248" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Safe artificial intelligence in finance</title>
		<author>
			<persName><forename type="first">P</forename><surname>Giudici</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Raffinetti</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Finance research Letters</title>
		<imprint>
			<biblScope unit="volume">56</biblScope>
			<biblScope unit="page">104088</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">A study of the behavior of several methods for balancing machine learning training data</title>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">E</forename><surname>Batista</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">C</forename><surname>Prati</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">C</forename><surname>Monard</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">ACM SIGKDD explorations newsletter</title>
		<imprint>
			<biblScope unit="volume">6</biblScope>
			<biblScope unit="page" from="20" to="29" />
			<date type="published" when="2004">2004</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">A unified approach to interpreting model predictions</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">M</forename><surname>Lundberg</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S.-I</forename><surname>Lee</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Advances in neural information processing systems</title>
		<imprint>
			<biblScope unit="volume">30</biblScope>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Countering racial discrimination in algorithmic lending: A case for model-agnostic interpretation methods</title>
		<author>
			<persName><forename type="first">S</forename><surname>Agarwal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Cal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Neelakantan</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Economics Letters</title>
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Explaining prediction models and individual predictions with feature contributions</title>
		<author>
			<persName><forename type="first">E</forename><surname>Štrumbelj</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Kononenko</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Knowledge and information systems</title>
		<imprint>
			<biblScope unit="volume">41</biblScope>
			<biblScope unit="page" from="647" to="665" />
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<monogr>
		<author>
			<persName><forename type="first">L</forename><surname>Shapley</surname></persName>
		</author>
		<title level="m">Contributions to the Theory of Games: Volume II. A value for n-person games</title>
				<imprint>
			<publisher>Princeton University Press</publisher>
			<date type="published" when="1953">1953</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">The interpretation of interaction in contingency tables</title>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">H</forename><surname>Simpson</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of the Royal Statistical Society: Series B (Methodological)</title>
		<imprint>
			<biblScope unit="volume">13</biblScope>
			<biblScope unit="page" from="238" to="241" />
			<date type="published" when="1951">1951</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
