<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Towards Accountability Driven Development for Machine Learning Systems</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Chiu</forename><surname>Pang Fung</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">School of Mathematical and Computer Sciences</orgName>
								<orgName type="institution">Heriot-Watt University Edinburgh</orgName>
								<address>
									<postCode>EH14 4AS</postCode>
									<country key="GB">UK</country>
								</address>
							</affiliation>
							<affiliation key="aff2">
								<orgName type="department">School of Computing</orgName>
								<orgName type="institution">University of Leeds</orgName>
								<address>
									<postCode>LS2 9JT</postCode>
									<settlement>Leeds</settlement>
									<country key="GB">UK</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Wei</forename><surname>Pang</surname></persName>
							<email>w.pang@hw.ac.uk</email>
							<affiliation key="aff1">
								<orgName type="department">School of Natural and Computing Sciences</orgName>
								<orgName type="institution">University of Aberdeen Aberdeen</orgName>
								<address>
									<postCode>AB24 3UE</postCode>
									<country key="GB">UK</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Iman</forename><surname>Naja</surname></persName>
							<email>iman.naja@abdn.ac.uk</email>
							<affiliation key="aff0">
								<orgName type="department">School of Mathematical and Computer Sciences</orgName>
								<orgName type="institution">Heriot-Watt University Edinburgh</orgName>
								<address>
									<postCode>EH14 4AS</postCode>
									<country key="GB">UK</country>
								</address>
							</affiliation>
							<affiliation key="aff1">
								<orgName type="department">School of Natural and Computing Sciences</orgName>
								<orgName type="institution">University of Aberdeen Aberdeen</orgName>
								<address>
									<postCode>AB24 3UE</postCode>
									<country key="GB">UK</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Milan</forename><surname>Markovic</surname></persName>
							<email>milan.markovic@abdn.ac.uk</email>
							<affiliation key="aff1">
								<orgName type="department">School of Natural and Computing Sciences</orgName>
								<orgName type="institution">University of Aberdeen Aberdeen</orgName>
								<address>
									<postCode>AB24 3UE</postCode>
									<country key="GB">UK</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Peter</forename><surname>Edwards</surname></persName>
							<email>p.edwards@abdn.ac.uk</email>
							<affiliation key="aff1">
								<orgName type="department">School of Natural and Computing Sciences</orgName>
								<orgName type="institution">University of Aberdeen Aberdeen</orgName>
								<address>
									<postCode>AB24 3UE</postCode>
									<country key="GB">UK</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Towards Accountability Driven Development for Machine Learning Systems</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">5F8072267D1F7E745F54014373E4E6EF</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T04:46+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Behaviour Driven Development</term>
					<term>Machine Learning</term>
					<term>Model Card</term>
					<term>Accountability</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>With rapid deployment of Machine Learning (ML) systems into diverse domains such as healthcare and autonomous driving, important questions regarding accountability in case of incidents resulting from ML errors remain largely unsolved. To improve accountability of ML systems, we introduce a framework called Accountability Driven Development (ADD). Our framework reuses Behaviour Driven Development (BDD) approach to describe testing scenarios and system behaviours in ML Systems' development using natural language, guides and forces developers and intended users to actively record necessary accountability information in the design and implementation stages. In this paper, we illustrate how to transform accountability requirements to specific scenarios and provide syntax to describe them. The use of natural language allows non technical collaborators such as stakeholders and non ML domain experts deeply engaged in ML system development to provide more comprehensive evidence to support system's accountability. This framework also attributes the responsibility to the whole project team including the intended users rather than putting all the accountability burden on ML engineers only. Moreover, this framework can be considered as a combination of both system test and acceptance test, thus making the development more efficient. We hope this work can attract more engineers to use our idea, which enables them to create more accountable ML systems.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>There are many definitions of accountability in AI systems, and following our previous work <ref type="bibr" target="#b9">[10]</ref>, we define accountability as follows: "The ability to inspect, review or otherwise interrogate an AI system with the goal of (i ) making processes associated with each of its life cycle stages transparent; (ii ) demonstrating compliance with hard laws (i.e. laws and regulations), and soft laws (i.e. standards and guidelines); and (iii ) aiding investigations into the cause(s) of failure or erroneous decisions and supporting the identification of responsible parties." <ref type="bibr" target="#b9">[10]</ref>. This definition can be adopted to the narrower scope -ML systems. Currently ML system development is a straight forward process <ref type="bibr" target="#b16">[17,</ref><ref type="bibr" target="#b3">4]</ref> (Figure <ref type="figure" target="#fig_0">1</ref>). In this kind of processes, the relevant personnel within the project are in charge of different tasks based on their roles. Decision makers care about the purpose; stakeholders are concerned with the business value of the potential ML application; domain experts explain what the data means; data scientists and ML engineers focus on technical perspective, including data processing, model training and evaluating. This division of labor may cause ambiguity for accountability of ML systems. Many data scientists and ML engineers are keen on fulfilling the system performance requirements rather than considering the accountability issues. The Model Card framework <ref type="bibr" target="#b8">[9]</ref> offers a standardized documentation procedure to capture useful information such as model purpose, owners, data information, algorithms' details. But the information they collected is not comprehensive regarding to accountability of ML systems. The lack of information about system's explainability, transparency, fairness and performance makes it difficult to audit and investigate the potential failures of ML systems. ML system's life cycle can be divided into four high level stages: Design, Implementation, Deployment, and Operation &amp; Maintenance <ref type="bibr" target="#b9">[10]</ref>. In this paper, we focus on illustrating how to ask the ML system to generate those information in our ADD framework in the implementation stage, and we also give examples based on a simulation using our framework to develop an auto decision making system for mortgage applications.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Background and related work</head><p>Artificial Intelligence (AI) systems which employ ML models are being put to use in various applications affecting our daily lives. However, few companies and organizations are dedicating resources to mitigate AI risks despite some of them having to increasingly manage such risks related to AI <ref type="bibr" target="#b7">[8]</ref>. In AI/ML research, the same situation happens that many researchers focus on improving the performance of ML algorithms rather than looking into the accountability of these systems. Research about ML accountability has been emerging in recent years. Hajian et al. <ref type="bibr" target="#b4">[5]</ref> propose an idea on how to process raw data to prevent discrimination in AI based intrusion and crime detection. Datta et al. <ref type="bibr" target="#b2">[3]</ref> introduce the Quantitative Input Influence (QII) metric to improve the transparency of   <ref type="bibr" target="#b13">[14]</ref> Deep Learning Important FeaTures (DeepLIFT). Finallu, Lundberg et al. <ref type="bibr" target="#b6">[7]</ref> propose the SHapley Additive exPlanations (SHAP) method. These articles started the upsurge of ML's explainability research. All the above researches make it feasible for the information of ML system's explainability to be collected. For the other aspect, transparency, Mitchell et al. introduce Model Card <ref type="bibr" target="#b8">[9]</ref> to create documentations for ML models. Their framework allows people to look deeply into the model's development and its performance.</p><p>On the other side, in software development there is a methodology called Behavior Driven Development (BDD) <ref type="bibr" target="#b11">[12,</ref><ref type="bibr" target="#b5">6,</ref><ref type="bibr" target="#b15">16]</ref>, which is developed from Test Driven Development (TDD) <ref type="bibr" target="#b1">[2,</ref><ref type="bibr" target="#b10">11]</ref>. Same as TDD, BDD is still a test driven development approach, but BDD designs user story scenarios instead of designing test cases; it uses natural language to describe software system's behaviours, forcing the behaviours fulfill the corresponding requirements in all scenarios to test the system. As a test-first approach, in BDD the scenarios are designed at the beginning of the development when there is no existing code, so that the first test must fail until developers start to write code. The use of natural languages to describe user story scenarios in BDD creates a bridge among software designers, developers and users, encouraging all the development team members such as developers and non-technical collaborators to work together with less technical barriers. Inspired by BDD, we use natural language to describe the ML system's behaviours in the form of user story scenarios that are designed according to both technical and accountability requirements, making the ML system fulfill those requirements. That means the ML system must generate information about its performance, explainability, transparency etc. For example, there is a performance requirement-Accuracy, when the ML system passes the corresponding scenario test, it must generate information on accuracy. Those information can be collected to improve system's accountability. If there is a failure about the system's accuracy after its deployment, we can trace back what kind of accuracy test was done and what condition had be set in the corresponding testing scenario, and this information will help the failure investigation and auditing purpose. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">ML System Development with ADD</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">The workflow of ADD</head><p>In our framework (Figure <ref type="figure" target="#fig_2">2</ref>), starting from the Task/Requirements analysis, the development team and potential users should not only discuss technical requirements (e.g. performance requirements such as accuracy and efficiency), but also discuss the accountability requirements according to model's transparency, explainability, fairness and document all the requirements. Then based on the requirements, user stories and test scenarios are designed to start the testing. Similar to BDD, the first test must fail as no ML system has been developed. Afterwards, the developers start to implement the ML system and test it again, until the developed system passes all the tests. Eventually the system can be considered to have fulfilled all the requirements and ready to be deployed. In this case, the testing in our framework can be considered combining system test and acceptance test.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">Technical and Accountability Requirements</head><p>Technical requirements normally refer to performance characteristics. Some of these requirements originate from potential users, such as performance requirements, safety requirements, hardware architecture requirements. However, to improve the entire technical requirements, the development team should proactively propose some extra technical requirements in ML Systems, such as metrics for specific algorithms that do not violate customer's requirements. For example, a customer asks for a 90% accuracy for a classifier, and the development team should offer a confusion matrix analysis. According to transparency, a document which is extended from Model Card <ref type="bibr" target="#b8">[9]</ref> to collect more information should be established to record details of the entire model development process such as requirement analysis and useful information generated when the system passes the tests. Corresponding to explainability, there should be requirements to explain how and why this output is generated by the system. For example, in the auto decision making system for mortgage applications, when an output is obtained, the weights of all the features (features in this paper refer to applicant's income, occupation, age etc.) of the corresponding applicant should be listed and recorded. Regarding fairness, it can be proposed in accordance with law that the system cannot be discriminatory based on ethnicity or gender etc.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3">Design user story scenarios</head><p>In BDD, the user story scenarios tests relate to some sub-modules of the whole software system. For example, in a supermarket management system, selling a product relates to the cash flow management module and stock management module, so that the scenarios descriptions are more comprehensive. In ADD, we consider three blocks (Data Collection, Data Analysis/Pre-processing and Model Construction. Figure <ref type="figure" target="#fig_2">2</ref>) inside the dotted rectangle as a component in development. That will simplify the design. For example, we can ask this component to provide data distribution information rather to ask the Data Analysis/Preprocessing block to do so. Not only the normal requirements, but also the edge use cases must be covered in the scenario design. For example, when there is an edge use case that an applicant with no features goes into the system, an error message "Wrong input given, can not create proper output" should be given rather a normal output as "Mortgage application is approved " or "Mortgage application is declined ". Also, all details of scenarios design must be recorded as they contain what situations have been considered for the use of the system which are useful for accountability.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.4">Syntax for describing scenarios and system's behaviour</head><p>The ML technical terms may be difficult to understand, and obscure statistical methods, evaluation standards and metrics keep non-technical personnel out of the loop. These non-technical collaborators may have better understanding of other accountability factors, such as laws and ethical requirements, which can help improving a system's accountability. Removing this technical barrier is necessary to gain contributions for accountability from both technical and non-technical personnel in the development stages. In ADD, a method similar to BDD is used, and we use natural language to describe user story scenarios and the behaviors of the system. We give two examples below to explain how to transfer one fairness requirement into scenarios and how to describe system's behaviour in that scenario. In the second example, the system is forced to generate explanation information. (It is pointed out that in these examples, we just want to demonstrate how ADD can be used, and we do not consider the profit requirement of the intended users, the banks. We also point out that the below examples are the starting point to engage all parties, and they will be further refined after discussions among related people, including users and ML developers/designers. The final version of the scenarios may be produced after several iterations until all parties reach a consensus. If a consensus cannot be reached, the information on disagreement also needs to be recorded. )</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Title: Producing non-discriminatory output</head><p>As a Bank. I want the system that produces fair outputs for different applicants regardless of their ethnicity. so that the System is not racially biased.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Scenario 1:</head><p>The System should produce fair outputs for applicants from different ethnic groups. Given that a mortgage application approved for applicant A, When another applicant B, with the same features but different ethnicity applies for the same mortgage amount, Then the mortgage application should be approved for applicant B.</p><p>Scenario 2: The system should produce fair outputs for applicants from different ethnic groups. Given that a mortgage application from applicant A is declined, When another applicant B with the same features but different ethnicity is applying for the same mortgage amount, Then mortgage application from applicant B should be declined.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Title: Proving explanations for outputs</head><p>As a Bank. I want the system that provides explanation for every mortgage application decision. so that we know why the output is produced and can explain it to our clients.</p><p>Scenario 1: The system should provide explanation for a successful application. Given an applicant with all necessary features, When mortgage application is approved, Then the system should create a report with all the weights corresponding to the features of the applicant.</p><p>Scenario 2: The system should provide explanation for a failed application. Given an applicant with all necessary features, When mortgage application is declined, Then the system should create a report with all the weights corresponding to the features of the applicant.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Discussion and Future work</head><p>There is an urgent need to improve ML system's accountability during its development stage. In this paper, we briefly introduce our ADD framework to develop ML systems and explain how the ADD can facilitate accountability information capture regarding the system's performance, explainability, fairness and transparency. We believe ADD can help reduce misunderstanding among users, ML system designers and developers by engaging them and confirming accountability requirements.</p><p>In the future, we are going to further improve this framework by making a deeper study in the requirements and scenarios design sections (Sections 3.2, 3.3), developing an example model and giving a full report by extending the Model Card framework <ref type="bibr" target="#b8">[9]</ref>. Further more, we point out that ADD is a generic methodology for facilitating accountability in ML, and therefore, we will extend our framework to the whole life cycle of ML systems, making it more widely used. Finally, we are considering developing a Cucumber-like <ref type="bibr" target="#b14">[15]</ref> (cucumber is a tool that supports BDD.) tool that is specifically for ML system development.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Fig. 1 .</head><label>1</label><figDesc>Fig. 1. A typical Machine Learning system Development Workflow.</figDesc><graphic coords="3,134.77,115.83,345.83,165.62" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head></head><label></label><figDesc>decision-making systems by measuring the influence of the input features. Bach et al. [1] propose Layer-Wise Relevance Propagation to explain deep neural networks. Ribeiro et al. [13] propose the called Local Interpretable Model-Agnostic Explanations (LIME) algorithm to locally explain predictions of classifiers and regressors by training an interpretable model to approximate the original model. Afterwards, Shrikumar et al. propose</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Fig. 2 .</head><label>2</label><figDesc>Fig. 2. Machine Learning System Development workflow in ADD framework. Three blocks (Data Collection, Data Analysis/Pre-processing and Model Construction) in the dotted rectangle are considered as an implementation component, which creates the ML system. The system can go to the deployment stage if it passes all the scenarios tests or it must be improved if it fails the test in any of those scenarios.</figDesc><graphic coords="4,203.93,257.62,207.49,161.36" type="bitmap" /></figure>
		</body>
		<back>

			<div type="funding">
<div xmlns="http://www.tei-c.org/ns/1.0"><p>This research is supported by the RAInS project funded by EPSRC</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">On pixel-wise explanations for non-linear classifier decisions by layer-wise relevance propagation</title>
		<author>
			<persName><forename type="first">S</forename><surname>Bach</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Binder</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Montavon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Klauschen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">R</forename><surname>Müller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Samek</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">PloS one</title>
		<imprint>
			<biblScope unit="volume">10</biblScope>
			<biblScope unit="issue">7</biblScope>
			<biblScope unit="page">e0130140</biblScope>
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<monogr>
		<title level="m" type="main">Test-driven development: by example</title>
		<author>
			<persName><forename type="first">K</forename><surname>Beck</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2003">2003</date>
			<publisher>Addison-Wesley Professional</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Algorithmic transparency via quantitative input influence: Theory and experiments with learning systems</title>
		<author>
			<persName><forename type="first">A</forename><surname>Datta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Sen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Zick</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IEEE symposium on security and privacy (SP)</title>
				<imprint>
			<publisher>IEEE</publisher>
			<date type="published" when="2016">2016. 2016</date>
			<biblScope unit="page" from="598" to="617" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<monogr>
		<ptr target="https://cloud.google.com/ai-platform/docs/ml-solutions-overview" />
		<title level="m">Google: Machine learning workflow</title>
				<imprint>
			<date type="published" when="2021-03-25">Mar 25, 2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Discrimination prevention in data mining for intrusion and crime detection</title>
		<author>
			<persName><forename type="first">S</forename><surname>Hajian</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Domingo-Ferrer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Martinez-Balleste</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IEEE Symposium on Computational Intelligence in Cyber Security (CICS)</title>
				<imprint>
			<publisher>IEEE</publisher>
			<date type="published" when="2011">2011. 2011</date>
			<biblScope unit="page" from="47" to="54" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Behavior driven development: Beter dan test driven development</title>
		<author>
			<persName><forename type="first">R</forename><surname>Haring</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>De Ruiter</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Java Magazine</title>
		<imprint>
			<biblScope unit="volume">29</biblScope>
			<date type="published" when="2011">2011</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<title level="m" type="main">A unified approach to interpreting model predictions</title>
		<author>
			<persName><forename type="first">S</forename><surname>Lundberg</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">I</forename><surname>Lee</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1705.07874</idno>
		<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b7">
	<monogr>
		<author>
			<persName><surname>Mckinsey</surname></persName>
		</author>
		<ptr target="https://www.mckinsey.com/business-functions/mckinsey-analytics/our-insights/global-survey-the-state-of-ai-in-2020" />
		<title level="m">The state of ai in 2020</title>
				<imprint>
			<date type="published" when="2021-03-28">Mar 28, 2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Model cards for model reporting</title>
		<author>
			<persName><forename type="first">M</forename><surname>Mitchell</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Zaldivar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Barnes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Vasserman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Hutchinson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Spitzer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><forename type="middle">D</forename><surname>Raji</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Gebru</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the conference on fairness, accountability, and transparency</title>
				<meeting>the conference on fairness, accountability, and transparency</meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="220" to="229" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<title level="m" type="main">A semantic framework to support ai system accountability and audit</title>
		<author>
			<persName><forename type="first">I</forename><surname>Naja</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Markovic</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Edwards</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Cottrill</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<author>
			<persName><forename type="first">J</forename><surname>Newkirk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">A</forename><surname>Vorontsov</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Test-driven development in Microsoft</title>
				<meeting><address><addrLine>, WA</addrLine></address></meeting>
		<imprint>
			<publisher>Microsoft Press Redmond</publisher>
			<date type="published" when="2004">2004</date>
			<biblScope unit="volume">1</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<monogr>
		<author>
			<persName><forename type="first">D</forename><surname>North</surname></persName>
		</author>
		<ptr target="http://dannorth.net/introducing-bdd" />
		<title level="m">faster organizations, faster software</title>
				<imprint>
			<date type="published" when="2021-03-30">Mar 30, 2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">why should i trust you?&quot; explaining the predictions of any classifier</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">T</forename><surname>Ribeiro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Singh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Guestrin</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining</title>
				<meeting>the 22nd ACM SIGKDD international conference on knowledge discovery and data mining</meeting>
		<imprint>
			<date type="published" when="2016">2016</date>
			<biblScope unit="page" from="1135" to="1144" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Learning important features through propagating activation differences</title>
		<author>
			<persName><forename type="first">A</forename><surname>Shrikumar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Greenside</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Kundaje</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">International Conference on Machine Learning</title>
				<imprint>
			<publisher>PMLR</publisher>
			<date type="published" when="2017">2017</date>
			<biblScope unit="page" from="3145" to="3153" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<monogr>
		<ptr target="https://cucumber.io/docs/guides/overview/,on-line" />
		<title level="m">what is cucumber</title>
				<imprint/>
	</monogr>
	<note>SmartBear</note>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">A study of the characteristics of behaviour driven development</title>
		<author>
			<persName><forename type="first">C</forename><surname>Solis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Wang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">37th EUROMICRO Conference on Software Engineering and Advanced Applications</title>
				<imprint>
			<publisher>IEEE</publisher>
			<date type="published" when="2011">2011. 2011</date>
			<biblScope unit="page" from="383" to="387" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Machine learning for networking: Workflow, advances and opportunities</title>
		<author>
			<persName><forename type="first">M</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Cui</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Xiao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Jiang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Network</title>
		<imprint>
			<biblScope unit="volume">32</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="92" to="99" />
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
