<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">The Effect of Severity Ratings on Software Developers&apos; Priority of Usability Inspection Results</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author role="corresp">
							<persName><forename type="first">Asbjørn</forename><surname>Følstad</surname></persName>
							<email>asf@sintef.no</email>
							<affiliation key="aff0">
								<orgName type="institution">SINTEF ICT</orgName>
								<address>
									<addrLine>Forskningsveien 1</addrLine>
									<postCode>0314, +47 22067515</postCode>
									<settlement>Oslo</settlement>
									<country key="NO">Norway</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">The Effect of Severity Ratings on Software Developers&apos; Priority of Usability Inspection Results</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">F5B442460630B028B341F11C69092B40</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T09:56+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>H5.m.Information interfaces and presentation (e.g.</term>
					<term>HCI): Miscellaneous Usability evaluation</term>
					<term>usability inspection</term>
					<term>developers&apos; priority</term>
					<term>impact</term>
					<term>severity ratings</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Knowledge of the factors that affect developers' priority of usability evaluation results is important in order to improve the interplay between usability evaluation and software development.</p><p>In the presented study, the effect of usability inspection severity ratings on the developers' priority of evaluation results was investigated. The usability inspection results with higher severity ratings were associated with higher developer priority. This result contradicts Sawyer et al. [7], but is in line with Law's <ref type="bibr" target="#b5">[5,</ref><ref type="bibr" target="#b6">6]</ref> finding related to the impact of user test results. The findings serve as a reminder for HCI professionals to focus on high severity issues.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">INTRODUCTION</head><p>One important indicator of successful interplay between usability evaluation and software development is the extent to which evaluation results are associated with subsequent changes in the system under development. This indicator, termed the "impact" <ref type="bibr">[7]</ref> or "persuasive power" <ref type="bibr" target="#b4">[4]</ref> of usability evaluation results, may reflect whether or not a usability evaluation has generated results that are needed in the development process.</p><p>Problem severity is a characteristic of usability evaluation results that has been suggested to affect the impact of usability evaluation results. There is, however, divergence in the literature regarding the actual effect of severity ratings on developers' prioritizing of usability evaluation results. Sawyer et al.'s <ref type="bibr">[7]</ref> study of the impact of usability inspection results indicated that usability inspectors' severity ratings had no effect on the impact of the evaluation results; reported impact ratios were 72% (low severity issues), 71% (medium severity issues), 72% (high severity issues). In contrast to this finding Law <ref type="bibr" target="#b5">[5,</ref><ref type="bibr" target="#b6">6]</ref>, in a study of the impact of user tests, reported a tendency towards higher severity results having higher impact; reported impact ratios were 26% (minor problems), 42% (moderate problems), 47% (severe problems). Law's findings, however, were not statistically significant <ref type="bibr" target="#b5">[5]</ref>. Hertzum <ref type="bibr" target="#b3">[3]</ref> suggested that the effect of severity classifications may change across the development process, e.g. high severity evaluation results may have relatively higher impact in later phases of development. Law's study was conducted relatively late in the development process, on the running prototype of a digital library. Sawyer et al. did not report in which development phases their usability inspections were conducted.</p><p>In order to complement the existing research on the effect of severity ratings on the impact of evaluation results, an empirical study of the impact of usability inspection results is presented. The data of the present study was collected as part of a larger study reported by Følstad <ref type="bibr" target="#b2">[2]</ref>, but the results discussed below have not previously been presented.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">RESEARCH PROBLEM AND HYPOTHESIS</head><p>The research problem of the present study was formulated as:</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>What is the effect of usability inspectors' severity ratings on developers' priority of usability inspection results?</head><p>The null hypothesis of the study (no effect of severity ratings) followed the findings of Sawyer et al., and the alternative hypothesis (H1) was formulated in line with the findings presented by Law: H1: High severity issues will tend to be prioritized higher by developers than low severity issues.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">METHOD</head><p>Usability inspections were conducted as group-based expert walkthroughs <ref type="bibr" target="#b1">[1]</ref>. The objects of evaluation were three mobile work-support systems for medical personnel at hospitals, politicians and political advisors, and parking wardens respectively. All systems were in late phases of development, running prototypes close to market. The usability inspectors were 13 HCI professionals, all with &gt;1 year work experience (Mdn=5 Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. I-USED'08, September 24, 2008, Pisa, Italy years) <ref type="foot" target="#foot_0">1</ref> . Each inspector participated in one of three evaluation groups, one group for each object of evaluation. The walkthroughs were conducted as two-stage processes where (1) the individual evaluators noted down usability issues (usability problems and change suggestions) and (2) a common set of usability issues were agreed on in the group. All usability issues were to be classified as either Critical (will probably stop typical users in using the application to solve the task), Serious (will probably cause serious delay for typical users …), or Cosmetic (will probably cause minor delay …). The output of the usability inspections was one report for each object of evaluation, delivered to each of the three development teams respectively. Three months after the evaluation reports had been delivered individual interviews were conducted with development team representatives. The representatives were requested to prioritize all usability issues according to the following: High (change has already been done, or will be done no later than six months after receiving the evaluation report), Medium (change is relevant but will not be prioritized the first six months), Low (change will not be prioritized), Wrong (the item is perceived by the developer to be a misjudgment). In order to align the resulting developers' priorities with the impact ratio definitions of Law and Sawyer et al., the priority High was recoded as "Change", and the priorities Medium, Low and Wrong were recoded as "No change".</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">RESULTS</head><p>The evaluation groups generated totally 167 usability issues. The three objects of evaluation were associated with 44, 61, and 62 usability issues respectively. The total impact ratio (number of issues associated with change/total number of issues [following 7 and 6]) was 27%, which is relatively low. The relationship between the developers' priorities and the usability inspectors' severity ratings is presented in Table <ref type="table" target="#tab_0">1</ref>. Visual inspection of Table <ref type="table" target="#tab_0">1</ref> shows a tendency towards higher priority given to usability issues with severity ratings serious and critical. A Pearson Chi-Square test showed statistically significant differences in priority between severity rating groups; X 2 =14.446, df=3, p(one-sided)=.001.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">DISCUSSION</head><p>The presented results indicate that severity ratings may have significant impact on developers' priority of results from usability inspections. This finding contributes to our understanding of severity ratings as a characteristic of usability evaluation results that may help to identify which usability evaluation results that are needed in software development. phases their usability inspections were associated with, but their relatively high impact ratios suggest that their inspections possibly may have been conducted in earlier project phases.</p><p>The present study, as the study of Law, indicates that the identification of a low severity usability issue typically is of less value to software developers than the identification of a high severity issue. This should serve as a reminder for HCI professionals to spend evaluation resources on identification and communication of higher severity usability issues.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 . Usability issues distributed across developers' priorities and usability inspectors' severity ratings</head><label>1</label><figDesc></figDesc><table><row><cell></cell><cell>Not Classified</cell><cell>Cosmetic</cell><cell>Serious</cell><cell>Critical</cell></row><row><cell>Change</cell><cell>6</cell><cell>9</cell><cell>18</cell><cell>12</cell></row><row><cell>No change</cell><cell>46</cell><cell>31</cell><cell>26</cell><cell>16</cell></row><row><cell>Impact ratio</cell><cell>12%</cell><cell>23%</cell><cell>41%</cell><cell>43%</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head></head><label></label><figDesc>The finding is particularly interesting since it contradicts the conclusions of Sawyer et al. and therefore may provoke necessary rethinking regarding usability inspectors ability to provide severity assessments that are useful to software engineers.It is also interesting to note that the results are fully in line with Law's findings related to severity ratings of user test results. The present study may thus serve to strengthen Law's conclusions. Curiously, the impact ratios of the different severity levels in Law's study and the present study are close to being identical.Why, then, do the present study indicate that the severity ratings of usability inspection results may have an effect on the developers' priority, when Sawyer et al. did not find a similar effect? One reason may be the relatively high impact ratios reported bySawyer et  al., something that may well result in a greater proportion of low severity issues being prioritized. Another reason may be that the present study, as the study of Law, favored high severity evaluation results since the usability evaluations were conducted relatively late in the development process [cf. 3]. Sawyer et al. do not state which development</figDesc><table /></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">The study reported by Følstad also included separate evaluation groups with work-domain experts. The results of these groups were not included in the current study, in order to make a clearcut comparison with the findings of Law and Sawyer et al.</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">ACKNOWLEDGMENTS</head><p>This paper has been written as part of the RECORD project, supported by the VERDIKT program of the Norwegian Research Council.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<title/>
		<author>
			<persName><surname>References</surname></persName>
		</author>
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Group-based Expert Walkthrough</title>
		<author>
			<persName><forename type="first">A</forename><surname>Følstad</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">R3UEMs: Review, Report and Refine Usability Evaluation Methods. Proceedings of the 3rd. COST294-MAUSE International Workshop</title>
				<editor>
			<persName><forename type="first">D</forename><surname>Scapin</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">E</forename><forename type="middle">L</forename></persName>
		</editor>
		<editor>
			<persName><forename type="first">-C</forename><surname>Law</surname></persName>
		</editor>
		<imprint>
			<date type="published" when="2007">2007</date>
			<biblScope unit="page" from="58" to="60" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Work-Domain Experts as Evaluators: Usability Inspection of Domain-Specific Work-Support Systems</title>
		<author>
			<persName><forename type="first">A</forename><surname>Følstad</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">International Journal of Human-Computer Interaction</title>
		<imprint>
			<biblScope unit="volume">22</biblScope>
			<biblScope unit="issue">3</biblScope>
			<biblScope unit="page" from="217" to="245" />
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Problem Prioritization in Usability Evaluation: From Severity Assessments Toward Impact on Design</title>
		<author>
			<persName><forename type="first">M</forename><surname>Hertzum</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">International Journal of Human-Computer Interaction</title>
		<imprint>
			<biblScope unit="volume">21</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="125" to="146" />
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Tracking the effectiveness of usability evaluation methods</title>
		<author>
			<persName><forename type="first">B</forename><forename type="middle">E</forename><surname>John</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">J</forename><surname>Marks</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Behaviour &amp; Information Technology</title>
		<imprint>
			<biblScope unit="volume">16</biblScope>
			<biblScope unit="page" from="188" to="202" />
			<date type="published" when="1997">1997</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">A Multi-Perspective Approach to Tracking the Effectiveness of User Tests: A Case Study</title>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">L</forename><surname>Law</surname></persName>
		</author>
		<author>
			<persName><forename type="first">-C</forename></persName>
		</author>
		<idno>Report no. 2004/2</idno>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the NordiCHI Workshop on Improving the Interplay Between Usability Evaluation and User Interface Design</title>
				<editor>
			<persName><forename type="first">K</forename><surname>Hornbaek</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Stage</surname></persName>
		</editor>
		<meeting>the NordiCHI Workshop on Improving the Interplay Between Usability Evaluation and User Interface Design</meeting>
		<imprint>
			<date type="published" when="2004">2004</date>
			<biblScope unit="page" from="36" to="40" />
		</imprint>
		<respStmt>
			<orgName>University of Aalborg, HCI Lab</orgName>
		</respStmt>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Evaluating the Downstream Utility of User Tests and Examining the Developer Effect: A Case Study</title>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">L</forename><surname>Law</surname></persName>
		</author>
		<author>
			<persName><forename type="first">-C</forename></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">International Journal of Human-Computer Interaction</title>
		<imprint>
			<biblScope unit="volume">21</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="147" to="172" />
			<date type="published" when="2006">2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Making a Difference -The Impact of Inspections</title>
		<author>
			<persName><forename type="first">P</forename><surname>Sawyer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Flanders</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Wixon</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the CHI&apos;96 Conference on Human Factors in Computing Systems</title>
				<meeting>the CHI&apos;96 Conference on Human Factors in Computing Systems</meeting>
		<imprint>
			<date type="published" when="1996">1996</date>
			<biblScope unit="page" from="376" to="382" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
